var/home/core/zuul-output/0000755000175000017500000000000015110070575014526 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015110106247015466 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000005501025015110106237017667 0ustar rootrootNov 21 14:08:41 crc systemd[1]: Starting Kubernetes Kubelet... Nov 21 14:08:41 crc restorecon[4692]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:41 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 21 14:08:42 crc restorecon[4692]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 21 14:08:42 crc restorecon[4692]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Nov 21 14:08:43 crc kubenswrapper[4897]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 21 14:08:43 crc kubenswrapper[4897]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Nov 21 14:08:43 crc kubenswrapper[4897]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 21 14:08:43 crc kubenswrapper[4897]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 21 14:08:43 crc kubenswrapper[4897]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Nov 21 14:08:43 crc kubenswrapper[4897]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.669267 4897 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.674209 4897 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.674230 4897 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.674238 4897 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.674243 4897 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.674250 4897 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.674255 4897 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.674260 4897 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.674266 4897 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.674272 4897 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.674278 4897 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.674283 4897 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.674289 4897 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.674293 4897 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.674299 4897 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.674304 4897 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.674309 4897 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.674314 4897 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.674319 4897 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.674326 4897 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.674331 4897 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.674336 4897 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.674342 4897 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.674360 4897 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.674366 4897 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.674373 4897 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.674379 4897 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.674386 4897 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.674392 4897 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.674399 4897 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.674408 4897 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.674416 4897 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.674423 4897 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.674430 4897 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.674436 4897 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.674442 4897 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.674448 4897 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.674453 4897 feature_gate.go:330] unrecognized feature gate: Example Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.674458 4897 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.674463 4897 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.674468 4897 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.674474 4897 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.674479 4897 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.674484 4897 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.674489 4897 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.674494 4897 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.674520 4897 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.674528 4897 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.674534 4897 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.674540 4897 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.674547 4897 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.674554 4897 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.674562 4897 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.674568 4897 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.674574 4897 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.674580 4897 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.674585 4897 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.674590 4897 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.674597 4897 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.674602 4897 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.674607 4897 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.674612 4897 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.674618 4897 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.674623 4897 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.674630 4897 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.674637 4897 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.674642 4897 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.674648 4897 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.674654 4897 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.674662 4897 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.674668 4897 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.674673 4897 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.674775 4897 flags.go:64] FLAG: --address="0.0.0.0" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.674788 4897 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.674804 4897 flags.go:64] FLAG: --anonymous-auth="true" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.674813 4897 flags.go:64] FLAG: --application-metrics-count-limit="100" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.674821 4897 flags.go:64] FLAG: --authentication-token-webhook="false" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.674827 4897 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.674836 4897 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.674845 4897 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.674851 4897 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.674859 4897 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.674866 4897 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.674873 4897 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.674879 4897 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.674891 4897 flags.go:64] FLAG: --cgroup-root="" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.674898 4897 flags.go:64] FLAG: --cgroups-per-qos="true" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.674904 4897 flags.go:64] FLAG: --client-ca-file="" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.674910 4897 flags.go:64] FLAG: --cloud-config="" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.674916 4897 flags.go:64] FLAG: --cloud-provider="" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.674922 4897 flags.go:64] FLAG: --cluster-dns="[]" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.674929 4897 flags.go:64] FLAG: --cluster-domain="" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.674935 4897 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.674942 4897 flags.go:64] FLAG: --config-dir="" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.674947 4897 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.674955 4897 flags.go:64] FLAG: --container-log-max-files="5" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.674963 4897 flags.go:64] FLAG: --container-log-max-size="10Mi" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.674970 4897 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.674976 4897 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.674983 4897 flags.go:64] FLAG: --containerd-namespace="k8s.io" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.674989 4897 flags.go:64] FLAG: --contention-profiling="false" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.674995 4897 flags.go:64] FLAG: --cpu-cfs-quota="true" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675001 4897 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675007 4897 flags.go:64] FLAG: --cpu-manager-policy="none" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675013 4897 flags.go:64] FLAG: --cpu-manager-policy-options="" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675020 4897 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675026 4897 flags.go:64] FLAG: --enable-controller-attach-detach="true" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675032 4897 flags.go:64] FLAG: --enable-debugging-handlers="true" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675038 4897 flags.go:64] FLAG: --enable-load-reader="false" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675044 4897 flags.go:64] FLAG: --enable-server="true" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675050 4897 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675059 4897 flags.go:64] FLAG: --event-burst="100" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675065 4897 flags.go:64] FLAG: --event-qps="50" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675073 4897 flags.go:64] FLAG: --event-storage-age-limit="default=0" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675080 4897 flags.go:64] FLAG: --event-storage-event-limit="default=0" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675086 4897 flags.go:64] FLAG: --eviction-hard="" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675094 4897 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675100 4897 flags.go:64] FLAG: --eviction-minimum-reclaim="" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675106 4897 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675112 4897 flags.go:64] FLAG: --eviction-soft="" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675118 4897 flags.go:64] FLAG: --eviction-soft-grace-period="" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675132 4897 flags.go:64] FLAG: --exit-on-lock-contention="false" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675139 4897 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675145 4897 flags.go:64] FLAG: --experimental-mounter-path="" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675151 4897 flags.go:64] FLAG: --fail-cgroupv1="false" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675157 4897 flags.go:64] FLAG: --fail-swap-on="true" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675164 4897 flags.go:64] FLAG: --feature-gates="" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675173 4897 flags.go:64] FLAG: --file-check-frequency="20s" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675180 4897 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675188 4897 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675196 4897 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675203 4897 flags.go:64] FLAG: --healthz-port="10248" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675222 4897 flags.go:64] FLAG: --help="false" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675229 4897 flags.go:64] FLAG: --hostname-override="" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675235 4897 flags.go:64] FLAG: --housekeeping-interval="10s" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675241 4897 flags.go:64] FLAG: --http-check-frequency="20s" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675247 4897 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675254 4897 flags.go:64] FLAG: --image-credential-provider-config="" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675260 4897 flags.go:64] FLAG: --image-gc-high-threshold="85" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675266 4897 flags.go:64] FLAG: --image-gc-low-threshold="80" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675273 4897 flags.go:64] FLAG: --image-service-endpoint="" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675279 4897 flags.go:64] FLAG: --kernel-memcg-notification="false" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675285 4897 flags.go:64] FLAG: --kube-api-burst="100" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675291 4897 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675297 4897 flags.go:64] FLAG: --kube-api-qps="50" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675304 4897 flags.go:64] FLAG: --kube-reserved="" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675311 4897 flags.go:64] FLAG: --kube-reserved-cgroup="" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675317 4897 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675323 4897 flags.go:64] FLAG: --kubelet-cgroups="" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675329 4897 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675335 4897 flags.go:64] FLAG: --lock-file="" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675341 4897 flags.go:64] FLAG: --log-cadvisor-usage="false" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675347 4897 flags.go:64] FLAG: --log-flush-frequency="5s" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675354 4897 flags.go:64] FLAG: --log-json-info-buffer-size="0" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675363 4897 flags.go:64] FLAG: --log-json-split-stream="false" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675369 4897 flags.go:64] FLAG: --log-text-info-buffer-size="0" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675376 4897 flags.go:64] FLAG: --log-text-split-stream="false" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675383 4897 flags.go:64] FLAG: --logging-format="text" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675389 4897 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675398 4897 flags.go:64] FLAG: --make-iptables-util-chains="true" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675405 4897 flags.go:64] FLAG: --manifest-url="" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675411 4897 flags.go:64] FLAG: --manifest-url-header="" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675420 4897 flags.go:64] FLAG: --max-housekeeping-interval="15s" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675426 4897 flags.go:64] FLAG: --max-open-files="1000000" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675433 4897 flags.go:64] FLAG: --max-pods="110" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675439 4897 flags.go:64] FLAG: --maximum-dead-containers="-1" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675445 4897 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675451 4897 flags.go:64] FLAG: --memory-manager-policy="None" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675457 4897 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675463 4897 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675469 4897 flags.go:64] FLAG: --node-ip="192.168.126.11" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675475 4897 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675489 4897 flags.go:64] FLAG: --node-status-max-images="50" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675496 4897 flags.go:64] FLAG: --node-status-update-frequency="10s" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675519 4897 flags.go:64] FLAG: --oom-score-adj="-999" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675526 4897 flags.go:64] FLAG: --pod-cidr="" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675533 4897 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675544 4897 flags.go:64] FLAG: --pod-manifest-path="" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675551 4897 flags.go:64] FLAG: --pod-max-pids="-1" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675557 4897 flags.go:64] FLAG: --pods-per-core="0" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675563 4897 flags.go:64] FLAG: --port="10250" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675570 4897 flags.go:64] FLAG: --protect-kernel-defaults="false" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675576 4897 flags.go:64] FLAG: --provider-id="" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675582 4897 flags.go:64] FLAG: --qos-reserved="" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675588 4897 flags.go:64] FLAG: --read-only-port="10255" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675594 4897 flags.go:64] FLAG: --register-node="true" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675600 4897 flags.go:64] FLAG: --register-schedulable="true" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675607 4897 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675617 4897 flags.go:64] FLAG: --registry-burst="10" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675623 4897 flags.go:64] FLAG: --registry-qps="5" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675686 4897 flags.go:64] FLAG: --reserved-cpus="" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675692 4897 flags.go:64] FLAG: --reserved-memory="" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675701 4897 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675716 4897 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675723 4897 flags.go:64] FLAG: --rotate-certificates="false" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675729 4897 flags.go:64] FLAG: --rotate-server-certificates="false" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675734 4897 flags.go:64] FLAG: --runonce="false" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675740 4897 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675746 4897 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675752 4897 flags.go:64] FLAG: --seccomp-default="false" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675758 4897 flags.go:64] FLAG: --serialize-image-pulls="true" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675764 4897 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675770 4897 flags.go:64] FLAG: --storage-driver-db="cadvisor" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675779 4897 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675786 4897 flags.go:64] FLAG: --storage-driver-password="root" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675792 4897 flags.go:64] FLAG: --storage-driver-secure="false" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675798 4897 flags.go:64] FLAG: --storage-driver-table="stats" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675804 4897 flags.go:64] FLAG: --storage-driver-user="root" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675811 4897 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675818 4897 flags.go:64] FLAG: --sync-frequency="1m0s" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675824 4897 flags.go:64] FLAG: --system-cgroups="" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675831 4897 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675840 4897 flags.go:64] FLAG: --system-reserved-cgroup="" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675846 4897 flags.go:64] FLAG: --tls-cert-file="" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675852 4897 flags.go:64] FLAG: --tls-cipher-suites="[]" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675867 4897 flags.go:64] FLAG: --tls-min-version="" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675873 4897 flags.go:64] FLAG: --tls-private-key-file="" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675879 4897 flags.go:64] FLAG: --topology-manager-policy="none" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675885 4897 flags.go:64] FLAG: --topology-manager-policy-options="" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675891 4897 flags.go:64] FLAG: --topology-manager-scope="container" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675898 4897 flags.go:64] FLAG: --v="2" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675906 4897 flags.go:64] FLAG: --version="false" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675914 4897 flags.go:64] FLAG: --vmodule="" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675921 4897 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.675927 4897 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.676074 4897 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.676081 4897 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.676087 4897 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.676095 4897 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.676112 4897 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.676119 4897 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.676126 4897 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.676132 4897 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.676139 4897 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.676145 4897 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.676152 4897 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.676158 4897 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.676164 4897 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.676170 4897 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.676177 4897 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.676183 4897 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.676188 4897 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.676194 4897 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.676199 4897 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.676205 4897 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.676211 4897 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.676216 4897 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.676221 4897 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.676226 4897 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.676231 4897 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.676236 4897 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.676241 4897 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.676246 4897 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.676251 4897 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.676256 4897 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.676262 4897 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.676267 4897 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.676272 4897 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.676277 4897 feature_gate.go:330] unrecognized feature gate: Example Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.676284 4897 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.676291 4897 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.676299 4897 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.676305 4897 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.676313 4897 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.676320 4897 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.676329 4897 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.676336 4897 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.676343 4897 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.676350 4897 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.676358 4897 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.676364 4897 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.676371 4897 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.676377 4897 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.676383 4897 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.676390 4897 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.676397 4897 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.676404 4897 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.676410 4897 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.676416 4897 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.676423 4897 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.676430 4897 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.676436 4897 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.676443 4897 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.676449 4897 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.676454 4897 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.676459 4897 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.676465 4897 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.676470 4897 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.676475 4897 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.676481 4897 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.676486 4897 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.676491 4897 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.676496 4897 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.676522 4897 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.676529 4897 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.676536 4897 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.679173 4897 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.691734 4897 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.691792 4897 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.691882 4897 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.691894 4897 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.691902 4897 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.691908 4897 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.691913 4897 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.691918 4897 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.691922 4897 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.691927 4897 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.691932 4897 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.691936 4897 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.691941 4897 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.691945 4897 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.691949 4897 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.691954 4897 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.691958 4897 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.691963 4897 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.691968 4897 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.691976 4897 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.691982 4897 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.691988 4897 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.691995 4897 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692001 4897 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692007 4897 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692014 4897 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692021 4897 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692027 4897 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692033 4897 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692039 4897 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692044 4897 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692049 4897 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692054 4897 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692058 4897 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692063 4897 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692068 4897 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692074 4897 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692079 4897 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692084 4897 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692090 4897 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692094 4897 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692100 4897 feature_gate.go:330] unrecognized feature gate: Example Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692104 4897 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692111 4897 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692122 4897 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692128 4897 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692135 4897 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692140 4897 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692146 4897 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692150 4897 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692155 4897 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692161 4897 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692166 4897 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692173 4897 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692179 4897 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692186 4897 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692191 4897 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692196 4897 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692200 4897 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692205 4897 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692209 4897 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692213 4897 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692218 4897 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692223 4897 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692229 4897 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692234 4897 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692238 4897 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692242 4897 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692246 4897 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692250 4897 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692256 4897 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692260 4897 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692265 4897 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.692274 4897 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692446 4897 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692457 4897 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692463 4897 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692468 4897 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692473 4897 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692478 4897 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692482 4897 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692487 4897 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692493 4897 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692498 4897 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692535 4897 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692540 4897 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692545 4897 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692551 4897 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692557 4897 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692562 4897 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692567 4897 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692573 4897 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692578 4897 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692583 4897 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692588 4897 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692592 4897 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692596 4897 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692601 4897 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692605 4897 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692610 4897 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692614 4897 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692618 4897 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692622 4897 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692629 4897 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692634 4897 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692640 4897 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692645 4897 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692649 4897 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692654 4897 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692658 4897 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692663 4897 feature_gate.go:330] unrecognized feature gate: Example Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692667 4897 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692671 4897 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692675 4897 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692680 4897 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692686 4897 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692690 4897 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692695 4897 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692699 4897 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692704 4897 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692709 4897 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692714 4897 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692720 4897 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692726 4897 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692730 4897 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692736 4897 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692741 4897 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692747 4897 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692751 4897 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692758 4897 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692763 4897 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692768 4897 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692773 4897 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692779 4897 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692787 4897 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692793 4897 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692798 4897 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692803 4897 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692807 4897 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692813 4897 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692818 4897 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692822 4897 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692828 4897 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692833 4897 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.692837 4897 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.692845 4897 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.693094 4897 server.go:940] "Client rotation is on, will bootstrap in background" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.698861 4897 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.698986 4897 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.706197 4897 server.go:997] "Starting client certificate rotation" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.706235 4897 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.706436 4897 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2026-01-14 17:49:49.309242898 +0000 UTC Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.706859 4897 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 1299h41m5.602389151s for next certificate rotation Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.798396 4897 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.804063 4897 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.857635 4897 log.go:25] "Validated CRI v1 runtime API" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.923145 4897 log.go:25] "Validated CRI v1 image API" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.926221 4897 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.939012 4897 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-11-21-14-03-40-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.939070 4897 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:41 fsType:tmpfs blockSize:0}] Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.963786 4897 manager.go:217] Machine: {Timestamp:2025-11-21 14:08:43.960998726 +0000 UTC m=+1.245592221 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2799998 MemoryCapacity:33654124544 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:8a210af6-a71f-4ea9-90a3-aac32fb868b7 BootID:8b8bdf9e-0fef-41ac-98ad-710dc2137183 Filesystems:[{Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827060224 Type:vfs Inodes:4108169 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365408768 Type:vfs Inodes:821633 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:41 Capacity:1073741824 Type:vfs Inodes:4108169 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:e6:10:59 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:e6:10:59 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:ca:da:6a Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:bf:42:e2 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:90:69:f8 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:65:11:f6 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:4a:0e:a2:1b:2d:bb Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:2a:7e:df:3d:0a:6a Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654124544 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.964027 4897 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.964234 4897 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.968947 4897 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.969133 4897 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.969171 4897 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.969476 4897 topology_manager.go:138] "Creating topology manager with none policy" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.969489 4897 container_manager_linux.go:303] "Creating device plugin manager" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.970024 4897 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.970053 4897 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.970213 4897 state_mem.go:36] "Initialized new in-memory state store" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.970302 4897 server.go:1245] "Using root directory" path="/var/lib/kubelet" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.978108 4897 kubelet.go:418] "Attempting to sync node with API server" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.978168 4897 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.978226 4897 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.978247 4897 kubelet.go:324] "Adding apiserver pod source" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.978659 4897 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.991657 4897 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.992830 4897 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Nov 21 14:08:43 crc kubenswrapper[4897]: I1121 14:08:43.996560 4897 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.998937 4897 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.233:6443: connect: connection refused Nov 21 14:08:43 crc kubenswrapper[4897]: E1121 14:08:43.999039 4897 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.233:6443: connect: connection refused" logger="UnhandledError" Nov 21 14:08:43 crc kubenswrapper[4897]: W1121 14:08:43.999246 4897 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.233:6443: connect: connection refused Nov 21 14:08:43 crc kubenswrapper[4897]: E1121 14:08:43.999388 4897 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.233:6443: connect: connection refused" logger="UnhandledError" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.000634 4897 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.000681 4897 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.000694 4897 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.000704 4897 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.000721 4897 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.000734 4897 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.000745 4897 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.000761 4897 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.000773 4897 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.000784 4897 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.000812 4897 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.000822 4897 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.003833 4897 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.004542 4897 server.go:1280] "Started kubelet" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.004978 4897 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.004977 4897 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.005671 4897 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.233:6443: connect: connection refused Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.006012 4897 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Nov 21 14:08:44 crc systemd[1]: Started Kubernetes Kubelet. Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.007777 4897 server.go:460] "Adding debug handlers to kubelet server" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.009076 4897 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.009154 4897 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.009165 4897 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-11 12:24:31.739084375 +0000 UTC Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.009291 4897 volume_manager.go:287] "The desired_state_of_world populator starts" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.009320 4897 volume_manager.go:289] "Starting Kubelet Volume Manager" Nov 21 14:08:44 crc kubenswrapper[4897]: E1121 14:08:44.009331 4897 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.009421 4897 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Nov 21 14:08:44 crc kubenswrapper[4897]: W1121 14:08:44.010266 4897 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.233:6443: connect: connection refused Nov 21 14:08:44 crc kubenswrapper[4897]: E1121 14:08:44.010458 4897 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.233:6443: connect: connection refused" logger="UnhandledError" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.011093 4897 factory.go:55] Registering systemd factory Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.011136 4897 factory.go:221] Registration of the systemd container factory successfully Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.011641 4897 factory.go:153] Registering CRI-O factory Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.011682 4897 factory.go:221] Registration of the crio container factory successfully Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.011811 4897 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.011865 4897 factory.go:103] Registering Raw factory Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.011894 4897 manager.go:1196] Started watching for new ooms in manager Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.012912 4897 manager.go:319] Starting recovery of all containers Nov 21 14:08:44 crc kubenswrapper[4897]: E1121 14:08:44.012933 4897 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.233:6443: connect: connection refused" interval="200ms" Nov 21 14:08:44 crc kubenswrapper[4897]: E1121 14:08:44.012908 4897 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.233:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187a0ad75385817d default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-21 14:08:44.004475261 +0000 UTC m=+1.289068756,LastTimestamp:2025-11-21 14:08:44.004475261 +0000 UTC m=+1.289068756,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.023605 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.024232 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.024254 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.024269 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.024651 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.024678 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.024701 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.024715 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.024761 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.024779 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.024793 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.024812 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.024827 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.024854 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.024869 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.026102 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.026192 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.026213 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.026239 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.026259 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.026284 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.026303 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.026321 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.026346 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.026366 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.026390 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.026413 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.026443 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.026484 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.026520 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.026538 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.026560 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.026578 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.026595 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.026617 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.026635 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.026657 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.026674 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.026693 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.026715 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.026734 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.026758 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.026776 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.026792 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.026816 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.027018 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.027044 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.027069 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.027091 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.027112 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.027129 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.027150 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.027172 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.027197 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.027220 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.027239 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.027262 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.027280 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.027300 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.027317 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.027334 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.027354 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.027381 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.027439 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.027497 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.027528 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.027549 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.027563 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.027580 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.027592 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.027605 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.027621 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.027633 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.027665 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.038483 4897 manager.go:324] Recovery completed Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.050697 4897 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.050834 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.050857 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.050903 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.050933 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.050947 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.050991 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.051005 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.051024 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.051059 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.061429 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.061533 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.061597 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.061665 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.061741 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.061926 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.062008 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.062068 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.062124 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.062178 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.062235 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.062292 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.062356 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.062412 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.062467 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.062542 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.062602 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.062669 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.062728 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.062785 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.063568 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.063616 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.063643 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.063662 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.063682 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.063699 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.063718 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.063734 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.063750 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.063767 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.063781 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.063797 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.063812 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.063826 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.063840 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.063855 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.063869 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.063885 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.063900 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.063914 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.063929 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.063942 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.063956 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.063983 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.064047 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.064092 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.064114 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.064133 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.064154 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.064171 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.064188 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.064206 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.064225 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.064242 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.064261 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.064277 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.064290 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.064368 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.064401 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.064415 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.064429 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.064442 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.064457 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.064470 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.064484 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.064498 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.064533 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.064548 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.064561 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.064574 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.064587 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.064601 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.064615 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.064629 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.064641 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.064655 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.064669 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.064685 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.064698 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.064711 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.064726 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.064740 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.064754 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.064767 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.064796 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.064808 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.064822 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.064836 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.064850 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.064864 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.064879 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.064895 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.064908 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.064922 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.064935 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.064948 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.064962 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.064975 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.064988 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.065003 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.065019 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.065033 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.065046 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.065059 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.065072 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.065085 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.065099 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.065113 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.065127 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.065140 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.065154 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.065168 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.065182 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.065195 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.065209 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.065223 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.065238 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.065251 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.065267 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.065282 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.065296 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.065309 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.065323 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.065339 4897 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.065394 4897 reconstruct.go:97] "Volume reconstruction finished" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.065407 4897 reconciler.go:26] "Reconciler: start to sync state" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.068083 4897 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.070271 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.070299 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.070308 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.071035 4897 cpu_manager.go:225] "Starting CPU manager" policy="none" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.071055 4897 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.071170 4897 state_mem.go:36] "Initialized new in-memory state store" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.084210 4897 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.087378 4897 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.087432 4897 status_manager.go:217] "Starting to sync pod status with apiserver" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.087673 4897 kubelet.go:2335] "Starting kubelet main sync loop" Nov 21 14:08:44 crc kubenswrapper[4897]: E1121 14:08:44.087730 4897 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Nov 21 14:08:44 crc kubenswrapper[4897]: W1121 14:08:44.089121 4897 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.233:6443: connect: connection refused Nov 21 14:08:44 crc kubenswrapper[4897]: E1121 14:08:44.089196 4897 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.233:6443: connect: connection refused" logger="UnhandledError" Nov 21 14:08:44 crc kubenswrapper[4897]: E1121 14:08:44.110195 4897 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.149401 4897 policy_none.go:49] "None policy: Start" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.151752 4897 memory_manager.go:170] "Starting memorymanager" policy="None" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.151827 4897 state_mem.go:35] "Initializing new in-memory state store" Nov 21 14:08:44 crc kubenswrapper[4897]: E1121 14:08:44.188516 4897 kubelet.go:2359] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Nov 21 14:08:44 crc kubenswrapper[4897]: E1121 14:08:44.210324 4897 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.212626 4897 manager.go:334] "Starting Device Plugin manager" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.212676 4897 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.212691 4897 server.go:79] "Starting device plugin registration server" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.213139 4897 eviction_manager.go:189] "Eviction manager: starting control loop" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.213157 4897 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.213372 4897 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.213459 4897 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.213466 4897 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Nov 21 14:08:44 crc kubenswrapper[4897]: E1121 14:08:44.213547 4897 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.233:6443: connect: connection refused" interval="400ms" Nov 21 14:08:44 crc kubenswrapper[4897]: E1121 14:08:44.219825 4897 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.314803 4897 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.316376 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.316466 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.316493 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.316627 4897 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 21 14:08:44 crc kubenswrapper[4897]: E1121 14:08:44.317499 4897 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.233:6443: connect: connection refused" node="crc" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.388829 4897 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.388917 4897 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.389985 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.390020 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.390029 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.390131 4897 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.390271 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.390295 4897 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.391144 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.391207 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.391221 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.391258 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.391274 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.391281 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.391433 4897 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.391623 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.391678 4897 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.392800 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.392799 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.392836 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.392847 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.392824 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.392875 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.392963 4897 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.393148 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.393231 4897 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.394173 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.394229 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.394249 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.394383 4897 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.394442 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.394496 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.394541 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.394888 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.394934 4897 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.395985 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.396019 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.396032 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.396388 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.396434 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.396446 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.396765 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.396805 4897 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.398008 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.398078 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.398092 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.470803 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.471051 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.471137 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.471229 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.471308 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.471383 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.471460 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.471606 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.471694 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.471729 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.471755 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.471781 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.471898 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.471945 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.471973 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.518566 4897 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.520394 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.520453 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.520464 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.520533 4897 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 21 14:08:44 crc kubenswrapper[4897]: E1121 14:08:44.521237 4897 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.233:6443: connect: connection refused" node="crc" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.573550 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.573776 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.573945 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.573809 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.573984 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.574247 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.574282 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.574304 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.574331 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.574334 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.574356 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.574381 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.574389 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.574404 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.574452 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.574466 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.574454 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.574403 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.574490 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.574493 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.574568 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.574582 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.575155 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.575184 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.575125 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.575234 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.575262 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.575343 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.575400 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.575444 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 21 14:08:44 crc kubenswrapper[4897]: E1121 14:08:44.615601 4897 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.233:6443: connect: connection refused" interval="800ms" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.724065 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.732111 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.751264 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.756460 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.777924 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 21 14:08:44 crc kubenswrapper[4897]: W1121 14:08:44.807397 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-e8b6252c5088170516eba3f4b1e8af4254a095283148a533385c8f52b023af71 WatchSource:0}: Error finding container e8b6252c5088170516eba3f4b1e8af4254a095283148a533385c8f52b023af71: Status 404 returned error can't find the container with id e8b6252c5088170516eba3f4b1e8af4254a095283148a533385c8f52b023af71 Nov 21 14:08:44 crc kubenswrapper[4897]: W1121 14:08:44.808708 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-b897be5d76d91c2d0ec2c38b45b26974ef9d4eeb1b81534f66a6af43b82b97ae WatchSource:0}: Error finding container b897be5d76d91c2d0ec2c38b45b26974ef9d4eeb1b81534f66a6af43b82b97ae: Status 404 returned error can't find the container with id b897be5d76d91c2d0ec2c38b45b26974ef9d4eeb1b81534f66a6af43b82b97ae Nov 21 14:08:44 crc kubenswrapper[4897]: W1121 14:08:44.813041 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-9946b7327ccd6e8e4daedba2ea9805b3d9f6127ff3c31db68514034c92392c91 WatchSource:0}: Error finding container 9946b7327ccd6e8e4daedba2ea9805b3d9f6127ff3c31db68514034c92392c91: Status 404 returned error can't find the container with id 9946b7327ccd6e8e4daedba2ea9805b3d9f6127ff3c31db68514034c92392c91 Nov 21 14:08:44 crc kubenswrapper[4897]: W1121 14:08:44.813455 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-ad5d65c3b17b2bc3fcccff61737ecdc0a2b9e0ffdc4582f53cafd79c7de55eb8 WatchSource:0}: Error finding container ad5d65c3b17b2bc3fcccff61737ecdc0a2b9e0ffdc4582f53cafd79c7de55eb8: Status 404 returned error can't find the container with id ad5d65c3b17b2bc3fcccff61737ecdc0a2b9e0ffdc4582f53cafd79c7de55eb8 Nov 21 14:08:44 crc kubenswrapper[4897]: W1121 14:08:44.814292 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-28d286ca8d3a547731b5d92c9e47d2f409080aa0d9565958ffe364e9f9d0bc85 WatchSource:0}: Error finding container 28d286ca8d3a547731b5d92c9e47d2f409080aa0d9565958ffe364e9f9d0bc85: Status 404 returned error can't find the container with id 28d286ca8d3a547731b5d92c9e47d2f409080aa0d9565958ffe364e9f9d0bc85 Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.922055 4897 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.923704 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.923762 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.923775 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:08:44 crc kubenswrapper[4897]: I1121 14:08:44.923811 4897 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 21 14:08:44 crc kubenswrapper[4897]: E1121 14:08:44.924407 4897 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.233:6443: connect: connection refused" node="crc" Nov 21 14:08:45 crc kubenswrapper[4897]: I1121 14:08:45.006963 4897 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.233:6443: connect: connection refused Nov 21 14:08:45 crc kubenswrapper[4897]: I1121 14:08:45.010144 4897 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-16 21:55:45.572900877 +0000 UTC Nov 21 14:08:45 crc kubenswrapper[4897]: E1121 14:08:45.045975 4897 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.233:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187a0ad75385817d default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-21 14:08:44.004475261 +0000 UTC m=+1.289068756,LastTimestamp:2025-11-21 14:08:44.004475261 +0000 UTC m=+1.289068756,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 21 14:08:45 crc kubenswrapper[4897]: I1121 14:08:45.091496 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"e8b6252c5088170516eba3f4b1e8af4254a095283148a533385c8f52b023af71"} Nov 21 14:08:45 crc kubenswrapper[4897]: I1121 14:08:45.092702 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"28d286ca8d3a547731b5d92c9e47d2f409080aa0d9565958ffe364e9f9d0bc85"} Nov 21 14:08:45 crc kubenswrapper[4897]: I1121 14:08:45.094025 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"9946b7327ccd6e8e4daedba2ea9805b3d9f6127ff3c31db68514034c92392c91"} Nov 21 14:08:45 crc kubenswrapper[4897]: I1121 14:08:45.094939 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"ad5d65c3b17b2bc3fcccff61737ecdc0a2b9e0ffdc4582f53cafd79c7de55eb8"} Nov 21 14:08:45 crc kubenswrapper[4897]: I1121 14:08:45.095864 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"b897be5d76d91c2d0ec2c38b45b26974ef9d4eeb1b81534f66a6af43b82b97ae"} Nov 21 14:08:45 crc kubenswrapper[4897]: W1121 14:08:45.158302 4897 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.233:6443: connect: connection refused Nov 21 14:08:45 crc kubenswrapper[4897]: E1121 14:08:45.158415 4897 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.233:6443: connect: connection refused" logger="UnhandledError" Nov 21 14:08:45 crc kubenswrapper[4897]: W1121 14:08:45.272051 4897 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.233:6443: connect: connection refused Nov 21 14:08:45 crc kubenswrapper[4897]: E1121 14:08:45.272197 4897 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.233:6443: connect: connection refused" logger="UnhandledError" Nov 21 14:08:45 crc kubenswrapper[4897]: E1121 14:08:45.417094 4897 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.233:6443: connect: connection refused" interval="1.6s" Nov 21 14:08:45 crc kubenswrapper[4897]: W1121 14:08:45.424412 4897 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.233:6443: connect: connection refused Nov 21 14:08:45 crc kubenswrapper[4897]: E1121 14:08:45.424525 4897 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.233:6443: connect: connection refused" logger="UnhandledError" Nov 21 14:08:45 crc kubenswrapper[4897]: W1121 14:08:45.530753 4897 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.233:6443: connect: connection refused Nov 21 14:08:45 crc kubenswrapper[4897]: E1121 14:08:45.530941 4897 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.233:6443: connect: connection refused" logger="UnhandledError" Nov 21 14:08:45 crc kubenswrapper[4897]: I1121 14:08:45.725416 4897 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 21 14:08:45 crc kubenswrapper[4897]: I1121 14:08:45.727062 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:08:45 crc kubenswrapper[4897]: I1121 14:08:45.727123 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:08:45 crc kubenswrapper[4897]: I1121 14:08:45.727141 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:08:45 crc kubenswrapper[4897]: I1121 14:08:45.727175 4897 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 21 14:08:45 crc kubenswrapper[4897]: E1121 14:08:45.727688 4897 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.233:6443: connect: connection refused" node="crc" Nov 21 14:08:46 crc kubenswrapper[4897]: I1121 14:08:46.006759 4897 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.233:6443: connect: connection refused Nov 21 14:08:46 crc kubenswrapper[4897]: I1121 14:08:46.010956 4897 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-26 10:40:24.628517775 +0000 UTC Nov 21 14:08:46 crc kubenswrapper[4897]: I1121 14:08:46.010994 4897 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 116h31m38.61752748s for next certificate rotation Nov 21 14:08:46 crc kubenswrapper[4897]: I1121 14:08:46.102947 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"ca61ced18b5737bdd0f0f711d97c0b06ab34d5fc2412db2f11a3375e0baf5740"} Nov 21 14:08:46 crc kubenswrapper[4897]: I1121 14:08:46.104887 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"c256823f2fb49ecc121f6ef564cd86196f9b4a473e9bdea670173596ef14a823"} Nov 21 14:08:46 crc kubenswrapper[4897]: I1121 14:08:46.106182 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"e0fc71bebc463919826d01260f38b63bfdb01199fd8dd6ae90d9ac9cd75db703"} Nov 21 14:08:46 crc kubenswrapper[4897]: I1121 14:08:46.107306 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"101fb592fef069af009c613a1e9c9825024ed205498ab72334211ba90d6f7397"} Nov 21 14:08:46 crc kubenswrapper[4897]: I1121 14:08:46.108563 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"6eb2ac410b49098e5b3603de5e6bf5ed4484fab0b300021d0bed2e5ed8a42527"} Nov 21 14:08:47 crc kubenswrapper[4897]: I1121 14:08:47.007585 4897 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.233:6443: connect: connection refused Nov 21 14:08:47 crc kubenswrapper[4897]: E1121 14:08:47.018688 4897 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.233:6443: connect: connection refused" interval="3.2s" Nov 21 14:08:47 crc kubenswrapper[4897]: I1121 14:08:47.115652 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"d1dbe88e7e6ad63899e107ebb884a16edb9a42b21ae7c86d7fbfaf89785b0d36"} Nov 21 14:08:47 crc kubenswrapper[4897]: I1121 14:08:47.115736 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"c65f71f8746044f5d41687c670a0308cd09dc41a6531f9b5720e852b3d240756"} Nov 21 14:08:47 crc kubenswrapper[4897]: I1121 14:08:47.115768 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"ce3a2602338a1c1e9380ee55f78360f3d9cf6ccc62678886dc5ae574a72e3d9e"} Nov 21 14:08:47 crc kubenswrapper[4897]: I1121 14:08:47.115784 4897 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 21 14:08:47 crc kubenswrapper[4897]: I1121 14:08:47.118035 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:08:47 crc kubenswrapper[4897]: I1121 14:08:47.118091 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:08:47 crc kubenswrapper[4897]: I1121 14:08:47.118113 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:08:47 crc kubenswrapper[4897]: I1121 14:08:47.118179 4897 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="e0fc71bebc463919826d01260f38b63bfdb01199fd8dd6ae90d9ac9cd75db703" exitCode=0 Nov 21 14:08:47 crc kubenswrapper[4897]: I1121 14:08:47.118283 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"e0fc71bebc463919826d01260f38b63bfdb01199fd8dd6ae90d9ac9cd75db703"} Nov 21 14:08:47 crc kubenswrapper[4897]: I1121 14:08:47.118413 4897 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 21 14:08:47 crc kubenswrapper[4897]: I1121 14:08:47.119766 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:08:47 crc kubenswrapper[4897]: I1121 14:08:47.119826 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:08:47 crc kubenswrapper[4897]: I1121 14:08:47.119986 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:08:47 crc kubenswrapper[4897]: I1121 14:08:47.121425 4897 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="101fb592fef069af009c613a1e9c9825024ed205498ab72334211ba90d6f7397" exitCode=0 Nov 21 14:08:47 crc kubenswrapper[4897]: I1121 14:08:47.121559 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"101fb592fef069af009c613a1e9c9825024ed205498ab72334211ba90d6f7397"} Nov 21 14:08:47 crc kubenswrapper[4897]: I1121 14:08:47.121587 4897 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 21 14:08:47 crc kubenswrapper[4897]: I1121 14:08:47.122939 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:08:47 crc kubenswrapper[4897]: I1121 14:08:47.122992 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:08:47 crc kubenswrapper[4897]: I1121 14:08:47.123017 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:08:47 crc kubenswrapper[4897]: I1121 14:08:47.124837 4897 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="6eb2ac410b49098e5b3603de5e6bf5ed4484fab0b300021d0bed2e5ed8a42527" exitCode=0 Nov 21 14:08:47 crc kubenswrapper[4897]: I1121 14:08:47.124903 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"6eb2ac410b49098e5b3603de5e6bf5ed4484fab0b300021d0bed2e5ed8a42527"} Nov 21 14:08:47 crc kubenswrapper[4897]: I1121 14:08:47.125018 4897 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 21 14:08:47 crc kubenswrapper[4897]: I1121 14:08:47.125924 4897 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 21 14:08:47 crc kubenswrapper[4897]: I1121 14:08:47.126999 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:08:47 crc kubenswrapper[4897]: I1121 14:08:47.127045 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:08:47 crc kubenswrapper[4897]: I1121 14:08:47.127061 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:08:47 crc kubenswrapper[4897]: I1121 14:08:47.127005 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:08:47 crc kubenswrapper[4897]: I1121 14:08:47.127119 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:08:47 crc kubenswrapper[4897]: I1121 14:08:47.127143 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:08:47 crc kubenswrapper[4897]: I1121 14:08:47.127455 4897 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="ca61ced18b5737bdd0f0f711d97c0b06ab34d5fc2412db2f11a3375e0baf5740" exitCode=0 Nov 21 14:08:47 crc kubenswrapper[4897]: I1121 14:08:47.127550 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"ca61ced18b5737bdd0f0f711d97c0b06ab34d5fc2412db2f11a3375e0baf5740"} Nov 21 14:08:47 crc kubenswrapper[4897]: I1121 14:08:47.127647 4897 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 21 14:08:47 crc kubenswrapper[4897]: I1121 14:08:47.128342 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:08:47 crc kubenswrapper[4897]: I1121 14:08:47.128390 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:08:47 crc kubenswrapper[4897]: I1121 14:08:47.128407 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:08:47 crc kubenswrapper[4897]: W1121 14:08:47.301040 4897 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.233:6443: connect: connection refused Nov 21 14:08:47 crc kubenswrapper[4897]: E1121 14:08:47.301133 4897 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.233:6443: connect: connection refused" logger="UnhandledError" Nov 21 14:08:47 crc kubenswrapper[4897]: I1121 14:08:47.327898 4897 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 21 14:08:47 crc kubenswrapper[4897]: I1121 14:08:47.400896 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:08:47 crc kubenswrapper[4897]: I1121 14:08:47.400964 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:08:47 crc kubenswrapper[4897]: I1121 14:08:47.401106 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:08:47 crc kubenswrapper[4897]: I1121 14:08:47.401141 4897 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 21 14:08:47 crc kubenswrapper[4897]: E1121 14:08:47.401817 4897 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.233:6443: connect: connection refused" node="crc" Nov 21 14:08:47 crc kubenswrapper[4897]: I1121 14:08:47.611931 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 21 14:08:48 crc kubenswrapper[4897]: I1121 14:08:48.007048 4897 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.233:6443: connect: connection refused Nov 21 14:08:48 crc kubenswrapper[4897]: I1121 14:08:48.131298 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"1091bce69830e7805c81aea0fcabd87cb16fd9feef5becdd44d9d2baba613437"} Nov 21 14:08:48 crc kubenswrapper[4897]: I1121 14:08:48.131350 4897 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 21 14:08:48 crc kubenswrapper[4897]: I1121 14:08:48.132314 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:08:48 crc kubenswrapper[4897]: I1121 14:08:48.132348 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:08:48 crc kubenswrapper[4897]: I1121 14:08:48.132357 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:08:48 crc kubenswrapper[4897]: I1121 14:08:48.134339 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"419ab9403696491ef53a3d812d4c19d676f8fb1a1a4dfcdfcc0c6255377d2918"} Nov 21 14:08:48 crc kubenswrapper[4897]: I1121 14:08:48.134369 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"08651f8381818947818e0bb640c089be4a07022efc650befbb65ad58466bf8c3"} Nov 21 14:08:48 crc kubenswrapper[4897]: I1121 14:08:48.136752 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"de079956e3d8de0101f3f87dcc120ef23876c4972382639e419b3f34a38cdfca"} Nov 21 14:08:48 crc kubenswrapper[4897]: I1121 14:08:48.136775 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"737553dd1d16e2019ea89a376e3a868f9792ed5239aa7d6cc40a05d3d4f286ae"} Nov 21 14:08:48 crc kubenswrapper[4897]: I1121 14:08:48.138287 4897 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="03385e0c78e4806f60c160227ef4905a11003ed3d70f0e5c0dd0a477979d1d4b" exitCode=0 Nov 21 14:08:48 crc kubenswrapper[4897]: I1121 14:08:48.138333 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"03385e0c78e4806f60c160227ef4905a11003ed3d70f0e5c0dd0a477979d1d4b"} Nov 21 14:08:48 crc kubenswrapper[4897]: I1121 14:08:48.138376 4897 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 21 14:08:48 crc kubenswrapper[4897]: I1121 14:08:48.138442 4897 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 21 14:08:48 crc kubenswrapper[4897]: I1121 14:08:48.139175 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:08:48 crc kubenswrapper[4897]: I1121 14:08:48.139212 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:08:48 crc kubenswrapper[4897]: I1121 14:08:48.139224 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:08:48 crc kubenswrapper[4897]: I1121 14:08:48.139453 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:08:48 crc kubenswrapper[4897]: I1121 14:08:48.139477 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:08:48 crc kubenswrapper[4897]: I1121 14:08:48.139486 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:08:48 crc kubenswrapper[4897]: W1121 14:08:48.227643 4897 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.233:6443: connect: connection refused Nov 21 14:08:48 crc kubenswrapper[4897]: E1121 14:08:48.227750 4897 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.233:6443: connect: connection refused" logger="UnhandledError" Nov 21 14:08:48 crc kubenswrapper[4897]: W1121 14:08:48.416385 4897 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.233:6443: connect: connection refused Nov 21 14:08:48 crc kubenswrapper[4897]: E1121 14:08:48.416487 4897 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.233:6443: connect: connection refused" logger="UnhandledError" Nov 21 14:08:48 crc kubenswrapper[4897]: W1121 14:08:48.456562 4897 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.233:6443: connect: connection refused Nov 21 14:08:48 crc kubenswrapper[4897]: E1121 14:08:48.456655 4897 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.233:6443: connect: connection refused" logger="UnhandledError" Nov 21 14:08:48 crc kubenswrapper[4897]: I1121 14:08:48.737026 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 21 14:08:49 crc kubenswrapper[4897]: I1121 14:08:49.007298 4897 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.233:6443: connect: connection refused Nov 21 14:08:49 crc kubenswrapper[4897]: I1121 14:08:49.144480 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"d78c87c440c7d8e30075f6143111b3f504fb99686484b8f19a9c8d61a55eccc2"} Nov 21 14:08:49 crc kubenswrapper[4897]: I1121 14:08:49.144558 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"6682a8886d2b9bce069cf9261d4701ef6907f19574fd035bd89bd05626023e5c"} Nov 21 14:08:49 crc kubenswrapper[4897]: I1121 14:08:49.147187 4897 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="50aea2fa8b3b141b4330ddce2c8e0715230c6f06dab64b3abfd98a3067224bec" exitCode=0 Nov 21 14:08:49 crc kubenswrapper[4897]: I1121 14:08:49.147293 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"50aea2fa8b3b141b4330ddce2c8e0715230c6f06dab64b3abfd98a3067224bec"} Nov 21 14:08:49 crc kubenswrapper[4897]: I1121 14:08:49.147323 4897 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 21 14:08:49 crc kubenswrapper[4897]: I1121 14:08:49.149404 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:08:49 crc kubenswrapper[4897]: I1121 14:08:49.149438 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:08:49 crc kubenswrapper[4897]: I1121 14:08:49.149450 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:08:49 crc kubenswrapper[4897]: I1121 14:08:49.152255 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"acc9629bb6b0dfbe5b4efe2f9f980e4e3865fdc6742782ebc49a5a7e0f775b9d"} Nov 21 14:08:49 crc kubenswrapper[4897]: I1121 14:08:49.152378 4897 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 21 14:08:49 crc kubenswrapper[4897]: I1121 14:08:49.152405 4897 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 21 14:08:49 crc kubenswrapper[4897]: I1121 14:08:49.153052 4897 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 21 14:08:49 crc kubenswrapper[4897]: I1121 14:08:49.153339 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:08:49 crc kubenswrapper[4897]: I1121 14:08:49.153361 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:08:49 crc kubenswrapper[4897]: I1121 14:08:49.153387 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:08:49 crc kubenswrapper[4897]: I1121 14:08:49.153399 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:08:49 crc kubenswrapper[4897]: I1121 14:08:49.153367 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:08:49 crc kubenswrapper[4897]: I1121 14:08:49.153466 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:08:49 crc kubenswrapper[4897]: I1121 14:08:49.154574 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:08:49 crc kubenswrapper[4897]: I1121 14:08:49.154610 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:08:49 crc kubenswrapper[4897]: I1121 14:08:49.154623 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:08:49 crc kubenswrapper[4897]: I1121 14:08:49.323415 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 21 14:08:50 crc kubenswrapper[4897]: I1121 14:08:50.007244 4897 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.233:6443: connect: connection refused Nov 21 14:08:50 crc kubenswrapper[4897]: I1121 14:08:50.157938 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"1bd64b9b938bc80284bcee54141e90cad613d339c4f27ef2934b29bc2a318916"} Nov 21 14:08:50 crc kubenswrapper[4897]: I1121 14:08:50.157992 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"66c521c26d541bbddb064dd9710f1a7b41090d2f47343977922272dfc6dab7af"} Nov 21 14:08:50 crc kubenswrapper[4897]: I1121 14:08:50.161043 4897 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 21 14:08:50 crc kubenswrapper[4897]: I1121 14:08:50.161707 4897 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 21 14:08:50 crc kubenswrapper[4897]: I1121 14:08:50.161809 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"f0caf48d11ff2e3804c80431df8c7c88e081dcde5f0f357aae1f09765bfa2b18"} Nov 21 14:08:50 crc kubenswrapper[4897]: I1121 14:08:50.161920 4897 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 21 14:08:50 crc kubenswrapper[4897]: I1121 14:08:50.162306 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:08:50 crc kubenswrapper[4897]: I1121 14:08:50.162341 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:08:50 crc kubenswrapper[4897]: I1121 14:08:50.162354 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:08:50 crc kubenswrapper[4897]: I1121 14:08:50.162467 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:08:50 crc kubenswrapper[4897]: I1121 14:08:50.162494 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:08:50 crc kubenswrapper[4897]: I1121 14:08:50.162558 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:08:50 crc kubenswrapper[4897]: I1121 14:08:50.162657 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:08:50 crc kubenswrapper[4897]: I1121 14:08:50.162675 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:08:50 crc kubenswrapper[4897]: I1121 14:08:50.162684 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:08:50 crc kubenswrapper[4897]: E1121 14:08:50.219675 4897 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.233:6443: connect: connection refused" interval="6.4s" Nov 21 14:08:50 crc kubenswrapper[4897]: I1121 14:08:50.602673 4897 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 21 14:08:50 crc kubenswrapper[4897]: I1121 14:08:50.604362 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:08:50 crc kubenswrapper[4897]: I1121 14:08:50.604435 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:08:50 crc kubenswrapper[4897]: I1121 14:08:50.604453 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:08:50 crc kubenswrapper[4897]: I1121 14:08:50.604491 4897 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 21 14:08:50 crc kubenswrapper[4897]: E1121 14:08:50.605264 4897 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.233:6443: connect: connection refused" node="crc" Nov 21 14:08:51 crc kubenswrapper[4897]: I1121 14:08:51.007293 4897 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.233:6443: connect: connection refused Nov 21 14:08:51 crc kubenswrapper[4897]: I1121 14:08:51.172328 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"ca74b28a069dc2e609b0779164c3aa470a15ed9848af0dd5fc3479aaf750157b"} Nov 21 14:08:51 crc kubenswrapper[4897]: I1121 14:08:51.172395 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"936f59996d26e42099dcd055f87fff9e3ac4283b57e1956fe97463f2334152d7"} Nov 21 14:08:51 crc kubenswrapper[4897]: I1121 14:08:51.172399 4897 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 21 14:08:51 crc kubenswrapper[4897]: I1121 14:08:51.172548 4897 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 21 14:08:51 crc kubenswrapper[4897]: I1121 14:08:51.172894 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 21 14:08:51 crc kubenswrapper[4897]: I1121 14:08:51.173066 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 21 14:08:51 crc kubenswrapper[4897]: I1121 14:08:51.173221 4897 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="Get \"https://192.168.126.11:6443/livez\": dial tcp 192.168.126.11:6443: connect: connection refused" start-of-body= Nov 21 14:08:51 crc kubenswrapper[4897]: I1121 14:08:51.173263 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:08:51 crc kubenswrapper[4897]: I1121 14:08:51.173290 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:08:51 crc kubenswrapper[4897]: I1121 14:08:51.173255 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="Get \"https://192.168.126.11:6443/livez\": dial tcp 192.168.126.11:6443: connect: connection refused" Nov 21 14:08:51 crc kubenswrapper[4897]: I1121 14:08:51.173301 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:08:51 crc kubenswrapper[4897]: I1121 14:08:51.173428 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:08:51 crc kubenswrapper[4897]: I1121 14:08:51.173442 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:08:51 crc kubenswrapper[4897]: I1121 14:08:51.173449 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:08:51 crc kubenswrapper[4897]: W1121 14:08:51.732691 4897 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.233:6443: connect: connection refused Nov 21 14:08:51 crc kubenswrapper[4897]: E1121 14:08:51.732812 4897 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.233:6443: connect: connection refused" logger="UnhandledError" Nov 21 14:08:52 crc kubenswrapper[4897]: I1121 14:08:52.006588 4897 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.233:6443: connect: connection refused Nov 21 14:08:52 crc kubenswrapper[4897]: I1121 14:08:52.179198 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"270c53b9a342f26be077a01cd4e87e9ccd0eb5d2c84208ae38cbf720463245e5"} Nov 21 14:08:52 crc kubenswrapper[4897]: I1121 14:08:52.179282 4897 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 21 14:08:52 crc kubenswrapper[4897]: I1121 14:08:52.180565 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:08:52 crc kubenswrapper[4897]: I1121 14:08:52.180612 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:08:52 crc kubenswrapper[4897]: I1121 14:08:52.180630 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:08:52 crc kubenswrapper[4897]: I1121 14:08:52.180915 4897 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 21 14:08:52 crc kubenswrapper[4897]: I1121 14:08:52.182489 4897 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="f0caf48d11ff2e3804c80431df8c7c88e081dcde5f0f357aae1f09765bfa2b18" exitCode=255 Nov 21 14:08:52 crc kubenswrapper[4897]: I1121 14:08:52.182567 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"f0caf48d11ff2e3804c80431df8c7c88e081dcde5f0f357aae1f09765bfa2b18"} Nov 21 14:08:52 crc kubenswrapper[4897]: I1121 14:08:52.182580 4897 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 21 14:08:52 crc kubenswrapper[4897]: I1121 14:08:52.183568 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:08:52 crc kubenswrapper[4897]: I1121 14:08:52.183610 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:08:52 crc kubenswrapper[4897]: I1121 14:08:52.183628 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:08:52 crc kubenswrapper[4897]: I1121 14:08:52.184212 4897 scope.go:117] "RemoveContainer" containerID="f0caf48d11ff2e3804c80431df8c7c88e081dcde5f0f357aae1f09765bfa2b18" Nov 21 14:08:52 crc kubenswrapper[4897]: I1121 14:08:52.984919 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 21 14:08:53 crc kubenswrapper[4897]: I1121 14:08:53.110661 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 21 14:08:53 crc kubenswrapper[4897]: I1121 14:08:53.110829 4897 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 21 14:08:53 crc kubenswrapper[4897]: I1121 14:08:53.111802 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:08:53 crc kubenswrapper[4897]: I1121 14:08:53.111834 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:08:53 crc kubenswrapper[4897]: I1121 14:08:53.111842 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:08:53 crc kubenswrapper[4897]: I1121 14:08:53.187652 4897 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 21 14:08:53 crc kubenswrapper[4897]: I1121 14:08:53.189426 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"1907bfbad7eb34da642e30d6aa81820aab457f92347333a364ac12396b0e9dae"} Nov 21 14:08:53 crc kubenswrapper[4897]: I1121 14:08:53.189486 4897 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 21 14:08:53 crc kubenswrapper[4897]: I1121 14:08:53.189580 4897 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 21 14:08:53 crc kubenswrapper[4897]: I1121 14:08:53.190818 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:08:53 crc kubenswrapper[4897]: I1121 14:08:53.190818 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:08:53 crc kubenswrapper[4897]: I1121 14:08:53.190899 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:08:53 crc kubenswrapper[4897]: I1121 14:08:53.190867 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:08:53 crc kubenswrapper[4897]: I1121 14:08:53.190918 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:08:53 crc kubenswrapper[4897]: I1121 14:08:53.190927 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:08:54 crc kubenswrapper[4897]: I1121 14:08:54.192467 4897 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 21 14:08:54 crc kubenswrapper[4897]: I1121 14:08:54.192576 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 21 14:08:54 crc kubenswrapper[4897]: I1121 14:08:54.193644 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:08:54 crc kubenswrapper[4897]: I1121 14:08:54.193677 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:08:54 crc kubenswrapper[4897]: I1121 14:08:54.193688 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:08:54 crc kubenswrapper[4897]: E1121 14:08:54.219948 4897 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 21 14:08:55 crc kubenswrapper[4897]: I1121 14:08:55.197812 4897 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 21 14:08:55 crc kubenswrapper[4897]: I1121 14:08:55.199169 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:08:55 crc kubenswrapper[4897]: I1121 14:08:55.199220 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:08:55 crc kubenswrapper[4897]: I1121 14:08:55.199234 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:08:55 crc kubenswrapper[4897]: I1121 14:08:55.371977 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Nov 21 14:08:55 crc kubenswrapper[4897]: I1121 14:08:55.372203 4897 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 21 14:08:55 crc kubenswrapper[4897]: I1121 14:08:55.373406 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:08:55 crc kubenswrapper[4897]: I1121 14:08:55.373447 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:08:55 crc kubenswrapper[4897]: I1121 14:08:55.373459 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:08:56 crc kubenswrapper[4897]: I1121 14:08:56.110746 4897 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 21 14:08:56 crc kubenswrapper[4897]: I1121 14:08:56.110853 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 21 14:08:56 crc kubenswrapper[4897]: I1121 14:08:56.236129 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 21 14:08:56 crc kubenswrapper[4897]: I1121 14:08:56.236776 4897 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 21 14:08:56 crc kubenswrapper[4897]: I1121 14:08:56.237889 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:08:56 crc kubenswrapper[4897]: I1121 14:08:56.237919 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:08:56 crc kubenswrapper[4897]: I1121 14:08:56.237930 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:08:57 crc kubenswrapper[4897]: I1121 14:08:57.005678 4897 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 21 14:08:57 crc kubenswrapper[4897]: I1121 14:08:57.007026 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:08:57 crc kubenswrapper[4897]: I1121 14:08:57.007093 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:08:57 crc kubenswrapper[4897]: I1121 14:08:57.007106 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:08:57 crc kubenswrapper[4897]: I1121 14:08:57.007160 4897 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 21 14:08:57 crc kubenswrapper[4897]: I1121 14:08:57.242654 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 21 14:08:57 crc kubenswrapper[4897]: I1121 14:08:57.242864 4897 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 21 14:08:57 crc kubenswrapper[4897]: I1121 14:08:57.244460 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:08:57 crc kubenswrapper[4897]: I1121 14:08:57.244569 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:08:57 crc kubenswrapper[4897]: I1121 14:08:57.244586 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:08:57 crc kubenswrapper[4897]: I1121 14:08:57.249546 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 21 14:08:58 crc kubenswrapper[4897]: I1121 14:08:58.205257 4897 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 21 14:08:58 crc kubenswrapper[4897]: I1121 14:08:58.206259 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:08:58 crc kubenswrapper[4897]: I1121 14:08:58.206314 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:08:58 crc kubenswrapper[4897]: I1121 14:08:58.206330 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:08:59 crc kubenswrapper[4897]: I1121 14:08:59.140345 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Nov 21 14:08:59 crc kubenswrapper[4897]: I1121 14:08:59.140661 4897 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 21 14:08:59 crc kubenswrapper[4897]: I1121 14:08:59.141883 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:08:59 crc kubenswrapper[4897]: I1121 14:08:59.141923 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:08:59 crc kubenswrapper[4897]: I1121 14:08:59.141935 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:00 crc kubenswrapper[4897]: I1121 14:09:00.140385 4897 patch_prober.go:28] interesting pod/etcd-crc container/etcd namespace/openshift-etcd: Startup probe status=failure output="Get \"https://192.168.126.11:9980/readyz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 21 14:09:00 crc kubenswrapper[4897]: I1121 14:09:00.140499 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-etcd/etcd-crc" podUID="2139d3e2895fc6797b9c76a1b4c9886d" containerName="etcd" probeResult="failure" output="Get \"https://192.168.126.11:9980/readyz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 21 14:09:02 crc kubenswrapper[4897]: I1121 14:09:02.988320 4897 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Nov 21 14:09:02 crc kubenswrapper[4897]: I1121 14:09:02.988391 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Nov 21 14:09:02 crc kubenswrapper[4897]: I1121 14:09:02.992451 4897 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\": RBAC: [clusterrole.rbac.authorization.k8s.io \"system:openshift:public-info-viewer\" not found, clusterrole.rbac.authorization.k8s.io \"system:public-info-viewer\" not found]","reason":"Forbidden","details":{},"code":403} Nov 21 14:09:02 crc kubenswrapper[4897]: I1121 14:09:02.992575 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Nov 21 14:09:04 crc kubenswrapper[4897]: E1121 14:09:04.220123 4897 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 21 14:09:06 crc kubenswrapper[4897]: I1121 14:09:06.110750 4897 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 21 14:09:06 crc kubenswrapper[4897]: I1121 14:09:06.110888 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Nov 21 14:09:06 crc kubenswrapper[4897]: I1121 14:09:06.180727 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 21 14:09:06 crc kubenswrapper[4897]: I1121 14:09:06.180940 4897 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 21 14:09:06 crc kubenswrapper[4897]: I1121 14:09:06.181255 4897 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Nov 21 14:09:06 crc kubenswrapper[4897]: I1121 14:09:06.181306 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Nov 21 14:09:06 crc kubenswrapper[4897]: I1121 14:09:06.182129 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:06 crc kubenswrapper[4897]: I1121 14:09:06.182156 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:06 crc kubenswrapper[4897]: I1121 14:09:06.182166 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:06 crc kubenswrapper[4897]: I1121 14:09:06.188647 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 21 14:09:06 crc kubenswrapper[4897]: I1121 14:09:06.225037 4897 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 21 14:09:06 crc kubenswrapper[4897]: I1121 14:09:06.225425 4897 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Nov 21 14:09:06 crc kubenswrapper[4897]: I1121 14:09:06.225484 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Nov 21 14:09:06 crc kubenswrapper[4897]: I1121 14:09:06.226292 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:06 crc kubenswrapper[4897]: I1121 14:09:06.226350 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:06 crc kubenswrapper[4897]: I1121 14:09:06.226364 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:07 crc kubenswrapper[4897]: E1121 14:09:07.979876 4897 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="7s" Nov 21 14:09:07 crc kubenswrapper[4897]: I1121 14:09:07.982563 4897 trace.go:236] Trace[1627053337]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (21-Nov-2025 14:08:53.102) (total time: 14879ms): Nov 21 14:09:07 crc kubenswrapper[4897]: Trace[1627053337]: ---"Objects listed" error: 14879ms (14:09:07.982) Nov 21 14:09:07 crc kubenswrapper[4897]: Trace[1627053337]: [14.879802657s] [14.879802657s] END Nov 21 14:09:07 crc kubenswrapper[4897]: I1121 14:09:07.982592 4897 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Nov 21 14:09:07 crc kubenswrapper[4897]: I1121 14:09:07.983731 4897 trace.go:236] Trace[2118818051]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (21-Nov-2025 14:08:54.308) (total time: 13674ms): Nov 21 14:09:07 crc kubenswrapper[4897]: Trace[2118818051]: ---"Objects listed" error: 13674ms (14:09:07.983) Nov 21 14:09:07 crc kubenswrapper[4897]: Trace[2118818051]: [13.674683732s] [13.674683732s] END Nov 21 14:09:07 crc kubenswrapper[4897]: I1121 14:09:07.983770 4897 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Nov 21 14:09:07 crc kubenswrapper[4897]: E1121 14:09:07.984164 4897 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Nov 21 14:09:07 crc kubenswrapper[4897]: I1121 14:09:07.984266 4897 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Nov 21 14:09:07 crc kubenswrapper[4897]: I1121 14:09:07.986435 4897 trace.go:236] Trace[424812421]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (21-Nov-2025 14:08:54.766) (total time: 13219ms): Nov 21 14:09:07 crc kubenswrapper[4897]: Trace[424812421]: ---"Objects listed" error: 13219ms (14:09:07.986) Nov 21 14:09:07 crc kubenswrapper[4897]: Trace[424812421]: [13.219950116s] [13.219950116s] END Nov 21 14:09:07 crc kubenswrapper[4897]: I1121 14:09:07.986470 4897 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Nov 21 14:09:08 crc kubenswrapper[4897]: I1121 14:09:08.233213 4897 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Nov 21 14:09:08 crc kubenswrapper[4897]: I1121 14:09:08.233600 4897 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 21 14:09:08 crc kubenswrapper[4897]: I1121 14:09:08.235219 4897 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="1907bfbad7eb34da642e30d6aa81820aab457f92347333a364ac12396b0e9dae" exitCode=255 Nov 21 14:09:08 crc kubenswrapper[4897]: I1121 14:09:08.235269 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"1907bfbad7eb34da642e30d6aa81820aab457f92347333a364ac12396b0e9dae"} Nov 21 14:09:08 crc kubenswrapper[4897]: I1121 14:09:08.235340 4897 scope.go:117] "RemoveContainer" containerID="f0caf48d11ff2e3804c80431df8c7c88e081dcde5f0f357aae1f09765bfa2b18" Nov 21 14:09:08 crc kubenswrapper[4897]: I1121 14:09:08.235583 4897 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 21 14:09:08 crc kubenswrapper[4897]: I1121 14:09:08.236471 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:08 crc kubenswrapper[4897]: I1121 14:09:08.236527 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:08 crc kubenswrapper[4897]: I1121 14:09:08.236543 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:08 crc kubenswrapper[4897]: I1121 14:09:08.237361 4897 scope.go:117] "RemoveContainer" containerID="1907bfbad7eb34da642e30d6aa81820aab457f92347333a364ac12396b0e9dae" Nov 21 14:09:08 crc kubenswrapper[4897]: E1121 14:09:08.237625 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Nov 21 14:09:08 crc kubenswrapper[4897]: I1121 14:09:08.703923 4897 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Nov 21 14:09:08 crc kubenswrapper[4897]: I1121 14:09:08.993756 4897 apiserver.go:52] "Watching apiserver" Nov 21 14:09:08 crc kubenswrapper[4897]: I1121 14:09:08.998444 4897 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Nov 21 14:09:08 crc kubenswrapper[4897]: I1121 14:09:08.998733 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h"] Nov 21 14:09:08 crc kubenswrapper[4897]: I1121 14:09:08.999201 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 21 14:09:08 crc kubenswrapper[4897]: I1121 14:09:08.999286 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 21 14:09:08 crc kubenswrapper[4897]: I1121 14:09:08.999364 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 21 14:09:08 crc kubenswrapper[4897]: E1121 14:09:08.999408 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 21 14:09:08 crc kubenswrapper[4897]: I1121 14:09:08.999464 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 21 14:09:08 crc kubenswrapper[4897]: E1121 14:09:08.999529 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.000183 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.000557 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 21 14:09:09 crc kubenswrapper[4897]: E1121 14:09:09.000616 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.003665 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.003758 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.003873 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.004096 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.004302 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.004333 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.004609 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.005065 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.006201 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.010692 4897 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.032222 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.048969 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.063934 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.077350 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.089852 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.089980 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.090045 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.090080 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.090104 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.090126 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.090149 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.090171 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.090196 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.090218 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.090240 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.090265 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.090286 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.090306 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.090325 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.090347 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.090370 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.090390 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.090410 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.090435 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.090455 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.090474 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.090533 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.090561 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.090591 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.090778 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.090889 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.090900 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.091122 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.091139 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.091269 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.091419 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.091582 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.091590 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.091711 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.091752 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.091789 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.091260 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.091819 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.091913 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.091955 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.092019 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.092048 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.092771 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.092805 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.091913 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.092016 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.092257 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.092861 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.092330 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.092432 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.092483 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.092666 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.092927 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.092956 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.092979 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.093011 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.093032 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.093056 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.093078 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.093098 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.093121 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.093144 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.093168 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.093195 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.093220 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.093247 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.093274 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.093299 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.093315 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.093334 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.093353 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.093372 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.093389 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.093437 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.093459 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.093477 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.093494 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.093529 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.093551 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.093568 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.093587 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.092753 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.092761 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.093036 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.093132 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.093148 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.093184 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.092475 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.093605 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.093716 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.093756 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.093794 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.093793 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.093782 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.093835 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.093863 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.093917 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.093920 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.093933 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.093945 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.093974 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.094129 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.094167 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.094197 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.094174 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.094269 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.094366 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.095554 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.095569 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.095745 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.095836 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.095856 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.096231 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.096360 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.096422 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.096643 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.096880 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.096885 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.096913 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.097028 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.097175 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.097282 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.097311 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.097331 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.097324 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.097349 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.097356 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.097253 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.097771 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.098142 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.098175 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.098338 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.097362 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.098308 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.098480 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.098608 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.098772 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.098838 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.098906 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.098969 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.099380 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.099537 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.099440 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.099702 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.099716 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.099761 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.100102 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.100304 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.100413 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.100389 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.100817 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.100873 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.100918 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.100957 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.100988 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.101021 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.101070 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.101048 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.101120 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.101283 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.101301 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.101378 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.101449 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.101490 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.101605 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.101730 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.101736 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.101799 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.101827 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.101851 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.101863 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.101882 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.101904 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: E1121 14:09:09.102150 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-21 14:09:09.602003409 +0000 UTC m=+26.886596884 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.102296 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.102658 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.102757 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.103646 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.103655 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.103843 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.104126 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.102807 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.105417 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.105452 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.105487 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.105581 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.105608 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.105633 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.105657 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.105684 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.105711 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.105742 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.105769 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.105808 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.105835 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.105863 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.105891 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.105917 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.105940 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.105945 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.105987 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.105991 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.106010 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.106036 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.106059 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.106081 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.106138 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.106163 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.106196 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.106229 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.106259 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.106285 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.106253 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.106312 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.106352 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.106363 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.106439 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.106468 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.106523 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.106585 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.106613 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.106611 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.106665 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.107590 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.106672 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.106950 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.106956 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.107038 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.107054 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.107189 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.107304 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.107322 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.107337 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.103616 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.107885 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.107908 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.108173 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.108182 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.108314 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.108589 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.106637 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.108892 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.108931 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.108965 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.109004 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.109038 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.109071 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.109102 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.109131 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.109163 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.109194 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.109222 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.109251 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.109278 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.109305 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.109333 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.109361 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.109388 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.109446 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.109478 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.109536 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.109564 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.109595 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.109621 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.109650 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.109682 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.109714 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.109959 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.110019 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.110058 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.110118 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.110147 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.110203 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.110236 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.110298 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.110567 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.110640 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.110667 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.110729 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.110837 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.110873 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.110902 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.110938 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.111002 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.111032 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.111097 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.111129 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.111182 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.111209 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.111260 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.111290 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.111343 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.111371 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.111438 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.111524 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.111575 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.111610 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.111646 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.111679 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.111712 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.111740 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.111769 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.113748 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.108928 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.108541 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.113844 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.108955 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.108974 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.113894 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.113942 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.113990 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.114029 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.114279 4897 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.114308 4897 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.114335 4897 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.114354 4897 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.114372 4897 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.114389 4897 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.114405 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.114424 4897 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.114439 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.114455 4897 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.114469 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.114484 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.114517 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.114533 4897 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.114547 4897 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.114560 4897 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.114572 4897 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.114926 4897 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.114950 4897 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.114968 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.114986 4897 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.115000 4897 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.115015 4897 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.115082 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.115100 4897 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.115115 4897 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.115130 4897 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.115145 4897 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.115160 4897 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.115174 4897 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.115188 4897 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.115204 4897 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.115219 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.115234 4897 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.115286 4897 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.115301 4897 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.115317 4897 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.115331 4897 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.115345 4897 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.115359 4897 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.115373 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.115387 4897 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.115404 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.115419 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.115433 4897 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.115456 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.115470 4897 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.115485 4897 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.115529 4897 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.115544 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.115560 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.115576 4897 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.115591 4897 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.115605 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.115623 4897 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.115638 4897 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.115653 4897 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.115667 4897 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.115681 4897 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.115695 4897 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.115714 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.115728 4897 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.115741 4897 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.115756 4897 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.115773 4897 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.115787 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.115802 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.115780 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.115816 4897 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.116136 4897 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.116183 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.116209 4897 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.116248 4897 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.116272 4897 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.116289 4897 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.116305 4897 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.116323 4897 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.116351 4897 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.116369 4897 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.116395 4897 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.116415 4897 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.116431 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.116447 4897 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.116461 4897 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.116494 4897 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.116541 4897 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.116556 4897 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.116570 4897 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.116586 4897 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.116599 4897 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.116614 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.116629 4897 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.116644 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.116658 4897 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.116674 4897 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.116690 4897 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.116704 4897 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.116718 4897 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.116733 4897 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.116733 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.116746 4897 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.116763 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.116797 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.109076 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.109275 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.109314 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.109629 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.109777 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.109809 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.110801 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.110921 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.112261 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.112713 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.113101 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.113188 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.113268 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.117258 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.117599 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.117858 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.113310 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.113648 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.114068 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.114156 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.113769 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.114438 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.114558 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.114647 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.114808 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.118185 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.114949 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.115201 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.115424 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: E1121 14:09:09.115733 4897 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.115844 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.115875 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.116153 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.116222 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.116881 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: E1121 14:09:09.116894 4897 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 21 14:09:09 crc kubenswrapper[4897]: E1121 14:09:09.118345 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-21 14:09:09.618317668 +0000 UTC m=+26.902911363 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.118847 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.118876 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: E1121 14:09:09.118970 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-21 14:09:09.618955935 +0000 UTC m=+26.903549610 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.119166 4897 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.120147 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.120214 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.118151 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.120546 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.120675 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.120745 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.121618 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.121657 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.121822 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.121929 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.122012 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.122274 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.122617 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.122661 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.122886 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.123162 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.123928 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.125641 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.125760 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.126117 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.126229 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.126449 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.129161 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.129216 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.129447 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.130018 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.130861 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.130905 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.130124 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.131958 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: E1121 14:09:09.132277 4897 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 21 14:09:09 crc kubenswrapper[4897]: E1121 14:09:09.132402 4897 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 21 14:09:09 crc kubenswrapper[4897]: E1121 14:09:09.132582 4897 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 21 14:09:09 crc kubenswrapper[4897]: E1121 14:09:09.132881 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-21 14:09:09.632863069 +0000 UTC m=+26.917456544 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.137334 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.141283 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.141763 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.142617 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.142841 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.145216 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 21 14:09:09 crc kubenswrapper[4897]: E1121 14:09:09.145596 4897 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 21 14:09:09 crc kubenswrapper[4897]: E1121 14:09:09.145639 4897 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 21 14:09:09 crc kubenswrapper[4897]: E1121 14:09:09.145662 4897 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 21 14:09:09 crc kubenswrapper[4897]: E1121 14:09:09.145734 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-21 14:09:09.645707327 +0000 UTC m=+26.930301002 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.152066 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.152195 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.153612 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.157731 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.158276 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.158260 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.158303 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.159155 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.159411 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.159554 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.162600 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.162797 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.163569 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.163633 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.164269 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.165719 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.167160 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.168683 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.171807 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.177340 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.177564 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.177597 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.177685 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.178534 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.179166 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.179481 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.179677 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.186915 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.187208 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.192615 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.205637 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.209427 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.214245 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.215684 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.219138 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.219453 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.219554 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.219599 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.219612 4897 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.219624 4897 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.219634 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.219633 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.219646 4897 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.219702 4897 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.219719 4897 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.219734 4897 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.219744 4897 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.219756 4897 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.219767 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.219777 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.219796 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.219807 4897 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.219816 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.219825 4897 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.219835 4897 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.219739 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.219844 4897 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.219893 4897 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.219908 4897 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.219925 4897 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.219938 4897 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.219952 4897 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.219962 4897 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.219972 4897 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.219982 4897 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.219992 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.220003 4897 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.220014 4897 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.220024 4897 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.220036 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.220050 4897 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.220059 4897 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.220069 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.220078 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.220088 4897 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.220097 4897 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.220106 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.220116 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.220125 4897 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.220134 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.220143 4897 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.220151 4897 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.220160 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.220169 4897 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.220182 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.220190 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.220198 4897 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.220207 4897 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.220216 4897 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.220225 4897 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.220233 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.220242 4897 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.220253 4897 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.220290 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.220311 4897 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.220346 4897 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.220364 4897 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.220373 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.220383 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.220391 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.220400 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.220410 4897 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.220418 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.220427 4897 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.220436 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.220456 4897 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.220467 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.220480 4897 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.220491 4897 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.220526 4897 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.220538 4897 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.220564 4897 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.220576 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.220587 4897 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.220599 4897 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.220610 4897 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.220621 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.220633 4897 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.220646 4897 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.220664 4897 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.220674 4897 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.220683 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.220692 4897 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.220713 4897 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.220721 4897 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.220730 4897 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.220738 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.220748 4897 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.220757 4897 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.220770 4897 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.220779 4897 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.220787 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.234484 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.239848 4897 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.254841 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.265965 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.276590 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.314976 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.322301 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.331380 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.340555 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.623814 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 21 14:09:09 crc kubenswrapper[4897]: E1121 14:09:09.624046 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-21 14:09:10.624011731 +0000 UTC m=+27.908605336 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.624468 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.624522 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 21 14:09:09 crc kubenswrapper[4897]: E1121 14:09:09.624664 4897 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 21 14:09:09 crc kubenswrapper[4897]: E1121 14:09:09.624735 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-21 14:09:10.62471572 +0000 UTC m=+27.909309366 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 21 14:09:09 crc kubenswrapper[4897]: E1121 14:09:09.624788 4897 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 21 14:09:09 crc kubenswrapper[4897]: E1121 14:09:09.624814 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-21 14:09:10.624806953 +0000 UTC m=+27.909400428 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.725370 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 21 14:09:09 crc kubenswrapper[4897]: I1121 14:09:09.725418 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 21 14:09:09 crc kubenswrapper[4897]: E1121 14:09:09.725574 4897 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 21 14:09:09 crc kubenswrapper[4897]: E1121 14:09:09.725593 4897 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 21 14:09:09 crc kubenswrapper[4897]: E1121 14:09:09.725604 4897 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 21 14:09:09 crc kubenswrapper[4897]: E1121 14:09:09.725650 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-21 14:09:10.725634779 +0000 UTC m=+28.010228254 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 21 14:09:09 crc kubenswrapper[4897]: E1121 14:09:09.726141 4897 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 21 14:09:09 crc kubenswrapper[4897]: E1121 14:09:09.726182 4897 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 21 14:09:09 crc kubenswrapper[4897]: E1121 14:09:09.726198 4897 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 21 14:09:09 crc kubenswrapper[4897]: E1121 14:09:09.726278 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-21 14:09:10.726255825 +0000 UTC m=+28.010849300 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.040848 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-krv5b"] Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.041374 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.041823 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-4qwqj"] Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.042120 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-6ljgx"] Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.042322 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-4qwqj" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.042340 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-6ljgx" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.045316 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-nl76g"] Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.046131 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-nl76g" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.047235 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.047563 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.047753 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.047784 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-tpfb6"] Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.048003 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.048133 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.048276 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.048451 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.048590 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.048739 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.048768 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.048882 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.049379 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.049460 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.049525 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.049739 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.050033 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.050265 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.055568 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.056108 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.056460 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.056468 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.056487 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.056524 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.058816 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.072049 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.085376 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.093195 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.093770 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.094723 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.095332 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.096053 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.096738 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.097496 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.099664 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.100165 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7670227-d280-4847-b882-754429f56b0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-krv5b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.101554 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.102152 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.102711 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.103877 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.104395 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.105363 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.105977 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.106944 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.107665 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.108050 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.109178 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.109798 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.110393 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.111439 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.111488 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.112047 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.113397 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.113884 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.115165 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.115840 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.116356 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.117834 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.118419 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.119706 4897 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.120733 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.123406 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.123803 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.125256 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.125861 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.127978 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.128563 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/5a7d6101-d0d2-4828-b3ce-d01040c04640-cni-binary-copy\") pod \"multus-additional-cni-plugins-nl76g\" (UID: \"5a7d6101-d0d2-4828-b3ce-d01040c04640\") " pod="openshift-multus/multus-additional-cni-plugins-nl76g" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.128702 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/222f65f2-0b82-4760-bcf7-779244720f01-host-run-ovn-kubernetes\") pod \"ovnkube-node-tpfb6\" (UID: \"222f65f2-0b82-4760-bcf7-779244720f01\") " pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.128798 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mmvwn\" (UniqueName: \"kubernetes.io/projected/222f65f2-0b82-4760-bcf7-779244720f01-kube-api-access-mmvwn\") pod \"ovnkube-node-tpfb6\" (UID: \"222f65f2-0b82-4760-bcf7-779244720f01\") " pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.128883 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/5f3b70b5-9294-4f69-9528-500d28f34c89-host-run-netns\") pod \"multus-6ljgx\" (UID: \"5f3b70b5-9294-4f69-9528-500d28f34c89\") " pod="openshift-multus/multus-6ljgx" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.129006 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/222f65f2-0b82-4760-bcf7-779244720f01-etc-openvswitch\") pod \"ovnkube-node-tpfb6\" (UID: \"222f65f2-0b82-4760-bcf7-779244720f01\") " pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.129097 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/f46e2ba8-9ee0-406d-a668-2ba424618b69-hosts-file\") pod \"node-resolver-4qwqj\" (UID: \"f46e2ba8-9ee0-406d-a668-2ba424618b69\") " pod="openshift-dns/node-resolver-4qwqj" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.129179 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4592l\" (UniqueName: \"kubernetes.io/projected/5a7d6101-d0d2-4828-b3ce-d01040c04640-kube-api-access-4592l\") pod \"multus-additional-cni-plugins-nl76g\" (UID: \"5a7d6101-d0d2-4828-b3ce-d01040c04640\") " pod="openshift-multus/multus-additional-cni-plugins-nl76g" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.129274 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e7670227-d280-4847-b882-754429f56b0f-proxy-tls\") pod \"machine-config-daemon-krv5b\" (UID: \"e7670227-d280-4847-b882-754429f56b0f\") " pod="openshift-machine-config-operator/machine-config-daemon-krv5b" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.129131 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.129427 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/5f3b70b5-9294-4f69-9528-500d28f34c89-multus-cni-dir\") pod \"multus-6ljgx\" (UID: \"5f3b70b5-9294-4f69-9528-500d28f34c89\") " pod="openshift-multus/multus-6ljgx" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.129528 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/5a7d6101-d0d2-4828-b3ce-d01040c04640-system-cni-dir\") pod \"multus-additional-cni-plugins-nl76g\" (UID: \"5a7d6101-d0d2-4828-b3ce-d01040c04640\") " pod="openshift-multus/multus-additional-cni-plugins-nl76g" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.129628 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/222f65f2-0b82-4760-bcf7-779244720f01-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-tpfb6\" (UID: \"222f65f2-0b82-4760-bcf7-779244720f01\") " pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.129729 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/5f3b70b5-9294-4f69-9528-500d28f34c89-system-cni-dir\") pod \"multus-6ljgx\" (UID: \"5f3b70b5-9294-4f69-9528-500d28f34c89\") " pod="openshift-multus/multus-6ljgx" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.129808 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/5a7d6101-d0d2-4828-b3ce-d01040c04640-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-nl76g\" (UID: \"5a7d6101-d0d2-4828-b3ce-d01040c04640\") " pod="openshift-multus/multus-additional-cni-plugins-nl76g" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.129888 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/222f65f2-0b82-4760-bcf7-779244720f01-run-systemd\") pod \"ovnkube-node-tpfb6\" (UID: \"222f65f2-0b82-4760-bcf7-779244720f01\") " pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.129973 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/222f65f2-0b82-4760-bcf7-779244720f01-log-socket\") pod \"ovnkube-node-tpfb6\" (UID: \"222f65f2-0b82-4760-bcf7-779244720f01\") " pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.130071 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/5f3b70b5-9294-4f69-9528-500d28f34c89-host-var-lib-cni-bin\") pod \"multus-6ljgx\" (UID: \"5f3b70b5-9294-4f69-9528-500d28f34c89\") " pod="openshift-multus/multus-6ljgx" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.130156 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-78jhb\" (UniqueName: \"kubernetes.io/projected/5f3b70b5-9294-4f69-9528-500d28f34c89-kube-api-access-78jhb\") pod \"multus-6ljgx\" (UID: \"5f3b70b5-9294-4f69-9528-500d28f34c89\") " pod="openshift-multus/multus-6ljgx" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.130233 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/222f65f2-0b82-4760-bcf7-779244720f01-env-overrides\") pod \"ovnkube-node-tpfb6\" (UID: \"222f65f2-0b82-4760-bcf7-779244720f01\") " pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.130311 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/5f3b70b5-9294-4f69-9528-500d28f34c89-multus-socket-dir-parent\") pod \"multus-6ljgx\" (UID: \"5f3b70b5-9294-4f69-9528-500d28f34c89\") " pod="openshift-multus/multus-6ljgx" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.130399 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/5f3b70b5-9294-4f69-9528-500d28f34c89-host-var-lib-kubelet\") pod \"multus-6ljgx\" (UID: \"5f3b70b5-9294-4f69-9528-500d28f34c89\") " pod="openshift-multus/multus-6ljgx" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.130481 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/5f3b70b5-9294-4f69-9528-500d28f34c89-multus-conf-dir\") pod \"multus-6ljgx\" (UID: \"5f3b70b5-9294-4f69-9528-500d28f34c89\") " pod="openshift-multus/multus-6ljgx" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.130614 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/222f65f2-0b82-4760-bcf7-779244720f01-systemd-units\") pod \"ovnkube-node-tpfb6\" (UID: \"222f65f2-0b82-4760-bcf7-779244720f01\") " pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.130702 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/222f65f2-0b82-4760-bcf7-779244720f01-ovn-node-metrics-cert\") pod \"ovnkube-node-tpfb6\" (UID: \"222f65f2-0b82-4760-bcf7-779244720f01\") " pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.130790 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/5f3b70b5-9294-4f69-9528-500d28f34c89-cni-binary-copy\") pod \"multus-6ljgx\" (UID: \"5f3b70b5-9294-4f69-9528-500d28f34c89\") " pod="openshift-multus/multus-6ljgx" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.130867 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/5f3b70b5-9294-4f69-9528-500d28f34c89-host-run-k8s-cni-cncf-io\") pod \"multus-6ljgx\" (UID: \"5f3b70b5-9294-4f69-9528-500d28f34c89\") " pod="openshift-multus/multus-6ljgx" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.130952 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/5f3b70b5-9294-4f69-9528-500d28f34c89-multus-daemon-config\") pod \"multus-6ljgx\" (UID: \"5f3b70b5-9294-4f69-9528-500d28f34c89\") " pod="openshift-multus/multus-6ljgx" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.131033 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/222f65f2-0b82-4760-bcf7-779244720f01-run-openvswitch\") pod \"ovnkube-node-tpfb6\" (UID: \"222f65f2-0b82-4760-bcf7-779244720f01\") " pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.131111 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/5f3b70b5-9294-4f69-9528-500d28f34c89-hostroot\") pod \"multus-6ljgx\" (UID: \"5f3b70b5-9294-4f69-9528-500d28f34c89\") " pod="openshift-multus/multus-6ljgx" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.131191 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5f3b70b5-9294-4f69-9528-500d28f34c89-etc-kubernetes\") pod \"multus-6ljgx\" (UID: \"5f3b70b5-9294-4f69-9528-500d28f34c89\") " pod="openshift-multus/multus-6ljgx" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.131267 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/5a7d6101-d0d2-4828-b3ce-d01040c04640-cnibin\") pod \"multus-additional-cni-plugins-nl76g\" (UID: \"5a7d6101-d0d2-4828-b3ce-d01040c04640\") " pod="openshift-multus/multus-additional-cni-plugins-nl76g" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.131352 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/222f65f2-0b82-4760-bcf7-779244720f01-run-ovn\") pod \"ovnkube-node-tpfb6\" (UID: \"222f65f2-0b82-4760-bcf7-779244720f01\") " pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.131434 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/222f65f2-0b82-4760-bcf7-779244720f01-ovnkube-script-lib\") pod \"ovnkube-node-tpfb6\" (UID: \"222f65f2-0b82-4760-bcf7-779244720f01\") " pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.131553 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/e7670227-d280-4847-b882-754429f56b0f-rootfs\") pod \"machine-config-daemon-krv5b\" (UID: \"e7670227-d280-4847-b882-754429f56b0f\") " pod="openshift-machine-config-operator/machine-config-daemon-krv5b" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.130530 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.131696 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jfbnm\" (UniqueName: \"kubernetes.io/projected/e7670227-d280-4847-b882-754429f56b0f-kube-api-access-jfbnm\") pod \"machine-config-daemon-krv5b\" (UID: \"e7670227-d280-4847-b882-754429f56b0f\") " pod="openshift-machine-config-operator/machine-config-daemon-krv5b" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.131750 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/5a7d6101-d0d2-4828-b3ce-d01040c04640-os-release\") pod \"multus-additional-cni-plugins-nl76g\" (UID: \"5a7d6101-d0d2-4828-b3ce-d01040c04640\") " pod="openshift-multus/multus-additional-cni-plugins-nl76g" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.131813 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/222f65f2-0b82-4760-bcf7-779244720f01-host-kubelet\") pod \"ovnkube-node-tpfb6\" (UID: \"222f65f2-0b82-4760-bcf7-779244720f01\") " pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.131909 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qbsgb\" (UniqueName: \"kubernetes.io/projected/f46e2ba8-9ee0-406d-a668-2ba424618b69-kube-api-access-qbsgb\") pod \"node-resolver-4qwqj\" (UID: \"f46e2ba8-9ee0-406d-a668-2ba424618b69\") " pod="openshift-dns/node-resolver-4qwqj" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.132006 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/5a7d6101-d0d2-4828-b3ce-d01040c04640-tuning-conf-dir\") pod \"multus-additional-cni-plugins-nl76g\" (UID: \"5a7d6101-d0d2-4828-b3ce-d01040c04640\") " pod="openshift-multus/multus-additional-cni-plugins-nl76g" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.132141 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/222f65f2-0b82-4760-bcf7-779244720f01-host-cni-bin\") pod \"ovnkube-node-tpfb6\" (UID: \"222f65f2-0b82-4760-bcf7-779244720f01\") " pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.132181 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e7670227-d280-4847-b882-754429f56b0f-mcd-auth-proxy-config\") pod \"machine-config-daemon-krv5b\" (UID: \"e7670227-d280-4847-b882-754429f56b0f\") " pod="openshift-machine-config-operator/machine-config-daemon-krv5b" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.132209 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/5f3b70b5-9294-4f69-9528-500d28f34c89-os-release\") pod \"multus-6ljgx\" (UID: \"5f3b70b5-9294-4f69-9528-500d28f34c89\") " pod="openshift-multus/multus-6ljgx" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.132251 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/5f3b70b5-9294-4f69-9528-500d28f34c89-host-var-lib-cni-multus\") pod \"multus-6ljgx\" (UID: \"5f3b70b5-9294-4f69-9528-500d28f34c89\") " pod="openshift-multus/multus-6ljgx" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.132297 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/222f65f2-0b82-4760-bcf7-779244720f01-var-lib-openvswitch\") pod \"ovnkube-node-tpfb6\" (UID: \"222f65f2-0b82-4760-bcf7-779244720f01\") " pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.132408 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/222f65f2-0b82-4760-bcf7-779244720f01-host-cni-netd\") pod \"ovnkube-node-tpfb6\" (UID: \"222f65f2-0b82-4760-bcf7-779244720f01\") " pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.132449 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/5f3b70b5-9294-4f69-9528-500d28f34c89-host-run-multus-certs\") pod \"multus-6ljgx\" (UID: \"5f3b70b5-9294-4f69-9528-500d28f34c89\") " pod="openshift-multus/multus-6ljgx" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.132467 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/222f65f2-0b82-4760-bcf7-779244720f01-node-log\") pod \"ovnkube-node-tpfb6\" (UID: \"222f65f2-0b82-4760-bcf7-779244720f01\") " pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.132484 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/5f3b70b5-9294-4f69-9528-500d28f34c89-cnibin\") pod \"multus-6ljgx\" (UID: \"5f3b70b5-9294-4f69-9528-500d28f34c89\") " pod="openshift-multus/multus-6ljgx" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.132517 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/222f65f2-0b82-4760-bcf7-779244720f01-host-slash\") pod \"ovnkube-node-tpfb6\" (UID: \"222f65f2-0b82-4760-bcf7-779244720f01\") " pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.132559 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/222f65f2-0b82-4760-bcf7-779244720f01-host-run-netns\") pod \"ovnkube-node-tpfb6\" (UID: \"222f65f2-0b82-4760-bcf7-779244720f01\") " pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.132575 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.132594 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/222f65f2-0b82-4760-bcf7-779244720f01-ovnkube-config\") pod \"ovnkube-node-tpfb6\" (UID: \"222f65f2-0b82-4760-bcf7-779244720f01\") " pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.133888 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.134413 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.135840 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.136650 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.137989 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.138734 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.139904 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.140524 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.142147 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.142335 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6cd330fd-e8a6-49d5-bfc9-5f6122ccd49e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1bd64b9b938bc80284bcee54141e90cad613d339c4f27ef2934b29bc2a318916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://936f59996d26e42099dcd055f87fff9e3ac4283b57e1956fe97463f2334152d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca74b28a069dc2e609b0779164c3aa470a15ed9848af0dd5fc3479aaf750157b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://270c53b9a342f26be077a01cd4e87e9ccd0eb5d2c84208ae38cbf720463245e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66c521c26d541bbddb064dd9710f1a7b41090d2f47343977922272dfc6dab7af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6eb2ac410b49098e5b3603de5e6bf5ed4484fab0b300021d0bed2e5ed8a42527\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6eb2ac410b49098e5b3603de5e6bf5ed4484fab0b300021d0bed2e5ed8a42527\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03385e0c78e4806f60c160227ef4905a11003ed3d70f0e5c0dd0a477979d1d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03385e0c78e4806f60c160227ef4905a11003ed3d70f0e5c0dd0a477979d1d4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://50aea2fa8b3b141b4330ddce2c8e0715230c6f06dab64b3abfd98a3067224bec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50aea2fa8b3b141b4330ddce2c8e0715230c6f06dab64b3abfd98a3067224bec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.142756 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.143763 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.144243 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.145333 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.145953 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.146444 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.155652 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.168427 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.181119 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7670227-d280-4847-b882-754429f56b0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-krv5b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.199460 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"222f65f2-0b82-4760-bcf7-779244720f01\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tpfb6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.218258 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6cd330fd-e8a6-49d5-bfc9-5f6122ccd49e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1bd64b9b938bc80284bcee54141e90cad613d339c4f27ef2934b29bc2a318916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://936f59996d26e42099dcd055f87fff9e3ac4283b57e1956fe97463f2334152d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca74b28a069dc2e609b0779164c3aa470a15ed9848af0dd5fc3479aaf750157b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://270c53b9a342f26be077a01cd4e87e9ccd0eb5d2c84208ae38cbf720463245e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66c521c26d541bbddb064dd9710f1a7b41090d2f47343977922272dfc6dab7af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6eb2ac410b49098e5b3603de5e6bf5ed4484fab0b300021d0bed2e5ed8a42527\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6eb2ac410b49098e5b3603de5e6bf5ed4484fab0b300021d0bed2e5ed8a42527\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03385e0c78e4806f60c160227ef4905a11003ed3d70f0e5c0dd0a477979d1d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03385e0c78e4806f60c160227ef4905a11003ed3d70f0e5c0dd0a477979d1d4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://50aea2fa8b3b141b4330ddce2c8e0715230c6f06dab64b3abfd98a3067224bec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50aea2fa8b3b141b4330ddce2c8e0715230c6f06dab64b3abfd98a3067224bec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.230385 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.233742 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/5f3b70b5-9294-4f69-9528-500d28f34c89-host-var-lib-cni-bin\") pod \"multus-6ljgx\" (UID: \"5f3b70b5-9294-4f69-9528-500d28f34c89\") " pod="openshift-multus/multus-6ljgx" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.233782 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-78jhb\" (UniqueName: \"kubernetes.io/projected/5f3b70b5-9294-4f69-9528-500d28f34c89-kube-api-access-78jhb\") pod \"multus-6ljgx\" (UID: \"5f3b70b5-9294-4f69-9528-500d28f34c89\") " pod="openshift-multus/multus-6ljgx" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.233847 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/222f65f2-0b82-4760-bcf7-779244720f01-env-overrides\") pod \"ovnkube-node-tpfb6\" (UID: \"222f65f2-0b82-4760-bcf7-779244720f01\") " pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.233870 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/5f3b70b5-9294-4f69-9528-500d28f34c89-multus-socket-dir-parent\") pod \"multus-6ljgx\" (UID: \"5f3b70b5-9294-4f69-9528-500d28f34c89\") " pod="openshift-multus/multus-6ljgx" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.233886 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/5f3b70b5-9294-4f69-9528-500d28f34c89-host-var-lib-kubelet\") pod \"multus-6ljgx\" (UID: \"5f3b70b5-9294-4f69-9528-500d28f34c89\") " pod="openshift-multus/multus-6ljgx" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.233902 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/5f3b70b5-9294-4f69-9528-500d28f34c89-multus-conf-dir\") pod \"multus-6ljgx\" (UID: \"5f3b70b5-9294-4f69-9528-500d28f34c89\") " pod="openshift-multus/multus-6ljgx" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.233918 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/222f65f2-0b82-4760-bcf7-779244720f01-systemd-units\") pod \"ovnkube-node-tpfb6\" (UID: \"222f65f2-0b82-4760-bcf7-779244720f01\") " pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.233934 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/222f65f2-0b82-4760-bcf7-779244720f01-ovn-node-metrics-cert\") pod \"ovnkube-node-tpfb6\" (UID: \"222f65f2-0b82-4760-bcf7-779244720f01\") " pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.233963 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/5f3b70b5-9294-4f69-9528-500d28f34c89-cni-binary-copy\") pod \"multus-6ljgx\" (UID: \"5f3b70b5-9294-4f69-9528-500d28f34c89\") " pod="openshift-multus/multus-6ljgx" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.233966 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/5f3b70b5-9294-4f69-9528-500d28f34c89-multus-conf-dir\") pod \"multus-6ljgx\" (UID: \"5f3b70b5-9294-4f69-9528-500d28f34c89\") " pod="openshift-multus/multus-6ljgx" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.233989 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/5f3b70b5-9294-4f69-9528-500d28f34c89-host-run-k8s-cni-cncf-io\") pod \"multus-6ljgx\" (UID: \"5f3b70b5-9294-4f69-9528-500d28f34c89\") " pod="openshift-multus/multus-6ljgx" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.234041 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/5f3b70b5-9294-4f69-9528-500d28f34c89-host-run-k8s-cni-cncf-io\") pod \"multus-6ljgx\" (UID: \"5f3b70b5-9294-4f69-9528-500d28f34c89\") " pod="openshift-multus/multus-6ljgx" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.234041 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/5f3b70b5-9294-4f69-9528-500d28f34c89-multus-daemon-config\") pod \"multus-6ljgx\" (UID: \"5f3b70b5-9294-4f69-9528-500d28f34c89\") " pod="openshift-multus/multus-6ljgx" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.234060 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/5f3b70b5-9294-4f69-9528-500d28f34c89-multus-socket-dir-parent\") pod \"multus-6ljgx\" (UID: \"5f3b70b5-9294-4f69-9528-500d28f34c89\") " pod="openshift-multus/multus-6ljgx" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.234089 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/5f3b70b5-9294-4f69-9528-500d28f34c89-hostroot\") pod \"multus-6ljgx\" (UID: \"5f3b70b5-9294-4f69-9528-500d28f34c89\") " pod="openshift-multus/multus-6ljgx" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.234109 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5f3b70b5-9294-4f69-9528-500d28f34c89-etc-kubernetes\") pod \"multus-6ljgx\" (UID: \"5f3b70b5-9294-4f69-9528-500d28f34c89\") " pod="openshift-multus/multus-6ljgx" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.233925 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/5f3b70b5-9294-4f69-9528-500d28f34c89-host-var-lib-cni-bin\") pod \"multus-6ljgx\" (UID: \"5f3b70b5-9294-4f69-9528-500d28f34c89\") " pod="openshift-multus/multus-6ljgx" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.234131 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/5a7d6101-d0d2-4828-b3ce-d01040c04640-cnibin\") pod \"multus-additional-cni-plugins-nl76g\" (UID: \"5a7d6101-d0d2-4828-b3ce-d01040c04640\") " pod="openshift-multus/multus-additional-cni-plugins-nl76g" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.234152 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/222f65f2-0b82-4760-bcf7-779244720f01-run-openvswitch\") pod \"ovnkube-node-tpfb6\" (UID: \"222f65f2-0b82-4760-bcf7-779244720f01\") " pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.234169 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/222f65f2-0b82-4760-bcf7-779244720f01-run-ovn\") pod \"ovnkube-node-tpfb6\" (UID: \"222f65f2-0b82-4760-bcf7-779244720f01\") " pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.234190 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/222f65f2-0b82-4760-bcf7-779244720f01-ovnkube-script-lib\") pod \"ovnkube-node-tpfb6\" (UID: \"222f65f2-0b82-4760-bcf7-779244720f01\") " pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.234211 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/e7670227-d280-4847-b882-754429f56b0f-rootfs\") pod \"machine-config-daemon-krv5b\" (UID: \"e7670227-d280-4847-b882-754429f56b0f\") " pod="openshift-machine-config-operator/machine-config-daemon-krv5b" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.234233 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jfbnm\" (UniqueName: \"kubernetes.io/projected/e7670227-d280-4847-b882-754429f56b0f-kube-api-access-jfbnm\") pod \"machine-config-daemon-krv5b\" (UID: \"e7670227-d280-4847-b882-754429f56b0f\") " pod="openshift-machine-config-operator/machine-config-daemon-krv5b" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.234249 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/5a7d6101-d0d2-4828-b3ce-d01040c04640-os-release\") pod \"multus-additional-cni-plugins-nl76g\" (UID: \"5a7d6101-d0d2-4828-b3ce-d01040c04640\") " pod="openshift-multus/multus-additional-cni-plugins-nl76g" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.234265 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/222f65f2-0b82-4760-bcf7-779244720f01-host-kubelet\") pod \"ovnkube-node-tpfb6\" (UID: \"222f65f2-0b82-4760-bcf7-779244720f01\") " pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.234284 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qbsgb\" (UniqueName: \"kubernetes.io/projected/f46e2ba8-9ee0-406d-a668-2ba424618b69-kube-api-access-qbsgb\") pod \"node-resolver-4qwqj\" (UID: \"f46e2ba8-9ee0-406d-a668-2ba424618b69\") " pod="openshift-dns/node-resolver-4qwqj" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.234311 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/5a7d6101-d0d2-4828-b3ce-d01040c04640-tuning-conf-dir\") pod \"multus-additional-cni-plugins-nl76g\" (UID: \"5a7d6101-d0d2-4828-b3ce-d01040c04640\") " pod="openshift-multus/multus-additional-cni-plugins-nl76g" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.234329 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/222f65f2-0b82-4760-bcf7-779244720f01-host-cni-bin\") pod \"ovnkube-node-tpfb6\" (UID: \"222f65f2-0b82-4760-bcf7-779244720f01\") " pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.234169 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/222f65f2-0b82-4760-bcf7-779244720f01-systemd-units\") pod \"ovnkube-node-tpfb6\" (UID: \"222f65f2-0b82-4760-bcf7-779244720f01\") " pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.234351 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e7670227-d280-4847-b882-754429f56b0f-mcd-auth-proxy-config\") pod \"machine-config-daemon-krv5b\" (UID: \"e7670227-d280-4847-b882-754429f56b0f\") " pod="openshift-machine-config-operator/machine-config-daemon-krv5b" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.234470 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/5f3b70b5-9294-4f69-9528-500d28f34c89-os-release\") pod \"multus-6ljgx\" (UID: \"5f3b70b5-9294-4f69-9528-500d28f34c89\") " pod="openshift-multus/multus-6ljgx" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.234540 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/5f3b70b5-9294-4f69-9528-500d28f34c89-host-var-lib-cni-multus\") pod \"multus-6ljgx\" (UID: \"5f3b70b5-9294-4f69-9528-500d28f34c89\") " pod="openshift-multus/multus-6ljgx" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.234574 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/222f65f2-0b82-4760-bcf7-779244720f01-var-lib-openvswitch\") pod \"ovnkube-node-tpfb6\" (UID: \"222f65f2-0b82-4760-bcf7-779244720f01\") " pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.234578 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/5a7d6101-d0d2-4828-b3ce-d01040c04640-cnibin\") pod \"multus-additional-cni-plugins-nl76g\" (UID: \"5a7d6101-d0d2-4828-b3ce-d01040c04640\") " pod="openshift-multus/multus-additional-cni-plugins-nl76g" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.234611 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/222f65f2-0b82-4760-bcf7-779244720f01-host-cni-netd\") pod \"ovnkube-node-tpfb6\" (UID: \"222f65f2-0b82-4760-bcf7-779244720f01\") " pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.234672 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/5f3b70b5-9294-4f69-9528-500d28f34c89-host-run-multus-certs\") pod \"multus-6ljgx\" (UID: \"5f3b70b5-9294-4f69-9528-500d28f34c89\") " pod="openshift-multus/multus-6ljgx" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.234699 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/222f65f2-0b82-4760-bcf7-779244720f01-node-log\") pod \"ovnkube-node-tpfb6\" (UID: \"222f65f2-0b82-4760-bcf7-779244720f01\") " pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.234729 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/5f3b70b5-9294-4f69-9528-500d28f34c89-cnibin\") pod \"multus-6ljgx\" (UID: \"5f3b70b5-9294-4f69-9528-500d28f34c89\") " pod="openshift-multus/multus-6ljgx" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.234758 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/222f65f2-0b82-4760-bcf7-779244720f01-host-slash\") pod \"ovnkube-node-tpfb6\" (UID: \"222f65f2-0b82-4760-bcf7-779244720f01\") " pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.234788 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/222f65f2-0b82-4760-bcf7-779244720f01-host-run-netns\") pod \"ovnkube-node-tpfb6\" (UID: \"222f65f2-0b82-4760-bcf7-779244720f01\") " pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.234850 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/222f65f2-0b82-4760-bcf7-779244720f01-ovnkube-config\") pod \"ovnkube-node-tpfb6\" (UID: \"222f65f2-0b82-4760-bcf7-779244720f01\") " pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.234879 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/5a7d6101-d0d2-4828-b3ce-d01040c04640-cni-binary-copy\") pod \"multus-additional-cni-plugins-nl76g\" (UID: \"5a7d6101-d0d2-4828-b3ce-d01040c04640\") " pod="openshift-multus/multus-additional-cni-plugins-nl76g" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.234906 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/222f65f2-0b82-4760-bcf7-779244720f01-host-run-ovn-kubernetes\") pod \"ovnkube-node-tpfb6\" (UID: \"222f65f2-0b82-4760-bcf7-779244720f01\") " pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.234899 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/222f65f2-0b82-4760-bcf7-779244720f01-env-overrides\") pod \"ovnkube-node-tpfb6\" (UID: \"222f65f2-0b82-4760-bcf7-779244720f01\") " pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.234927 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/5f3b70b5-9294-4f69-9528-500d28f34c89-multus-daemon-config\") pod \"multus-6ljgx\" (UID: \"5f3b70b5-9294-4f69-9528-500d28f34c89\") " pod="openshift-multus/multus-6ljgx" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.234934 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mmvwn\" (UniqueName: \"kubernetes.io/projected/222f65f2-0b82-4760-bcf7-779244720f01-kube-api-access-mmvwn\") pod \"ovnkube-node-tpfb6\" (UID: \"222f65f2-0b82-4760-bcf7-779244720f01\") " pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.234970 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/5f3b70b5-9294-4f69-9528-500d28f34c89-host-run-netns\") pod \"multus-6ljgx\" (UID: \"5f3b70b5-9294-4f69-9528-500d28f34c89\") " pod="openshift-multus/multus-6ljgx" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.235059 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/222f65f2-0b82-4760-bcf7-779244720f01-node-log\") pod \"ovnkube-node-tpfb6\" (UID: \"222f65f2-0b82-4760-bcf7-779244720f01\") " pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.235162 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/e7670227-d280-4847-b882-754429f56b0f-rootfs\") pod \"machine-config-daemon-krv5b\" (UID: \"e7670227-d280-4847-b882-754429f56b0f\") " pod="openshift-machine-config-operator/machine-config-daemon-krv5b" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.235175 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e7670227-d280-4847-b882-754429f56b0f-mcd-auth-proxy-config\") pod \"machine-config-daemon-krv5b\" (UID: \"e7670227-d280-4847-b882-754429f56b0f\") " pod="openshift-machine-config-operator/machine-config-daemon-krv5b" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.235228 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/222f65f2-0b82-4760-bcf7-779244720f01-host-kubelet\") pod \"ovnkube-node-tpfb6\" (UID: \"222f65f2-0b82-4760-bcf7-779244720f01\") " pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.235228 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/5f3b70b5-9294-4f69-9528-500d28f34c89-cni-binary-copy\") pod \"multus-6ljgx\" (UID: \"5f3b70b5-9294-4f69-9528-500d28f34c89\") " pod="openshift-multus/multus-6ljgx" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.235260 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/5f3b70b5-9294-4f69-9528-500d28f34c89-hostroot\") pod \"multus-6ljgx\" (UID: \"5f3b70b5-9294-4f69-9528-500d28f34c89\") " pod="openshift-multus/multus-6ljgx" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.235294 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5f3b70b5-9294-4f69-9528-500d28f34c89-etc-kubernetes\") pod \"multus-6ljgx\" (UID: \"5f3b70b5-9294-4f69-9528-500d28f34c89\") " pod="openshift-multus/multus-6ljgx" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.235322 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/222f65f2-0b82-4760-bcf7-779244720f01-host-run-netns\") pod \"ovnkube-node-tpfb6\" (UID: \"222f65f2-0b82-4760-bcf7-779244720f01\") " pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.235340 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/222f65f2-0b82-4760-bcf7-779244720f01-run-openvswitch\") pod \"ovnkube-node-tpfb6\" (UID: \"222f65f2-0b82-4760-bcf7-779244720f01\") " pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.235384 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/222f65f2-0b82-4760-bcf7-779244720f01-etc-openvswitch\") pod \"ovnkube-node-tpfb6\" (UID: \"222f65f2-0b82-4760-bcf7-779244720f01\") " pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.235404 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/f46e2ba8-9ee0-406d-a668-2ba424618b69-hosts-file\") pod \"node-resolver-4qwqj\" (UID: \"f46e2ba8-9ee0-406d-a668-2ba424618b69\") " pod="openshift-dns/node-resolver-4qwqj" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.235431 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e7670227-d280-4847-b882-754429f56b0f-proxy-tls\") pod \"machine-config-daemon-krv5b\" (UID: \"e7670227-d280-4847-b882-754429f56b0f\") " pod="openshift-machine-config-operator/machine-config-daemon-krv5b" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.235518 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/5f3b70b5-9294-4f69-9528-500d28f34c89-multus-cni-dir\") pod \"multus-6ljgx\" (UID: \"5f3b70b5-9294-4f69-9528-500d28f34c89\") " pod="openshift-multus/multus-6ljgx" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.235542 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/5a7d6101-d0d2-4828-b3ce-d01040c04640-system-cni-dir\") pod \"multus-additional-cni-plugins-nl76g\" (UID: \"5a7d6101-d0d2-4828-b3ce-d01040c04640\") " pod="openshift-multus/multus-additional-cni-plugins-nl76g" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.235565 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4592l\" (UniqueName: \"kubernetes.io/projected/5a7d6101-d0d2-4828-b3ce-d01040c04640-kube-api-access-4592l\") pod \"multus-additional-cni-plugins-nl76g\" (UID: \"5a7d6101-d0d2-4828-b3ce-d01040c04640\") " pod="openshift-multus/multus-additional-cni-plugins-nl76g" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.235571 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/222f65f2-0b82-4760-bcf7-779244720f01-ovnkube-config\") pod \"ovnkube-node-tpfb6\" (UID: \"222f65f2-0b82-4760-bcf7-779244720f01\") " pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.235585 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/222f65f2-0b82-4760-bcf7-779244720f01-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-tpfb6\" (UID: \"222f65f2-0b82-4760-bcf7-779244720f01\") " pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.235606 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/222f65f2-0b82-4760-bcf7-779244720f01-host-cni-netd\") pod \"ovnkube-node-tpfb6\" (UID: \"222f65f2-0b82-4760-bcf7-779244720f01\") " pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.235607 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/5f3b70b5-9294-4f69-9528-500d28f34c89-system-cni-dir\") pod \"multus-6ljgx\" (UID: \"5f3b70b5-9294-4f69-9528-500d28f34c89\") " pod="openshift-multus/multus-6ljgx" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.235622 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/222f65f2-0b82-4760-bcf7-779244720f01-var-lib-openvswitch\") pod \"ovnkube-node-tpfb6\" (UID: \"222f65f2-0b82-4760-bcf7-779244720f01\") " pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.235644 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/5f3b70b5-9294-4f69-9528-500d28f34c89-system-cni-dir\") pod \"multus-6ljgx\" (UID: \"5f3b70b5-9294-4f69-9528-500d28f34c89\") " pod="openshift-multus/multus-6ljgx" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.235647 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/5a7d6101-d0d2-4828-b3ce-d01040c04640-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-nl76g\" (UID: \"5a7d6101-d0d2-4828-b3ce-d01040c04640\") " pod="openshift-multus/multus-additional-cni-plugins-nl76g" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.235659 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/5f3b70b5-9294-4f69-9528-500d28f34c89-host-var-lib-cni-multus\") pod \"multus-6ljgx\" (UID: \"5f3b70b5-9294-4f69-9528-500d28f34c89\") " pod="openshift-multus/multus-6ljgx" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.235676 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/222f65f2-0b82-4760-bcf7-779244720f01-run-systemd\") pod \"ovnkube-node-tpfb6\" (UID: \"222f65f2-0b82-4760-bcf7-779244720f01\") " pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.235704 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/222f65f2-0b82-4760-bcf7-779244720f01-log-socket\") pod \"ovnkube-node-tpfb6\" (UID: \"222f65f2-0b82-4760-bcf7-779244720f01\") " pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.235836 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/5a7d6101-d0d2-4828-b3ce-d01040c04640-tuning-conf-dir\") pod \"multus-additional-cni-plugins-nl76g\" (UID: \"5a7d6101-d0d2-4828-b3ce-d01040c04640\") " pod="openshift-multus/multus-additional-cni-plugins-nl76g" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.235890 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/5a7d6101-d0d2-4828-b3ce-d01040c04640-os-release\") pod \"multus-additional-cni-plugins-nl76g\" (UID: \"5a7d6101-d0d2-4828-b3ce-d01040c04640\") " pod="openshift-multus/multus-additional-cni-plugins-nl76g" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.235898 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/5f3b70b5-9294-4f69-9528-500d28f34c89-host-run-multus-certs\") pod \"multus-6ljgx\" (UID: \"5f3b70b5-9294-4f69-9528-500d28f34c89\") " pod="openshift-multus/multus-6ljgx" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.235937 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/222f65f2-0b82-4760-bcf7-779244720f01-etc-openvswitch\") pod \"ovnkube-node-tpfb6\" (UID: \"222f65f2-0b82-4760-bcf7-779244720f01\") " pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.235975 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/5f3b70b5-9294-4f69-9528-500d28f34c89-cnibin\") pod \"multus-6ljgx\" (UID: \"5f3b70b5-9294-4f69-9528-500d28f34c89\") " pod="openshift-multus/multus-6ljgx" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.235999 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/222f65f2-0b82-4760-bcf7-779244720f01-run-ovn\") pod \"ovnkube-node-tpfb6\" (UID: \"222f65f2-0b82-4760-bcf7-779244720f01\") " pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.236017 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/222f65f2-0b82-4760-bcf7-779244720f01-host-slash\") pod \"ovnkube-node-tpfb6\" (UID: \"222f65f2-0b82-4760-bcf7-779244720f01\") " pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.236046 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/5f3b70b5-9294-4f69-9528-500d28f34c89-os-release\") pod \"multus-6ljgx\" (UID: \"5f3b70b5-9294-4f69-9528-500d28f34c89\") " pod="openshift-multus/multus-6ljgx" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.236062 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/5f3b70b5-9294-4f69-9528-500d28f34c89-host-run-netns\") pod \"multus-6ljgx\" (UID: \"5f3b70b5-9294-4f69-9528-500d28f34c89\") " pod="openshift-multus/multus-6ljgx" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.236089 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/222f65f2-0b82-4760-bcf7-779244720f01-run-systemd\") pod \"ovnkube-node-tpfb6\" (UID: \"222f65f2-0b82-4760-bcf7-779244720f01\") " pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.236112 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/222f65f2-0b82-4760-bcf7-779244720f01-log-socket\") pod \"ovnkube-node-tpfb6\" (UID: \"222f65f2-0b82-4760-bcf7-779244720f01\") " pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.236134 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/222f65f2-0b82-4760-bcf7-779244720f01-host-run-ovn-kubernetes\") pod \"ovnkube-node-tpfb6\" (UID: \"222f65f2-0b82-4760-bcf7-779244720f01\") " pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.235498 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/222f65f2-0b82-4760-bcf7-779244720f01-host-cni-bin\") pod \"ovnkube-node-tpfb6\" (UID: \"222f65f2-0b82-4760-bcf7-779244720f01\") " pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.236175 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/f46e2ba8-9ee0-406d-a668-2ba424618b69-hosts-file\") pod \"node-resolver-4qwqj\" (UID: \"f46e2ba8-9ee0-406d-a668-2ba424618b69\") " pod="openshift-dns/node-resolver-4qwqj" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.236194 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/5a7d6101-d0d2-4828-b3ce-d01040c04640-cni-binary-copy\") pod \"multus-additional-cni-plugins-nl76g\" (UID: \"5a7d6101-d0d2-4828-b3ce-d01040c04640\") " pod="openshift-multus/multus-additional-cni-plugins-nl76g" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.236269 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/5f3b70b5-9294-4f69-9528-500d28f34c89-multus-cni-dir\") pod \"multus-6ljgx\" (UID: \"5f3b70b5-9294-4f69-9528-500d28f34c89\") " pod="openshift-multus/multus-6ljgx" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.236444 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/5a7d6101-d0d2-4828-b3ce-d01040c04640-system-cni-dir\") pod \"multus-additional-cni-plugins-nl76g\" (UID: \"5a7d6101-d0d2-4828-b3ce-d01040c04640\") " pod="openshift-multus/multus-additional-cni-plugins-nl76g" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.236455 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/5a7d6101-d0d2-4828-b3ce-d01040c04640-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-nl76g\" (UID: \"5a7d6101-d0d2-4828-b3ce-d01040c04640\") " pod="openshift-multus/multus-additional-cni-plugins-nl76g" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.236469 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/222f65f2-0b82-4760-bcf7-779244720f01-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-tpfb6\" (UID: \"222f65f2-0b82-4760-bcf7-779244720f01\") " pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.236617 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/5f3b70b5-9294-4f69-9528-500d28f34c89-host-var-lib-kubelet\") pod \"multus-6ljgx\" (UID: \"5f3b70b5-9294-4f69-9528-500d28f34c89\") " pod="openshift-multus/multus-6ljgx" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.236652 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/222f65f2-0b82-4760-bcf7-779244720f01-ovnkube-script-lib\") pod \"ovnkube-node-tpfb6\" (UID: \"222f65f2-0b82-4760-bcf7-779244720f01\") " pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.239898 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e7670227-d280-4847-b882-754429f56b0f-proxy-tls\") pod \"machine-config-daemon-krv5b\" (UID: \"e7670227-d280-4847-b882-754429f56b0f\") " pod="openshift-machine-config-operator/machine-config-daemon-krv5b" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.240356 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/222f65f2-0b82-4760-bcf7-779244720f01-ovn-node-metrics-cert\") pod \"ovnkube-node-tpfb6\" (UID: \"222f65f2-0b82-4760-bcf7-779244720f01\") " pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.246346 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.248768 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"f2480d5e27cb2f647f721edbabdca62343cbd016d024c1c1f7f15f42ae65f18d"} Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.252098 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"48c6b1ea7497b33c00b956e400426c1d4f3c6c66e0225133434460800eef9314"} Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.252131 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"1ea6068a6b6f8123b402de354e5374a0f02a7191a922d1718ab38f8898f47c52"} Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.252144 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"37eec6fd9f3838ee5d82e1ba471e9b74205258e739f26e3f63f023a04e9207e7"} Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.252138 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mmvwn\" (UniqueName: \"kubernetes.io/projected/222f65f2-0b82-4760-bcf7-779244720f01-kube-api-access-mmvwn\") pod \"ovnkube-node-tpfb6\" (UID: \"222f65f2-0b82-4760-bcf7-779244720f01\") " pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.254573 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jfbnm\" (UniqueName: \"kubernetes.io/projected/e7670227-d280-4847-b882-754429f56b0f-kube-api-access-jfbnm\") pod \"machine-config-daemon-krv5b\" (UID: \"e7670227-d280-4847-b882-754429f56b0f\") " pod="openshift-machine-config-operator/machine-config-daemon-krv5b" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.255599 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"c976d804a7ebeba2c879e920cbc389f96f7f5def4b4439e8c8980d5460ecdb65"} Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.255679 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"578022ed984cdab10cc838090bcd7b03e984625f869859d5d8acc3105630f5b0"} Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.255695 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4592l\" (UniqueName: \"kubernetes.io/projected/5a7d6101-d0d2-4828-b3ce-d01040c04640-kube-api-access-4592l\") pod \"multus-additional-cni-plugins-nl76g\" (UID: \"5a7d6101-d0d2-4828-b3ce-d01040c04640\") " pod="openshift-multus/multus-additional-cni-plugins-nl76g" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.257446 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qbsgb\" (UniqueName: \"kubernetes.io/projected/f46e2ba8-9ee0-406d-a668-2ba424618b69-kube-api-access-qbsgb\") pod \"node-resolver-4qwqj\" (UID: \"f46e2ba8-9ee0-406d-a668-2ba424618b69\") " pod="openshift-dns/node-resolver-4qwqj" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.259104 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6ljgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f3b70b5-9294-4f69-9528-500d28f34c89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78jhb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6ljgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.261847 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-78jhb\" (UniqueName: \"kubernetes.io/projected/5f3b70b5-9294-4f69-9528-500d28f34c89-kube-api-access-78jhb\") pod \"multus-6ljgx\" (UID: \"5f3b70b5-9294-4f69-9528-500d28f34c89\") " pod="openshift-multus/multus-6ljgx" Nov 21 14:09:10 crc kubenswrapper[4897]: E1121 14:09:10.264497 4897 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"etcd-crc\" already exists" pod="openshift-etcd/etcd-crc" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.269943 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.285585 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:10Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.302066 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nl76g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a7d6101-d0d2-4828-b3ce-d01040c04640\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nl76g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:10Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.318454 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:10Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.341872 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4qwqj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f46e2ba8-9ee0-406d-a668-2ba424618b69\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbsgb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4qwqj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:10Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.355015 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48c6b1ea7497b33c00b956e400426c1d4f3c6c66e0225133434460800eef9314\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ea6068a6b6f8123b402de354e5374a0f02a7191a922d1718ab38f8898f47c52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:10Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.357395 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.366026 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4qwqj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f46e2ba8-9ee0-406d-a668-2ba424618b69\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbsgb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4qwqj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:10Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.367929 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-4qwqj" Nov 21 14:09:10 crc kubenswrapper[4897]: W1121 14:09:10.371161 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode7670227_d280_4847_b882_754429f56b0f.slice/crio-64645d8ad59a4e8ac10d6faaceb001478b775eede50c9252407045a0bc25c23a WatchSource:0}: Error finding container 64645d8ad59a4e8ac10d6faaceb001478b775eede50c9252407045a0bc25c23a: Status 404 returned error can't find the container with id 64645d8ad59a4e8ac10d6faaceb001478b775eede50c9252407045a0bc25c23a Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.378355 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-6ljgx" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.381520 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:10Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.386237 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-nl76g" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.394522 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.396119 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7670227-d280-4847-b882-754429f56b0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-krv5b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:10Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:10 crc kubenswrapper[4897]: W1121 14:09:10.402054 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5f3b70b5_9294_4f69_9528_500d28f34c89.slice/crio-9e5c078433c6879ec760ae29a7d8b16bd19982c6e288c367e697678e7a40c1af WatchSource:0}: Error finding container 9e5c078433c6879ec760ae29a7d8b16bd19982c6e288c367e697678e7a40c1af: Status 404 returned error can't find the container with id 9e5c078433c6879ec760ae29a7d8b16bd19982c6e288c367e697678e7a40c1af Nov 21 14:09:10 crc kubenswrapper[4897]: W1121 14:09:10.411964 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5a7d6101_d0d2_4828_b3ce_d01040c04640.slice/crio-f6d50805cc20d5a225778ebfe52014943be7bc04421109088067d0e048857288 WatchSource:0}: Error finding container f6d50805cc20d5a225778ebfe52014943be7bc04421109088067d0e048857288: Status 404 returned error can't find the container with id f6d50805cc20d5a225778ebfe52014943be7bc04421109088067d0e048857288 Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.417616 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"222f65f2-0b82-4760-bcf7-779244720f01\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tpfb6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:10Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.438095 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6ljgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f3b70b5-9294-4f69-9528-500d28f34c89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78jhb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6ljgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:10Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.459702 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6cd330fd-e8a6-49d5-bfc9-5f6122ccd49e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1bd64b9b938bc80284bcee54141e90cad613d339c4f27ef2934b29bc2a318916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://936f59996d26e42099dcd055f87fff9e3ac4283b57e1956fe97463f2334152d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca74b28a069dc2e609b0779164c3aa470a15ed9848af0dd5fc3479aaf750157b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://270c53b9a342f26be077a01cd4e87e9ccd0eb5d2c84208ae38cbf720463245e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66c521c26d541bbddb064dd9710f1a7b41090d2f47343977922272dfc6dab7af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6eb2ac410b49098e5b3603de5e6bf5ed4484fab0b300021d0bed2e5ed8a42527\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6eb2ac410b49098e5b3603de5e6bf5ed4484fab0b300021d0bed2e5ed8a42527\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03385e0c78e4806f60c160227ef4905a11003ed3d70f0e5c0dd0a477979d1d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03385e0c78e4806f60c160227ef4905a11003ed3d70f0e5c0dd0a477979d1d4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://50aea2fa8b3b141b4330ddce2c8e0715230c6f06dab64b3abfd98a3067224bec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50aea2fa8b3b141b4330ddce2c8e0715230c6f06dab64b3abfd98a3067224bec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:10Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.479536 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:10Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.496395 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:10Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.498617 4897 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.509863 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:10Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.512097 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.512561 4897 scope.go:117] "RemoveContainer" containerID="1907bfbad7eb34da642e30d6aa81820aab457f92347333a364ac12396b0e9dae" Nov 21 14:09:10 crc kubenswrapper[4897]: E1121 14:09:10.513096 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.525852 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c976d804a7ebeba2c879e920cbc389f96f7f5def4b4439e8c8980d5460ecdb65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:10Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.546914 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nl76g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a7d6101-d0d2-4828-b3ce-d01040c04640\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nl76g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:10Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.566893 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:10Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.581031 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7670227-d280-4847-b882-754429f56b0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-krv5b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:10Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.600608 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"222f65f2-0b82-4760-bcf7-779244720f01\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tpfb6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:10Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.624543 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6cd330fd-e8a6-49d5-bfc9-5f6122ccd49e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1bd64b9b938bc80284bcee54141e90cad613d339c4f27ef2934b29bc2a318916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://936f59996d26e42099dcd055f87fff9e3ac4283b57e1956fe97463f2334152d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca74b28a069dc2e609b0779164c3aa470a15ed9848af0dd5fc3479aaf750157b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://270c53b9a342f26be077a01cd4e87e9ccd0eb5d2c84208ae38cbf720463245e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66c521c26d541bbddb064dd9710f1a7b41090d2f47343977922272dfc6dab7af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6eb2ac410b49098e5b3603de5e6bf5ed4484fab0b300021d0bed2e5ed8a42527\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6eb2ac410b49098e5b3603de5e6bf5ed4484fab0b300021d0bed2e5ed8a42527\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03385e0c78e4806f60c160227ef4905a11003ed3d70f0e5c0dd0a477979d1d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03385e0c78e4806f60c160227ef4905a11003ed3d70f0e5c0dd0a477979d1d4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://50aea2fa8b3b141b4330ddce2c8e0715230c6f06dab64b3abfd98a3067224bec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50aea2fa8b3b141b4330ddce2c8e0715230c6f06dab64b3abfd98a3067224bec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:10Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.638963 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a6c0a99-565b-41f9-a8b5-c8990a0c8eb3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://737553dd1d16e2019ea89a376e3a868f9792ed5239aa7d6cc40a05d3d4f286ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6682a8886d2b9bce069cf9261d4701ef6907f19574fd035bd89bd05626023e5c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de079956e3d8de0101f3f87dcc120ef23876c4972382639e419b3f34a38cdfca\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1907bfbad7eb34da642e30d6aa81820aab457f92347333a364ac12396b0e9dae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1907bfbad7eb34da642e30d6aa81820aab457f92347333a364ac12396b0e9dae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"le observer\\\\nW1121 14:09:07.994401 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1121 14:09:07.994601 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1121 14:09:07.995290 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1461162674/tls.crt::/tmp/serving-cert-1461162674/tls.key\\\\\\\"\\\\nI1121 14:09:08.162382 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1121 14:09:08.165301 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1121 14:09:08.165318 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1121 14:09:08.165343 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1121 14:09:08.165348 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1121 14:09:08.172015 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1121 14:09:08.172083 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1121 14:09:08.172089 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1121 14:09:08.172094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1121 14:09:08.172098 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1121 14:09:08.172102 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1121 14:09:08.172106 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1121 14:09:08.172023 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1121 14:09:08.174214 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d78c87c440c7d8e30075f6143111b3f504fb99686484b8f19a9c8d61a55eccc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://101fb592fef069af009c613a1e9c9825024ed205498ab72334211ba90d6f7397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://101fb592fef069af009c613a1e9c9825024ed205498ab72334211ba90d6f7397\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:10Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.639596 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 21 14:09:10 crc kubenswrapper[4897]: E1121 14:09:10.639868 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-21 14:09:12.639832926 +0000 UTC m=+29.924426401 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.639936 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.640004 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 21 14:09:10 crc kubenswrapper[4897]: E1121 14:09:10.640146 4897 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 21 14:09:10 crc kubenswrapper[4897]: E1121 14:09:10.640233 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-21 14:09:12.640212156 +0000 UTC m=+29.924805641 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 21 14:09:10 crc kubenswrapper[4897]: E1121 14:09:10.640411 4897 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 21 14:09:10 crc kubenswrapper[4897]: E1121 14:09:10.640461 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-21 14:09:12.640448182 +0000 UTC m=+29.925041657 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.651210 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:10Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.664883 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:10Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.678190 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6ljgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f3b70b5-9294-4f69-9528-500d28f34c89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78jhb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6ljgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:10Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.696726 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:10Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.712950 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c976d804a7ebeba2c879e920cbc389f96f7f5def4b4439e8c8980d5460ecdb65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:10Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.730342 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nl76g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a7d6101-d0d2-4828-b3ce-d01040c04640\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nl76g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:10Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.741694 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.741750 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 21 14:09:10 crc kubenswrapper[4897]: E1121 14:09:10.741929 4897 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 21 14:09:10 crc kubenswrapper[4897]: E1121 14:09:10.741956 4897 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 21 14:09:10 crc kubenswrapper[4897]: E1121 14:09:10.741971 4897 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 21 14:09:10 crc kubenswrapper[4897]: E1121 14:09:10.742030 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-21 14:09:12.742011659 +0000 UTC m=+30.026605134 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 21 14:09:10 crc kubenswrapper[4897]: E1121 14:09:10.742103 4897 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 21 14:09:10 crc kubenswrapper[4897]: E1121 14:09:10.742115 4897 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 21 14:09:10 crc kubenswrapper[4897]: E1121 14:09:10.742125 4897 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 21 14:09:10 crc kubenswrapper[4897]: E1121 14:09:10.742150 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-21 14:09:12.742142012 +0000 UTC m=+30.026735487 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.747766 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48c6b1ea7497b33c00b956e400426c1d4f3c6c66e0225133434460800eef9314\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ea6068a6b6f8123b402de354e5374a0f02a7191a922d1718ab38f8898f47c52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:10Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:10 crc kubenswrapper[4897]: I1121 14:09:10.765190 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4qwqj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f46e2ba8-9ee0-406d-a668-2ba424618b69\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbsgb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4qwqj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:10Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:11 crc kubenswrapper[4897]: I1121 14:09:11.088269 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 21 14:09:11 crc kubenswrapper[4897]: I1121 14:09:11.088324 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 21 14:09:11 crc kubenswrapper[4897]: I1121 14:09:11.088431 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 21 14:09:11 crc kubenswrapper[4897]: E1121 14:09:11.088442 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 21 14:09:11 crc kubenswrapper[4897]: E1121 14:09:11.088616 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 21 14:09:11 crc kubenswrapper[4897]: E1121 14:09:11.088697 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 21 14:09:11 crc kubenswrapper[4897]: I1121 14:09:11.259559 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-6ljgx" event={"ID":"5f3b70b5-9294-4f69-9528-500d28f34c89","Type":"ContainerStarted","Data":"527cf28abdeef1b42d9c19fb88e3171dc05c4516c3a542a461b5b1daa85c6d25"} Nov 21 14:09:11 crc kubenswrapper[4897]: I1121 14:09:11.259605 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-6ljgx" event={"ID":"5f3b70b5-9294-4f69-9528-500d28f34c89","Type":"ContainerStarted","Data":"9e5c078433c6879ec760ae29a7d8b16bd19982c6e288c367e697678e7a40c1af"} Nov 21 14:09:11 crc kubenswrapper[4897]: I1121 14:09:11.261371 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-4qwqj" event={"ID":"f46e2ba8-9ee0-406d-a668-2ba424618b69","Type":"ContainerStarted","Data":"3f43aa89d4d9defcb50d0dc6f90293b5f426b9cfc398b780c2e10f6a5c9f1e2d"} Nov 21 14:09:11 crc kubenswrapper[4897]: I1121 14:09:11.261398 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-4qwqj" event={"ID":"f46e2ba8-9ee0-406d-a668-2ba424618b69","Type":"ContainerStarted","Data":"7c3bedd0dadb64bdf8f551857bc2d7a88077b9258bd5ee1642cb3aaf58ed3ef8"} Nov 21 14:09:11 crc kubenswrapper[4897]: I1121 14:09:11.263595 4897 generic.go:334] "Generic (PLEG): container finished" podID="222f65f2-0b82-4760-bcf7-779244720f01" containerID="7877db6f5c2bf0eb31218c64d67bca48f921ff7ac62b4e495f29024c086bf738" exitCode=0 Nov 21 14:09:11 crc kubenswrapper[4897]: I1121 14:09:11.263651 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" event={"ID":"222f65f2-0b82-4760-bcf7-779244720f01","Type":"ContainerDied","Data":"7877db6f5c2bf0eb31218c64d67bca48f921ff7ac62b4e495f29024c086bf738"} Nov 21 14:09:11 crc kubenswrapper[4897]: I1121 14:09:11.263675 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" event={"ID":"222f65f2-0b82-4760-bcf7-779244720f01","Type":"ContainerStarted","Data":"deecb27cbb58ca8e141468253196c2b73bb738df87fcd915dc32ef490981f458"} Nov 21 14:09:11 crc kubenswrapper[4897]: I1121 14:09:11.266264 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" event={"ID":"e7670227-d280-4847-b882-754429f56b0f","Type":"ContainerStarted","Data":"961961696b8750c248598f4cda032b18395e5c802275dcfe6d3e9f7ef3510f1c"} Nov 21 14:09:11 crc kubenswrapper[4897]: I1121 14:09:11.266325 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" event={"ID":"e7670227-d280-4847-b882-754429f56b0f","Type":"ContainerStarted","Data":"5854028821e6a360a259427b68e7bb4c2a7220e0c68f4cc91639b2ece0540a02"} Nov 21 14:09:11 crc kubenswrapper[4897]: I1121 14:09:11.266346 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" event={"ID":"e7670227-d280-4847-b882-754429f56b0f","Type":"ContainerStarted","Data":"64645d8ad59a4e8ac10d6faaceb001478b775eede50c9252407045a0bc25c23a"} Nov 21 14:09:11 crc kubenswrapper[4897]: I1121 14:09:11.267850 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-nl76g" event={"ID":"5a7d6101-d0d2-4828-b3ce-d01040c04640","Type":"ContainerStarted","Data":"e56402f62821baf335f378084f4ed7c21b60ccbafe148419a0a9a403f106ad19"} Nov 21 14:09:11 crc kubenswrapper[4897]: I1121 14:09:11.267922 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-nl76g" event={"ID":"5a7d6101-d0d2-4828-b3ce-d01040c04640","Type":"ContainerStarted","Data":"f6d50805cc20d5a225778ebfe52014943be7bc04421109088067d0e048857288"} Nov 21 14:09:11 crc kubenswrapper[4897]: I1121 14:09:11.269197 4897 scope.go:117] "RemoveContainer" containerID="1907bfbad7eb34da642e30d6aa81820aab457f92347333a364ac12396b0e9dae" Nov 21 14:09:11 crc kubenswrapper[4897]: E1121 14:09:11.269666 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Nov 21 14:09:11 crc kubenswrapper[4897]: I1121 14:09:11.281302 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6ljgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f3b70b5-9294-4f69-9528-500d28f34c89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://527cf28abdeef1b42d9c19fb88e3171dc05c4516c3a542a461b5b1daa85c6d25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78jhb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6ljgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:11Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:11 crc kubenswrapper[4897]: I1121 14:09:11.303761 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6cd330fd-e8a6-49d5-bfc9-5f6122ccd49e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1bd64b9b938bc80284bcee54141e90cad613d339c4f27ef2934b29bc2a318916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://936f59996d26e42099dcd055f87fff9e3ac4283b57e1956fe97463f2334152d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca74b28a069dc2e609b0779164c3aa470a15ed9848af0dd5fc3479aaf750157b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://270c53b9a342f26be077a01cd4e87e9ccd0eb5d2c84208ae38cbf720463245e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66c521c26d541bbddb064dd9710f1a7b41090d2f47343977922272dfc6dab7af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6eb2ac410b49098e5b3603de5e6bf5ed4484fab0b300021d0bed2e5ed8a42527\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6eb2ac410b49098e5b3603de5e6bf5ed4484fab0b300021d0bed2e5ed8a42527\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03385e0c78e4806f60c160227ef4905a11003ed3d70f0e5c0dd0a477979d1d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03385e0c78e4806f60c160227ef4905a11003ed3d70f0e5c0dd0a477979d1d4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://50aea2fa8b3b141b4330ddce2c8e0715230c6f06dab64b3abfd98a3067224bec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50aea2fa8b3b141b4330ddce2c8e0715230c6f06dab64b3abfd98a3067224bec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:11Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:11 crc kubenswrapper[4897]: I1121 14:09:11.319980 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a6c0a99-565b-41f9-a8b5-c8990a0c8eb3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://737553dd1d16e2019ea89a376e3a868f9792ed5239aa7d6cc40a05d3d4f286ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6682a8886d2b9bce069cf9261d4701ef6907f19574fd035bd89bd05626023e5c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de079956e3d8de0101f3f87dcc120ef23876c4972382639e419b3f34a38cdfca\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1907bfbad7eb34da642e30d6aa81820aab457f92347333a364ac12396b0e9dae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1907bfbad7eb34da642e30d6aa81820aab457f92347333a364ac12396b0e9dae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"le observer\\\\nW1121 14:09:07.994401 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1121 14:09:07.994601 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1121 14:09:07.995290 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1461162674/tls.crt::/tmp/serving-cert-1461162674/tls.key\\\\\\\"\\\\nI1121 14:09:08.162382 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1121 14:09:08.165301 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1121 14:09:08.165318 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1121 14:09:08.165343 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1121 14:09:08.165348 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1121 14:09:08.172015 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1121 14:09:08.172083 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1121 14:09:08.172089 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1121 14:09:08.172094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1121 14:09:08.172098 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1121 14:09:08.172102 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1121 14:09:08.172106 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1121 14:09:08.172023 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1121 14:09:08.174214 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d78c87c440c7d8e30075f6143111b3f504fb99686484b8f19a9c8d61a55eccc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://101fb592fef069af009c613a1e9c9825024ed205498ab72334211ba90d6f7397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://101fb592fef069af009c613a1e9c9825024ed205498ab72334211ba90d6f7397\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:11Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:11 crc kubenswrapper[4897]: I1121 14:09:11.333874 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:11Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:11 crc kubenswrapper[4897]: I1121 14:09:11.348726 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:11Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:11 crc kubenswrapper[4897]: I1121 14:09:11.363961 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:11Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:11 crc kubenswrapper[4897]: I1121 14:09:11.378858 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c976d804a7ebeba2c879e920cbc389f96f7f5def4b4439e8c8980d5460ecdb65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:11Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:11 crc kubenswrapper[4897]: I1121 14:09:11.395543 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nl76g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a7d6101-d0d2-4828-b3ce-d01040c04640\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nl76g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:11Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:11 crc kubenswrapper[4897]: I1121 14:09:11.410253 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48c6b1ea7497b33c00b956e400426c1d4f3c6c66e0225133434460800eef9314\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ea6068a6b6f8123b402de354e5374a0f02a7191a922d1718ab38f8898f47c52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:11Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:11 crc kubenswrapper[4897]: I1121 14:09:11.421997 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4qwqj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f46e2ba8-9ee0-406d-a668-2ba424618b69\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbsgb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4qwqj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:11Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:11 crc kubenswrapper[4897]: I1121 14:09:11.437227 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:11Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:11 crc kubenswrapper[4897]: I1121 14:09:11.451637 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7670227-d280-4847-b882-754429f56b0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-krv5b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:11Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:11 crc kubenswrapper[4897]: I1121 14:09:11.472167 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"222f65f2-0b82-4760-bcf7-779244720f01\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tpfb6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:11Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:11 crc kubenswrapper[4897]: I1121 14:09:11.488050 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48c6b1ea7497b33c00b956e400426c1d4f3c6c66e0225133434460800eef9314\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ea6068a6b6f8123b402de354e5374a0f02a7191a922d1718ab38f8898f47c52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:11Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:11 crc kubenswrapper[4897]: I1121 14:09:11.501006 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4qwqj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f46e2ba8-9ee0-406d-a668-2ba424618b69\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f43aa89d4d9defcb50d0dc6f90293b5f426b9cfc398b780c2e10f6a5c9f1e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbsgb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4qwqj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:11Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:11 crc kubenswrapper[4897]: I1121 14:09:11.518061 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:11Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:11 crc kubenswrapper[4897]: I1121 14:09:11.532498 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7670227-d280-4847-b882-754429f56b0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://961961696b8750c248598f4cda032b18395e5c802275dcfe6d3e9f7ef3510f1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5854028821e6a360a259427b68e7bb4c2a7220e0c68f4cc91639b2ece0540a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-krv5b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:11Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:11 crc kubenswrapper[4897]: I1121 14:09:11.555731 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"222f65f2-0b82-4760-bcf7-779244720f01\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7877db6f5c2bf0eb31218c64d67bca48f921ff7ac62b4e495f29024c086bf738\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7877db6f5c2bf0eb31218c64d67bca48f921ff7ac62b4e495f29024c086bf738\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tpfb6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:11Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:11 crc kubenswrapper[4897]: I1121 14:09:11.584390 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:11Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:11 crc kubenswrapper[4897]: I1121 14:09:11.606475 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6ljgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f3b70b5-9294-4f69-9528-500d28f34c89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://527cf28abdeef1b42d9c19fb88e3171dc05c4516c3a542a461b5b1daa85c6d25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78jhb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6ljgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:11Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:11 crc kubenswrapper[4897]: I1121 14:09:11.638034 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6cd330fd-e8a6-49d5-bfc9-5f6122ccd49e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1bd64b9b938bc80284bcee54141e90cad613d339c4f27ef2934b29bc2a318916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://936f59996d26e42099dcd055f87fff9e3ac4283b57e1956fe97463f2334152d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca74b28a069dc2e609b0779164c3aa470a15ed9848af0dd5fc3479aaf750157b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://270c53b9a342f26be077a01cd4e87e9ccd0eb5d2c84208ae38cbf720463245e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66c521c26d541bbddb064dd9710f1a7b41090d2f47343977922272dfc6dab7af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6eb2ac410b49098e5b3603de5e6bf5ed4484fab0b300021d0bed2e5ed8a42527\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6eb2ac410b49098e5b3603de5e6bf5ed4484fab0b300021d0bed2e5ed8a42527\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03385e0c78e4806f60c160227ef4905a11003ed3d70f0e5c0dd0a477979d1d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03385e0c78e4806f60c160227ef4905a11003ed3d70f0e5c0dd0a477979d1d4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://50aea2fa8b3b141b4330ddce2c8e0715230c6f06dab64b3abfd98a3067224bec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50aea2fa8b3b141b4330ddce2c8e0715230c6f06dab64b3abfd98a3067224bec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:11Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:11 crc kubenswrapper[4897]: I1121 14:09:11.659973 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a6c0a99-565b-41f9-a8b5-c8990a0c8eb3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://737553dd1d16e2019ea89a376e3a868f9792ed5239aa7d6cc40a05d3d4f286ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6682a8886d2b9bce069cf9261d4701ef6907f19574fd035bd89bd05626023e5c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de079956e3d8de0101f3f87dcc120ef23876c4972382639e419b3f34a38cdfca\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1907bfbad7eb34da642e30d6aa81820aab457f92347333a364ac12396b0e9dae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1907bfbad7eb34da642e30d6aa81820aab457f92347333a364ac12396b0e9dae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"le observer\\\\nW1121 14:09:07.994401 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1121 14:09:07.994601 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1121 14:09:07.995290 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1461162674/tls.crt::/tmp/serving-cert-1461162674/tls.key\\\\\\\"\\\\nI1121 14:09:08.162382 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1121 14:09:08.165301 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1121 14:09:08.165318 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1121 14:09:08.165343 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1121 14:09:08.165348 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1121 14:09:08.172015 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1121 14:09:08.172083 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1121 14:09:08.172089 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1121 14:09:08.172094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1121 14:09:08.172098 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1121 14:09:08.172102 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1121 14:09:08.172106 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1121 14:09:08.172023 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1121 14:09:08.174214 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d78c87c440c7d8e30075f6143111b3f504fb99686484b8f19a9c8d61a55eccc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://101fb592fef069af009c613a1e9c9825024ed205498ab72334211ba90d6f7397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://101fb592fef069af009c613a1e9c9825024ed205498ab72334211ba90d6f7397\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:11Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:11 crc kubenswrapper[4897]: I1121 14:09:11.678704 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:11Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:11 crc kubenswrapper[4897]: I1121 14:09:11.700455 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:11Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:11 crc kubenswrapper[4897]: I1121 14:09:11.719270 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c976d804a7ebeba2c879e920cbc389f96f7f5def4b4439e8c8980d5460ecdb65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:11Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:11 crc kubenswrapper[4897]: I1121 14:09:11.743941 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nl76g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a7d6101-d0d2-4828-b3ce-d01040c04640\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e56402f62821baf335f378084f4ed7c21b60ccbafe148419a0a9a403f106ad19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nl76g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:11Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:12 crc kubenswrapper[4897]: I1121 14:09:12.280304 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" event={"ID":"222f65f2-0b82-4760-bcf7-779244720f01","Type":"ContainerStarted","Data":"8971c38daad45bda3d61545f52120dad40992dbf2e1d6e1e776500ed58d73f36"} Nov 21 14:09:12 crc kubenswrapper[4897]: I1121 14:09:12.280863 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" event={"ID":"222f65f2-0b82-4760-bcf7-779244720f01","Type":"ContainerStarted","Data":"cfee19f82e35869a641318a0c45bc7a2ae71d80039f5790e60e277be15c9d3ec"} Nov 21 14:09:12 crc kubenswrapper[4897]: I1121 14:09:12.280881 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" event={"ID":"222f65f2-0b82-4760-bcf7-779244720f01","Type":"ContainerStarted","Data":"ccd51aa0faf3b1327f207b355cddf86a7353c0051f7e5ce0396f3c4bc67e3b16"} Nov 21 14:09:12 crc kubenswrapper[4897]: I1121 14:09:12.280896 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" event={"ID":"222f65f2-0b82-4760-bcf7-779244720f01","Type":"ContainerStarted","Data":"92f1d932a83d65ba6a9a906cc489ff51e76547d07d42656630c59421f833c4db"} Nov 21 14:09:12 crc kubenswrapper[4897]: I1121 14:09:12.280907 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" event={"ID":"222f65f2-0b82-4760-bcf7-779244720f01","Type":"ContainerStarted","Data":"ec279f2f558a8ac653006321b18cbaf13185876fe303c5cc1c65988aef09edbf"} Nov 21 14:09:12 crc kubenswrapper[4897]: I1121 14:09:12.280917 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" event={"ID":"222f65f2-0b82-4760-bcf7-779244720f01","Type":"ContainerStarted","Data":"9de095f3c760b13d5197f624eac3806a1f4186577f2b95eb522532d453eb266e"} Nov 21 14:09:12 crc kubenswrapper[4897]: I1121 14:09:12.289355 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"09877d9780caddd5f7f73e95fdf818a427ed645ed5997984113151f4b4759c13"} Nov 21 14:09:12 crc kubenswrapper[4897]: I1121 14:09:12.294203 4897 generic.go:334] "Generic (PLEG): container finished" podID="5a7d6101-d0d2-4828-b3ce-d01040c04640" containerID="e56402f62821baf335f378084f4ed7c21b60ccbafe148419a0a9a403f106ad19" exitCode=0 Nov 21 14:09:12 crc kubenswrapper[4897]: I1121 14:09:12.294280 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-nl76g" event={"ID":"5a7d6101-d0d2-4828-b3ce-d01040c04640","Type":"ContainerDied","Data":"e56402f62821baf335f378084f4ed7c21b60ccbafe148419a0a9a403f106ad19"} Nov 21 14:09:12 crc kubenswrapper[4897]: I1121 14:09:12.302563 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09877d9780caddd5f7f73e95fdf818a427ed645ed5997984113151f4b4759c13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:12Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:12 crc kubenswrapper[4897]: I1121 14:09:12.316860 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:12Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:12 crc kubenswrapper[4897]: I1121 14:09:12.334182 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6ljgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f3b70b5-9294-4f69-9528-500d28f34c89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://527cf28abdeef1b42d9c19fb88e3171dc05c4516c3a542a461b5b1daa85c6d25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78jhb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6ljgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:12Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:12 crc kubenswrapper[4897]: I1121 14:09:12.357043 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6cd330fd-e8a6-49d5-bfc9-5f6122ccd49e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1bd64b9b938bc80284bcee54141e90cad613d339c4f27ef2934b29bc2a318916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://936f59996d26e42099dcd055f87fff9e3ac4283b57e1956fe97463f2334152d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca74b28a069dc2e609b0779164c3aa470a15ed9848af0dd5fc3479aaf750157b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://270c53b9a342f26be077a01cd4e87e9ccd0eb5d2c84208ae38cbf720463245e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66c521c26d541bbddb064dd9710f1a7b41090d2f47343977922272dfc6dab7af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6eb2ac410b49098e5b3603de5e6bf5ed4484fab0b300021d0bed2e5ed8a42527\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6eb2ac410b49098e5b3603de5e6bf5ed4484fab0b300021d0bed2e5ed8a42527\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03385e0c78e4806f60c160227ef4905a11003ed3d70f0e5c0dd0a477979d1d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03385e0c78e4806f60c160227ef4905a11003ed3d70f0e5c0dd0a477979d1d4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://50aea2fa8b3b141b4330ddce2c8e0715230c6f06dab64b3abfd98a3067224bec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50aea2fa8b3b141b4330ddce2c8e0715230c6f06dab64b3abfd98a3067224bec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:12Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:12 crc kubenswrapper[4897]: I1121 14:09:12.376369 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a6c0a99-565b-41f9-a8b5-c8990a0c8eb3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://737553dd1d16e2019ea89a376e3a868f9792ed5239aa7d6cc40a05d3d4f286ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6682a8886d2b9bce069cf9261d4701ef6907f19574fd035bd89bd05626023e5c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de079956e3d8de0101f3f87dcc120ef23876c4972382639e419b3f34a38cdfca\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1907bfbad7eb34da642e30d6aa81820aab457f92347333a364ac12396b0e9dae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1907bfbad7eb34da642e30d6aa81820aab457f92347333a364ac12396b0e9dae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"le observer\\\\nW1121 14:09:07.994401 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1121 14:09:07.994601 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1121 14:09:07.995290 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1461162674/tls.crt::/tmp/serving-cert-1461162674/tls.key\\\\\\\"\\\\nI1121 14:09:08.162382 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1121 14:09:08.165301 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1121 14:09:08.165318 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1121 14:09:08.165343 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1121 14:09:08.165348 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1121 14:09:08.172015 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1121 14:09:08.172083 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1121 14:09:08.172089 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1121 14:09:08.172094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1121 14:09:08.172098 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1121 14:09:08.172102 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1121 14:09:08.172106 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1121 14:09:08.172023 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1121 14:09:08.174214 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d78c87c440c7d8e30075f6143111b3f504fb99686484b8f19a9c8d61a55eccc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://101fb592fef069af009c613a1e9c9825024ed205498ab72334211ba90d6f7397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://101fb592fef069af009c613a1e9c9825024ed205498ab72334211ba90d6f7397\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:12Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:12 crc kubenswrapper[4897]: I1121 14:09:12.393764 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:12Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:12 crc kubenswrapper[4897]: I1121 14:09:12.408134 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c976d804a7ebeba2c879e920cbc389f96f7f5def4b4439e8c8980d5460ecdb65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:12Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:12 crc kubenswrapper[4897]: I1121 14:09:12.425075 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nl76g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a7d6101-d0d2-4828-b3ce-d01040c04640\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e56402f62821baf335f378084f4ed7c21b60ccbafe148419a0a9a403f106ad19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nl76g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:12Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:12 crc kubenswrapper[4897]: I1121 14:09:12.441737 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48c6b1ea7497b33c00b956e400426c1d4f3c6c66e0225133434460800eef9314\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ea6068a6b6f8123b402de354e5374a0f02a7191a922d1718ab38f8898f47c52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:12Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:12 crc kubenswrapper[4897]: I1121 14:09:12.454871 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4qwqj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f46e2ba8-9ee0-406d-a668-2ba424618b69\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f43aa89d4d9defcb50d0dc6f90293b5f426b9cfc398b780c2e10f6a5c9f1e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbsgb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4qwqj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:12Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:12 crc kubenswrapper[4897]: I1121 14:09:12.480896 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"222f65f2-0b82-4760-bcf7-779244720f01\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7877db6f5c2bf0eb31218c64d67bca48f921ff7ac62b4e495f29024c086bf738\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7877db6f5c2bf0eb31218c64d67bca48f921ff7ac62b4e495f29024c086bf738\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tpfb6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:12Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:12 crc kubenswrapper[4897]: I1121 14:09:12.494565 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:12Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:12 crc kubenswrapper[4897]: I1121 14:09:12.514967 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7670227-d280-4847-b882-754429f56b0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://961961696b8750c248598f4cda032b18395e5c802275dcfe6d3e9f7ef3510f1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5854028821e6a360a259427b68e7bb4c2a7220e0c68f4cc91639b2ece0540a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-krv5b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:12Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:12 crc kubenswrapper[4897]: I1121 14:09:12.552829 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:12Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:12 crc kubenswrapper[4897]: I1121 14:09:12.578440 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7670227-d280-4847-b882-754429f56b0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://961961696b8750c248598f4cda032b18395e5c802275dcfe6d3e9f7ef3510f1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5854028821e6a360a259427b68e7bb4c2a7220e0c68f4cc91639b2ece0540a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-krv5b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:12Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:12 crc kubenswrapper[4897]: I1121 14:09:12.598406 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"222f65f2-0b82-4760-bcf7-779244720f01\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7877db6f5c2bf0eb31218c64d67bca48f921ff7ac62b4e495f29024c086bf738\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7877db6f5c2bf0eb31218c64d67bca48f921ff7ac62b4e495f29024c086bf738\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tpfb6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:12Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:12 crc kubenswrapper[4897]: I1121 14:09:12.619274 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6cd330fd-e8a6-49d5-bfc9-5f6122ccd49e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1bd64b9b938bc80284bcee54141e90cad613d339c4f27ef2934b29bc2a318916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://936f59996d26e42099dcd055f87fff9e3ac4283b57e1956fe97463f2334152d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca74b28a069dc2e609b0779164c3aa470a15ed9848af0dd5fc3479aaf750157b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://270c53b9a342f26be077a01cd4e87e9ccd0eb5d2c84208ae38cbf720463245e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66c521c26d541bbddb064dd9710f1a7b41090d2f47343977922272dfc6dab7af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6eb2ac410b49098e5b3603de5e6bf5ed4484fab0b300021d0bed2e5ed8a42527\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6eb2ac410b49098e5b3603de5e6bf5ed4484fab0b300021d0bed2e5ed8a42527\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03385e0c78e4806f60c160227ef4905a11003ed3d70f0e5c0dd0a477979d1d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03385e0c78e4806f60c160227ef4905a11003ed3d70f0e5c0dd0a477979d1d4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://50aea2fa8b3b141b4330ddce2c8e0715230c6f06dab64b3abfd98a3067224bec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50aea2fa8b3b141b4330ddce2c8e0715230c6f06dab64b3abfd98a3067224bec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:12Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:12 crc kubenswrapper[4897]: I1121 14:09:12.635417 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a6c0a99-565b-41f9-a8b5-c8990a0c8eb3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://737553dd1d16e2019ea89a376e3a868f9792ed5239aa7d6cc40a05d3d4f286ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6682a8886d2b9bce069cf9261d4701ef6907f19574fd035bd89bd05626023e5c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de079956e3d8de0101f3f87dcc120ef23876c4972382639e419b3f34a38cdfca\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1907bfbad7eb34da642e30d6aa81820aab457f92347333a364ac12396b0e9dae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1907bfbad7eb34da642e30d6aa81820aab457f92347333a364ac12396b0e9dae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"le observer\\\\nW1121 14:09:07.994401 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1121 14:09:07.994601 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1121 14:09:07.995290 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1461162674/tls.crt::/tmp/serving-cert-1461162674/tls.key\\\\\\\"\\\\nI1121 14:09:08.162382 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1121 14:09:08.165301 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1121 14:09:08.165318 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1121 14:09:08.165343 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1121 14:09:08.165348 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1121 14:09:08.172015 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1121 14:09:08.172083 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1121 14:09:08.172089 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1121 14:09:08.172094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1121 14:09:08.172098 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1121 14:09:08.172102 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1121 14:09:08.172106 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1121 14:09:08.172023 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1121 14:09:08.174214 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d78c87c440c7d8e30075f6143111b3f504fb99686484b8f19a9c8d61a55eccc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://101fb592fef069af009c613a1e9c9825024ed205498ab72334211ba90d6f7397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://101fb592fef069af009c613a1e9c9825024ed205498ab72334211ba90d6f7397\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:12Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:12 crc kubenswrapper[4897]: I1121 14:09:12.650760 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09877d9780caddd5f7f73e95fdf818a427ed645ed5997984113151f4b4759c13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:12Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:12 crc kubenswrapper[4897]: I1121 14:09:12.666889 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:12Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:12 crc kubenswrapper[4897]: I1121 14:09:12.671107 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 21 14:09:12 crc kubenswrapper[4897]: E1121 14:09:12.671298 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-21 14:09:16.6712702 +0000 UTC m=+33.955863675 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:09:12 crc kubenswrapper[4897]: I1121 14:09:12.671451 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 21 14:09:12 crc kubenswrapper[4897]: I1121 14:09:12.671513 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 21 14:09:12 crc kubenswrapper[4897]: E1121 14:09:12.671627 4897 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 21 14:09:12 crc kubenswrapper[4897]: E1121 14:09:12.671656 4897 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 21 14:09:12 crc kubenswrapper[4897]: E1121 14:09:12.671684 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-21 14:09:16.671676 +0000 UTC m=+33.956269475 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 21 14:09:12 crc kubenswrapper[4897]: E1121 14:09:12.671776 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-21 14:09:16.671751912 +0000 UTC m=+33.956345527 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 21 14:09:12 crc kubenswrapper[4897]: I1121 14:09:12.681207 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6ljgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f3b70b5-9294-4f69-9528-500d28f34c89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://527cf28abdeef1b42d9c19fb88e3171dc05c4516c3a542a461b5b1daa85c6d25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78jhb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6ljgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:12Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:12 crc kubenswrapper[4897]: I1121 14:09:12.702888 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:12Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:12 crc kubenswrapper[4897]: I1121 14:09:12.717360 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c976d804a7ebeba2c879e920cbc389f96f7f5def4b4439e8c8980d5460ecdb65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:12Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:12 crc kubenswrapper[4897]: I1121 14:09:12.738684 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nl76g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a7d6101-d0d2-4828-b3ce-d01040c04640\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e56402f62821baf335f378084f4ed7c21b60ccbafe148419a0a9a403f106ad19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e56402f62821baf335f378084f4ed7c21b60ccbafe148419a0a9a403f106ad19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nl76g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:12Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:12 crc kubenswrapper[4897]: I1121 14:09:12.754142 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48c6b1ea7497b33c00b956e400426c1d4f3c6c66e0225133434460800eef9314\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ea6068a6b6f8123b402de354e5374a0f02a7191a922d1718ab38f8898f47c52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:12Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:12 crc kubenswrapper[4897]: I1121 14:09:12.765265 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4qwqj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f46e2ba8-9ee0-406d-a668-2ba424618b69\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f43aa89d4d9defcb50d0dc6f90293b5f426b9cfc398b780c2e10f6a5c9f1e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbsgb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4qwqj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:12Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:12 crc kubenswrapper[4897]: I1121 14:09:12.772965 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 21 14:09:12 crc kubenswrapper[4897]: I1121 14:09:12.773107 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 21 14:09:12 crc kubenswrapper[4897]: E1121 14:09:12.773220 4897 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 21 14:09:12 crc kubenswrapper[4897]: E1121 14:09:12.773270 4897 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 21 14:09:12 crc kubenswrapper[4897]: E1121 14:09:12.773285 4897 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 21 14:09:12 crc kubenswrapper[4897]: E1121 14:09:12.773300 4897 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 21 14:09:12 crc kubenswrapper[4897]: E1121 14:09:12.773328 4897 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 21 14:09:12 crc kubenswrapper[4897]: E1121 14:09:12.773344 4897 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 21 14:09:12 crc kubenswrapper[4897]: E1121 14:09:12.773362 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-21 14:09:16.773336689 +0000 UTC m=+34.057930164 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 21 14:09:12 crc kubenswrapper[4897]: E1121 14:09:12.773410 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-21 14:09:16.773389341 +0000 UTC m=+34.057982996 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 21 14:09:13 crc kubenswrapper[4897]: I1121 14:09:13.088074 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 21 14:09:13 crc kubenswrapper[4897]: I1121 14:09:13.088108 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 21 14:09:13 crc kubenswrapper[4897]: I1121 14:09:13.088218 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 21 14:09:13 crc kubenswrapper[4897]: E1121 14:09:13.088228 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 21 14:09:13 crc kubenswrapper[4897]: E1121 14:09:13.088340 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 21 14:09:13 crc kubenswrapper[4897]: E1121 14:09:13.088480 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 21 14:09:13 crc kubenswrapper[4897]: I1121 14:09:13.117070 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 21 14:09:13 crc kubenswrapper[4897]: I1121 14:09:13.120644 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 21 14:09:13 crc kubenswrapper[4897]: I1121 14:09:13.127856 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Nov 21 14:09:13 crc kubenswrapper[4897]: I1121 14:09:13.130696 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c976d804a7ebeba2c879e920cbc389f96f7f5def4b4439e8c8980d5460ecdb65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:13Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:13 crc kubenswrapper[4897]: I1121 14:09:13.149411 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nl76g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a7d6101-d0d2-4828-b3ce-d01040c04640\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e56402f62821baf335f378084f4ed7c21b60ccbafe148419a0a9a403f106ad19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e56402f62821baf335f378084f4ed7c21b60ccbafe148419a0a9a403f106ad19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nl76g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:13Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:13 crc kubenswrapper[4897]: I1121 14:09:13.164711 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:13Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:13 crc kubenswrapper[4897]: I1121 14:09:13.181404 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48c6b1ea7497b33c00b956e400426c1d4f3c6c66e0225133434460800eef9314\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ea6068a6b6f8123b402de354e5374a0f02a7191a922d1718ab38f8898f47c52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:13Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:13 crc kubenswrapper[4897]: I1121 14:09:13.194373 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4qwqj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f46e2ba8-9ee0-406d-a668-2ba424618b69\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f43aa89d4d9defcb50d0dc6f90293b5f426b9cfc398b780c2e10f6a5c9f1e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbsgb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4qwqj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:13Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:13 crc kubenswrapper[4897]: I1121 14:09:13.208260 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:13Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:13 crc kubenswrapper[4897]: I1121 14:09:13.223211 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7670227-d280-4847-b882-754429f56b0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://961961696b8750c248598f4cda032b18395e5c802275dcfe6d3e9f7ef3510f1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5854028821e6a360a259427b68e7bb4c2a7220e0c68f4cc91639b2ece0540a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-krv5b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:13Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:13 crc kubenswrapper[4897]: I1121 14:09:13.245030 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"222f65f2-0b82-4760-bcf7-779244720f01\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7877db6f5c2bf0eb31218c64d67bca48f921ff7ac62b4e495f29024c086bf738\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7877db6f5c2bf0eb31218c64d67bca48f921ff7ac62b4e495f29024c086bf738\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tpfb6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:13Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:13 crc kubenswrapper[4897]: I1121 14:09:13.264707 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a6c0a99-565b-41f9-a8b5-c8990a0c8eb3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://737553dd1d16e2019ea89a376e3a868f9792ed5239aa7d6cc40a05d3d4f286ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6682a8886d2b9bce069cf9261d4701ef6907f19574fd035bd89bd05626023e5c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de079956e3d8de0101f3f87dcc120ef23876c4972382639e419b3f34a38cdfca\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1907bfbad7eb34da642e30d6aa81820aab457f92347333a364ac12396b0e9dae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1907bfbad7eb34da642e30d6aa81820aab457f92347333a364ac12396b0e9dae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"le observer\\\\nW1121 14:09:07.994401 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1121 14:09:07.994601 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1121 14:09:07.995290 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1461162674/tls.crt::/tmp/serving-cert-1461162674/tls.key\\\\\\\"\\\\nI1121 14:09:08.162382 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1121 14:09:08.165301 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1121 14:09:08.165318 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1121 14:09:08.165343 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1121 14:09:08.165348 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1121 14:09:08.172015 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1121 14:09:08.172083 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1121 14:09:08.172089 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1121 14:09:08.172094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1121 14:09:08.172098 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1121 14:09:08.172102 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1121 14:09:08.172106 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1121 14:09:08.172023 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1121 14:09:08.174214 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d78c87c440c7d8e30075f6143111b3f504fb99686484b8f19a9c8d61a55eccc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://101fb592fef069af009c613a1e9c9825024ed205498ab72334211ba90d6f7397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://101fb592fef069af009c613a1e9c9825024ed205498ab72334211ba90d6f7397\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:13Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:13 crc kubenswrapper[4897]: I1121 14:09:13.280162 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09877d9780caddd5f7f73e95fdf818a427ed645ed5997984113151f4b4759c13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:13Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:13 crc kubenswrapper[4897]: I1121 14:09:13.300599 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:13Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:13 crc kubenswrapper[4897]: I1121 14:09:13.300812 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-nl76g" event={"ID":"5a7d6101-d0d2-4828-b3ce-d01040c04640","Type":"ContainerStarted","Data":"b6022ff28e83dafcd846a7f03b090262aa3229a0da89f1984e4b6a054a04ac24"} Nov 21 14:09:13 crc kubenswrapper[4897]: E1121 14:09:13.309019 4897 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-crc\" already exists" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 21 14:09:13 crc kubenswrapper[4897]: I1121 14:09:13.316438 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6ljgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f3b70b5-9294-4f69-9528-500d28f34c89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://527cf28abdeef1b42d9c19fb88e3171dc05c4516c3a542a461b5b1daa85c6d25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78jhb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6ljgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:13Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:13 crc kubenswrapper[4897]: I1121 14:09:13.338458 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6cd330fd-e8a6-49d5-bfc9-5f6122ccd49e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1bd64b9b938bc80284bcee54141e90cad613d339c4f27ef2934b29bc2a318916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://936f59996d26e42099dcd055f87fff9e3ac4283b57e1956fe97463f2334152d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca74b28a069dc2e609b0779164c3aa470a15ed9848af0dd5fc3479aaf750157b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://270c53b9a342f26be077a01cd4e87e9ccd0eb5d2c84208ae38cbf720463245e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66c521c26d541bbddb064dd9710f1a7b41090d2f47343977922272dfc6dab7af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6eb2ac410b49098e5b3603de5e6bf5ed4484fab0b300021d0bed2e5ed8a42527\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6eb2ac410b49098e5b3603de5e6bf5ed4484fab0b300021d0bed2e5ed8a42527\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03385e0c78e4806f60c160227ef4905a11003ed3d70f0e5c0dd0a477979d1d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03385e0c78e4806f60c160227ef4905a11003ed3d70f0e5c0dd0a477979d1d4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://50aea2fa8b3b141b4330ddce2c8e0715230c6f06dab64b3abfd98a3067224bec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50aea2fa8b3b141b4330ddce2c8e0715230c6f06dab64b3abfd98a3067224bec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:13Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:13 crc kubenswrapper[4897]: I1121 14:09:13.351452 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b78c993a-9966-4e84-85a1-b4f73aaf17f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce3a2602338a1c1e9380ee55f78360f3d9cf6ccc62678886dc5ae574a72e3d9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c256823f2fb49ecc121f6ef564cd86196f9b4a473e9bdea670173596ef14a823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c65f71f8746044f5d41687c670a0308cd09dc41a6531f9b5720e852b3d240756\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1dbe88e7e6ad63899e107ebb884a16edb9a42b21ae7c86d7fbfaf89785b0d36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:13Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:13 crc kubenswrapper[4897]: I1121 14:09:13.368258 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09877d9780caddd5f7f73e95fdf818a427ed645ed5997984113151f4b4759c13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:13Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:13 crc kubenswrapper[4897]: I1121 14:09:13.382304 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:13Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:13 crc kubenswrapper[4897]: I1121 14:09:13.397670 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6ljgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f3b70b5-9294-4f69-9528-500d28f34c89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://527cf28abdeef1b42d9c19fb88e3171dc05c4516c3a542a461b5b1daa85c6d25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78jhb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6ljgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:13Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:13 crc kubenswrapper[4897]: I1121 14:09:13.421462 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6cd330fd-e8a6-49d5-bfc9-5f6122ccd49e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1bd64b9b938bc80284bcee54141e90cad613d339c4f27ef2934b29bc2a318916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://936f59996d26e42099dcd055f87fff9e3ac4283b57e1956fe97463f2334152d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca74b28a069dc2e609b0779164c3aa470a15ed9848af0dd5fc3479aaf750157b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://270c53b9a342f26be077a01cd4e87e9ccd0eb5d2c84208ae38cbf720463245e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66c521c26d541bbddb064dd9710f1a7b41090d2f47343977922272dfc6dab7af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6eb2ac410b49098e5b3603de5e6bf5ed4484fab0b300021d0bed2e5ed8a42527\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6eb2ac410b49098e5b3603de5e6bf5ed4484fab0b300021d0bed2e5ed8a42527\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03385e0c78e4806f60c160227ef4905a11003ed3d70f0e5c0dd0a477979d1d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03385e0c78e4806f60c160227ef4905a11003ed3d70f0e5c0dd0a477979d1d4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://50aea2fa8b3b141b4330ddce2c8e0715230c6f06dab64b3abfd98a3067224bec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50aea2fa8b3b141b4330ddce2c8e0715230c6f06dab64b3abfd98a3067224bec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:13Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:13 crc kubenswrapper[4897]: I1121 14:09:13.439066 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a6c0a99-565b-41f9-a8b5-c8990a0c8eb3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://737553dd1d16e2019ea89a376e3a868f9792ed5239aa7d6cc40a05d3d4f286ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6682a8886d2b9bce069cf9261d4701ef6907f19574fd035bd89bd05626023e5c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de079956e3d8de0101f3f87dcc120ef23876c4972382639e419b3f34a38cdfca\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1907bfbad7eb34da642e30d6aa81820aab457f92347333a364ac12396b0e9dae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1907bfbad7eb34da642e30d6aa81820aab457f92347333a364ac12396b0e9dae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"le observer\\\\nW1121 14:09:07.994401 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1121 14:09:07.994601 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1121 14:09:07.995290 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1461162674/tls.crt::/tmp/serving-cert-1461162674/tls.key\\\\\\\"\\\\nI1121 14:09:08.162382 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1121 14:09:08.165301 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1121 14:09:08.165318 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1121 14:09:08.165343 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1121 14:09:08.165348 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1121 14:09:08.172015 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1121 14:09:08.172083 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1121 14:09:08.172089 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1121 14:09:08.172094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1121 14:09:08.172098 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1121 14:09:08.172102 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1121 14:09:08.172106 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1121 14:09:08.172023 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1121 14:09:08.174214 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d78c87c440c7d8e30075f6143111b3f504fb99686484b8f19a9c8d61a55eccc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://101fb592fef069af009c613a1e9c9825024ed205498ab72334211ba90d6f7397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://101fb592fef069af009c613a1e9c9825024ed205498ab72334211ba90d6f7397\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:13Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:13 crc kubenswrapper[4897]: I1121 14:09:13.455275 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nl76g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a7d6101-d0d2-4828-b3ce-d01040c04640\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e56402f62821baf335f378084f4ed7c21b60ccbafe148419a0a9a403f106ad19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e56402f62821baf335f378084f4ed7c21b60ccbafe148419a0a9a403f106ad19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6022ff28e83dafcd846a7f03b090262aa3229a0da89f1984e4b6a054a04ac24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nl76g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:13Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:13 crc kubenswrapper[4897]: I1121 14:09:13.470011 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:13Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:13 crc kubenswrapper[4897]: I1121 14:09:13.484662 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c976d804a7ebeba2c879e920cbc389f96f7f5def4b4439e8c8980d5460ecdb65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:13Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:13 crc kubenswrapper[4897]: I1121 14:09:13.496618 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4qwqj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f46e2ba8-9ee0-406d-a668-2ba424618b69\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f43aa89d4d9defcb50d0dc6f90293b5f426b9cfc398b780c2e10f6a5c9f1e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbsgb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4qwqj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:13Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:13 crc kubenswrapper[4897]: I1121 14:09:13.512808 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48c6b1ea7497b33c00b956e400426c1d4f3c6c66e0225133434460800eef9314\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ea6068a6b6f8123b402de354e5374a0f02a7191a922d1718ab38f8898f47c52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:13Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:13 crc kubenswrapper[4897]: I1121 14:09:13.553800 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7670227-d280-4847-b882-754429f56b0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://961961696b8750c248598f4cda032b18395e5c802275dcfe6d3e9f7ef3510f1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5854028821e6a360a259427b68e7bb4c2a7220e0c68f4cc91639b2ece0540a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-krv5b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:13Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:13 crc kubenswrapper[4897]: I1121 14:09:13.599191 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"222f65f2-0b82-4760-bcf7-779244720f01\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7877db6f5c2bf0eb31218c64d67bca48f921ff7ac62b4e495f29024c086bf738\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7877db6f5c2bf0eb31218c64d67bca48f921ff7ac62b4e495f29024c086bf738\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tpfb6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:13Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:13 crc kubenswrapper[4897]: I1121 14:09:13.633409 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:13Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:14 crc kubenswrapper[4897]: I1121 14:09:14.105724 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c976d804a7ebeba2c879e920cbc389f96f7f5def4b4439e8c8980d5460ecdb65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:14Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:14 crc kubenswrapper[4897]: I1121 14:09:14.124564 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nl76g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a7d6101-d0d2-4828-b3ce-d01040c04640\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e56402f62821baf335f378084f4ed7c21b60ccbafe148419a0a9a403f106ad19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e56402f62821baf335f378084f4ed7c21b60ccbafe148419a0a9a403f106ad19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6022ff28e83dafcd846a7f03b090262aa3229a0da89f1984e4b6a054a04ac24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nl76g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:14Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:14 crc kubenswrapper[4897]: I1121 14:09:14.141754 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:14Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:14 crc kubenswrapper[4897]: I1121 14:09:14.159030 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48c6b1ea7497b33c00b956e400426c1d4f3c6c66e0225133434460800eef9314\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ea6068a6b6f8123b402de354e5374a0f02a7191a922d1718ab38f8898f47c52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:14Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:14 crc kubenswrapper[4897]: I1121 14:09:14.172487 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4qwqj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f46e2ba8-9ee0-406d-a668-2ba424618b69\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f43aa89d4d9defcb50d0dc6f90293b5f426b9cfc398b780c2e10f6a5c9f1e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbsgb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4qwqj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:14Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:14 crc kubenswrapper[4897]: I1121 14:09:14.189977 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:14Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:14 crc kubenswrapper[4897]: I1121 14:09:14.204220 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7670227-d280-4847-b882-754429f56b0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://961961696b8750c248598f4cda032b18395e5c802275dcfe6d3e9f7ef3510f1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5854028821e6a360a259427b68e7bb4c2a7220e0c68f4cc91639b2ece0540a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-krv5b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:14Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:14 crc kubenswrapper[4897]: I1121 14:09:14.229419 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"222f65f2-0b82-4760-bcf7-779244720f01\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7877db6f5c2bf0eb31218c64d67bca48f921ff7ac62b4e495f29024c086bf738\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7877db6f5c2bf0eb31218c64d67bca48f921ff7ac62b4e495f29024c086bf738\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tpfb6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:14Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:14 crc kubenswrapper[4897]: I1121 14:09:14.246936 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a6c0a99-565b-41f9-a8b5-c8990a0c8eb3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://737553dd1d16e2019ea89a376e3a868f9792ed5239aa7d6cc40a05d3d4f286ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6682a8886d2b9bce069cf9261d4701ef6907f19574fd035bd89bd05626023e5c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de079956e3d8de0101f3f87dcc120ef23876c4972382639e419b3f34a38cdfca\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1907bfbad7eb34da642e30d6aa81820aab457f92347333a364ac12396b0e9dae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1907bfbad7eb34da642e30d6aa81820aab457f92347333a364ac12396b0e9dae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"le observer\\\\nW1121 14:09:07.994401 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1121 14:09:07.994601 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1121 14:09:07.995290 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1461162674/tls.crt::/tmp/serving-cert-1461162674/tls.key\\\\\\\"\\\\nI1121 14:09:08.162382 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1121 14:09:08.165301 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1121 14:09:08.165318 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1121 14:09:08.165343 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1121 14:09:08.165348 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1121 14:09:08.172015 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1121 14:09:08.172083 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1121 14:09:08.172089 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1121 14:09:08.172094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1121 14:09:08.172098 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1121 14:09:08.172102 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1121 14:09:08.172106 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1121 14:09:08.172023 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1121 14:09:08.174214 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d78c87c440c7d8e30075f6143111b3f504fb99686484b8f19a9c8d61a55eccc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://101fb592fef069af009c613a1e9c9825024ed205498ab72334211ba90d6f7397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://101fb592fef069af009c613a1e9c9825024ed205498ab72334211ba90d6f7397\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:14Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:14 crc kubenswrapper[4897]: I1121 14:09:14.261163 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b78c993a-9966-4e84-85a1-b4f73aaf17f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce3a2602338a1c1e9380ee55f78360f3d9cf6ccc62678886dc5ae574a72e3d9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c256823f2fb49ecc121f6ef564cd86196f9b4a473e9bdea670173596ef14a823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c65f71f8746044f5d41687c670a0308cd09dc41a6531f9b5720e852b3d240756\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1dbe88e7e6ad63899e107ebb884a16edb9a42b21ae7c86d7fbfaf89785b0d36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:14Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:14 crc kubenswrapper[4897]: I1121 14:09:14.278026 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09877d9780caddd5f7f73e95fdf818a427ed645ed5997984113151f4b4759c13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:14Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:14 crc kubenswrapper[4897]: I1121 14:09:14.294516 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:14Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:14 crc kubenswrapper[4897]: I1121 14:09:14.313167 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6ljgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f3b70b5-9294-4f69-9528-500d28f34c89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://527cf28abdeef1b42d9c19fb88e3171dc05c4516c3a542a461b5b1daa85c6d25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78jhb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6ljgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:14Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:14 crc kubenswrapper[4897]: I1121 14:09:14.335119 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" event={"ID":"222f65f2-0b82-4760-bcf7-779244720f01","Type":"ContainerStarted","Data":"b486a86e0e97ab409e590f30a1bf5cb9f584a7c67f96ff1236e42ca4f8675b56"} Nov 21 14:09:14 crc kubenswrapper[4897]: I1121 14:09:14.336045 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6cd330fd-e8a6-49d5-bfc9-5f6122ccd49e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1bd64b9b938bc80284bcee54141e90cad613d339c4f27ef2934b29bc2a318916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://936f59996d26e42099dcd055f87fff9e3ac4283b57e1956fe97463f2334152d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca74b28a069dc2e609b0779164c3aa470a15ed9848af0dd5fc3479aaf750157b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://270c53b9a342f26be077a01cd4e87e9ccd0eb5d2c84208ae38cbf720463245e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66c521c26d541bbddb064dd9710f1a7b41090d2f47343977922272dfc6dab7af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6eb2ac410b49098e5b3603de5e6bf5ed4484fab0b300021d0bed2e5ed8a42527\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6eb2ac410b49098e5b3603de5e6bf5ed4484fab0b300021d0bed2e5ed8a42527\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03385e0c78e4806f60c160227ef4905a11003ed3d70f0e5c0dd0a477979d1d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03385e0c78e4806f60c160227ef4905a11003ed3d70f0e5c0dd0a477979d1d4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://50aea2fa8b3b141b4330ddce2c8e0715230c6f06dab64b3abfd98a3067224bec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50aea2fa8b3b141b4330ddce2c8e0715230c6f06dab64b3abfd98a3067224bec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:14Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:14 crc kubenswrapper[4897]: I1121 14:09:14.336979 4897 generic.go:334] "Generic (PLEG): container finished" podID="5a7d6101-d0d2-4828-b3ce-d01040c04640" containerID="b6022ff28e83dafcd846a7f03b090262aa3229a0da89f1984e4b6a054a04ac24" exitCode=0 Nov 21 14:09:14 crc kubenswrapper[4897]: I1121 14:09:14.337011 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-nl76g" event={"ID":"5a7d6101-d0d2-4828-b3ce-d01040c04640","Type":"ContainerDied","Data":"b6022ff28e83dafcd846a7f03b090262aa3229a0da89f1984e4b6a054a04ac24"} Nov 21 14:09:14 crc kubenswrapper[4897]: I1121 14:09:14.352753 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:14Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:14 crc kubenswrapper[4897]: I1121 14:09:14.368707 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7670227-d280-4847-b882-754429f56b0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://961961696b8750c248598f4cda032b18395e5c802275dcfe6d3e9f7ef3510f1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5854028821e6a360a259427b68e7bb4c2a7220e0c68f4cc91639b2ece0540a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-krv5b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:14Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:14 crc kubenswrapper[4897]: I1121 14:09:14.392055 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"222f65f2-0b82-4760-bcf7-779244720f01\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7877db6f5c2bf0eb31218c64d67bca48f921ff7ac62b4e495f29024c086bf738\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7877db6f5c2bf0eb31218c64d67bca48f921ff7ac62b4e495f29024c086bf738\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tpfb6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:14Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:14 crc kubenswrapper[4897]: I1121 14:09:14.408952 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6ljgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f3b70b5-9294-4f69-9528-500d28f34c89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://527cf28abdeef1b42d9c19fb88e3171dc05c4516c3a542a461b5b1daa85c6d25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78jhb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6ljgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:14Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:14 crc kubenswrapper[4897]: I1121 14:09:14.432801 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6cd330fd-e8a6-49d5-bfc9-5f6122ccd49e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1bd64b9b938bc80284bcee54141e90cad613d339c4f27ef2934b29bc2a318916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://936f59996d26e42099dcd055f87fff9e3ac4283b57e1956fe97463f2334152d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca74b28a069dc2e609b0779164c3aa470a15ed9848af0dd5fc3479aaf750157b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://270c53b9a342f26be077a01cd4e87e9ccd0eb5d2c84208ae38cbf720463245e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66c521c26d541bbddb064dd9710f1a7b41090d2f47343977922272dfc6dab7af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6eb2ac410b49098e5b3603de5e6bf5ed4484fab0b300021d0bed2e5ed8a42527\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6eb2ac410b49098e5b3603de5e6bf5ed4484fab0b300021d0bed2e5ed8a42527\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03385e0c78e4806f60c160227ef4905a11003ed3d70f0e5c0dd0a477979d1d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03385e0c78e4806f60c160227ef4905a11003ed3d70f0e5c0dd0a477979d1d4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://50aea2fa8b3b141b4330ddce2c8e0715230c6f06dab64b3abfd98a3067224bec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50aea2fa8b3b141b4330ddce2c8e0715230c6f06dab64b3abfd98a3067224bec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:14Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:14 crc kubenswrapper[4897]: I1121 14:09:14.446344 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a6c0a99-565b-41f9-a8b5-c8990a0c8eb3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://737553dd1d16e2019ea89a376e3a868f9792ed5239aa7d6cc40a05d3d4f286ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6682a8886d2b9bce069cf9261d4701ef6907f19574fd035bd89bd05626023e5c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de079956e3d8de0101f3f87dcc120ef23876c4972382639e419b3f34a38cdfca\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1907bfbad7eb34da642e30d6aa81820aab457f92347333a364ac12396b0e9dae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1907bfbad7eb34da642e30d6aa81820aab457f92347333a364ac12396b0e9dae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"le observer\\\\nW1121 14:09:07.994401 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1121 14:09:07.994601 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1121 14:09:07.995290 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1461162674/tls.crt::/tmp/serving-cert-1461162674/tls.key\\\\\\\"\\\\nI1121 14:09:08.162382 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1121 14:09:08.165301 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1121 14:09:08.165318 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1121 14:09:08.165343 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1121 14:09:08.165348 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1121 14:09:08.172015 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1121 14:09:08.172083 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1121 14:09:08.172089 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1121 14:09:08.172094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1121 14:09:08.172098 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1121 14:09:08.172102 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1121 14:09:08.172106 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1121 14:09:08.172023 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1121 14:09:08.174214 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d78c87c440c7d8e30075f6143111b3f504fb99686484b8f19a9c8d61a55eccc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://101fb592fef069af009c613a1e9c9825024ed205498ab72334211ba90d6f7397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://101fb592fef069af009c613a1e9c9825024ed205498ab72334211ba90d6f7397\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:14Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:14 crc kubenswrapper[4897]: I1121 14:09:14.474151 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b78c993a-9966-4e84-85a1-b4f73aaf17f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce3a2602338a1c1e9380ee55f78360f3d9cf6ccc62678886dc5ae574a72e3d9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c256823f2fb49ecc121f6ef564cd86196f9b4a473e9bdea670173596ef14a823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c65f71f8746044f5d41687c670a0308cd09dc41a6531f9b5720e852b3d240756\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1dbe88e7e6ad63899e107ebb884a16edb9a42b21ae7c86d7fbfaf89785b0d36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:14Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:14 crc kubenswrapper[4897]: I1121 14:09:14.513156 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09877d9780caddd5f7f73e95fdf818a427ed645ed5997984113151f4b4759c13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:14Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:14 crc kubenswrapper[4897]: I1121 14:09:14.554647 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:14Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:14 crc kubenswrapper[4897]: I1121 14:09:14.594094 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:14Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:14 crc kubenswrapper[4897]: I1121 14:09:14.636422 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c976d804a7ebeba2c879e920cbc389f96f7f5def4b4439e8c8980d5460ecdb65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:14Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:14 crc kubenswrapper[4897]: I1121 14:09:14.676349 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nl76g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a7d6101-d0d2-4828-b3ce-d01040c04640\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e56402f62821baf335f378084f4ed7c21b60ccbafe148419a0a9a403f106ad19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e56402f62821baf335f378084f4ed7c21b60ccbafe148419a0a9a403f106ad19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6022ff28e83dafcd846a7f03b090262aa3229a0da89f1984e4b6a054a04ac24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6022ff28e83dafcd846a7f03b090262aa3229a0da89f1984e4b6a054a04ac24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nl76g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:14Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:14 crc kubenswrapper[4897]: I1121 14:09:14.718282 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48c6b1ea7497b33c00b956e400426c1d4f3c6c66e0225133434460800eef9314\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ea6068a6b6f8123b402de354e5374a0f02a7191a922d1718ab38f8898f47c52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:14Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:14 crc kubenswrapper[4897]: I1121 14:09:14.759988 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4qwqj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f46e2ba8-9ee0-406d-a668-2ba424618b69\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f43aa89d4d9defcb50d0dc6f90293b5f426b9cfc398b780c2e10f6a5c9f1e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbsgb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4qwqj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:14Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:14 crc kubenswrapper[4897]: I1121 14:09:14.985025 4897 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 21 14:09:14 crc kubenswrapper[4897]: I1121 14:09:14.989122 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:14 crc kubenswrapper[4897]: I1121 14:09:14.989168 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:14 crc kubenswrapper[4897]: I1121 14:09:14.989181 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:14 crc kubenswrapper[4897]: I1121 14:09:14.989314 4897 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 21 14:09:14 crc kubenswrapper[4897]: I1121 14:09:14.997307 4897 kubelet_node_status.go:115] "Node was previously registered" node="crc" Nov 21 14:09:14 crc kubenswrapper[4897]: I1121 14:09:14.997690 4897 kubelet_node_status.go:79] "Successfully registered node" node="crc" Nov 21 14:09:14 crc kubenswrapper[4897]: I1121 14:09:14.999221 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:14 crc kubenswrapper[4897]: I1121 14:09:14.999288 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:14 crc kubenswrapper[4897]: I1121 14:09:14.999308 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:14 crc kubenswrapper[4897]: I1121 14:09:14.999338 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:14 crc kubenswrapper[4897]: I1121 14:09:14.999359 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:14Z","lastTransitionTime":"2025-11-21T14:09:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:15 crc kubenswrapper[4897]: E1121 14:09:15.014552 4897 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8b8bdf9e-0fef-41ac-98ad-710dc2137183\\\",\\\"systemUUID\\\":\\\"8a210af6-a71f-4ea9-90a3-aac32fb868b7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:15Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:15 crc kubenswrapper[4897]: I1121 14:09:15.018990 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:15 crc kubenswrapper[4897]: I1121 14:09:15.019023 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:15 crc kubenswrapper[4897]: I1121 14:09:15.019032 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:15 crc kubenswrapper[4897]: I1121 14:09:15.019049 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:15 crc kubenswrapper[4897]: I1121 14:09:15.019060 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:15Z","lastTransitionTime":"2025-11-21T14:09:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:15 crc kubenswrapper[4897]: E1121 14:09:15.035581 4897 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8b8bdf9e-0fef-41ac-98ad-710dc2137183\\\",\\\"systemUUID\\\":\\\"8a210af6-a71f-4ea9-90a3-aac32fb868b7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:15Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:15 crc kubenswrapper[4897]: I1121 14:09:15.039016 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:15 crc kubenswrapper[4897]: I1121 14:09:15.039053 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:15 crc kubenswrapper[4897]: I1121 14:09:15.039066 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:15 crc kubenswrapper[4897]: I1121 14:09:15.039081 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:15 crc kubenswrapper[4897]: I1121 14:09:15.039092 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:15Z","lastTransitionTime":"2025-11-21T14:09:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:15 crc kubenswrapper[4897]: E1121 14:09:15.052614 4897 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8b8bdf9e-0fef-41ac-98ad-710dc2137183\\\",\\\"systemUUID\\\":\\\"8a210af6-a71f-4ea9-90a3-aac32fb868b7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:15Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:15 crc kubenswrapper[4897]: I1121 14:09:15.056576 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:15 crc kubenswrapper[4897]: I1121 14:09:15.056620 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:15 crc kubenswrapper[4897]: I1121 14:09:15.056631 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:15 crc kubenswrapper[4897]: I1121 14:09:15.056648 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:15 crc kubenswrapper[4897]: I1121 14:09:15.056665 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:15Z","lastTransitionTime":"2025-11-21T14:09:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:15 crc kubenswrapper[4897]: E1121 14:09:15.066837 4897 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8b8bdf9e-0fef-41ac-98ad-710dc2137183\\\",\\\"systemUUID\\\":\\\"8a210af6-a71f-4ea9-90a3-aac32fb868b7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:15Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:15 crc kubenswrapper[4897]: I1121 14:09:15.070292 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:15 crc kubenswrapper[4897]: I1121 14:09:15.070412 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:15 crc kubenswrapper[4897]: I1121 14:09:15.070424 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:15 crc kubenswrapper[4897]: I1121 14:09:15.070448 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:15 crc kubenswrapper[4897]: I1121 14:09:15.070459 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:15Z","lastTransitionTime":"2025-11-21T14:09:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:15 crc kubenswrapper[4897]: E1121 14:09:15.081761 4897 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8b8bdf9e-0fef-41ac-98ad-710dc2137183\\\",\\\"systemUUID\\\":\\\"8a210af6-a71f-4ea9-90a3-aac32fb868b7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:15Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:15 crc kubenswrapper[4897]: E1121 14:09:15.081879 4897 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 21 14:09:15 crc kubenswrapper[4897]: I1121 14:09:15.083353 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:15 crc kubenswrapper[4897]: I1121 14:09:15.083387 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:15 crc kubenswrapper[4897]: I1121 14:09:15.083399 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:15 crc kubenswrapper[4897]: I1121 14:09:15.083419 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:15 crc kubenswrapper[4897]: I1121 14:09:15.083432 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:15Z","lastTransitionTime":"2025-11-21T14:09:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:15 crc kubenswrapper[4897]: I1121 14:09:15.088678 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 21 14:09:15 crc kubenswrapper[4897]: I1121 14:09:15.088720 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 21 14:09:15 crc kubenswrapper[4897]: I1121 14:09:15.088751 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 21 14:09:15 crc kubenswrapper[4897]: E1121 14:09:15.088779 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 21 14:09:15 crc kubenswrapper[4897]: E1121 14:09:15.088888 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 21 14:09:15 crc kubenswrapper[4897]: E1121 14:09:15.089011 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 21 14:09:15 crc kubenswrapper[4897]: I1121 14:09:15.186904 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:15 crc kubenswrapper[4897]: I1121 14:09:15.186945 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:15 crc kubenswrapper[4897]: I1121 14:09:15.186955 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:15 crc kubenswrapper[4897]: I1121 14:09:15.186970 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:15 crc kubenswrapper[4897]: I1121 14:09:15.186981 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:15Z","lastTransitionTime":"2025-11-21T14:09:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:15 crc kubenswrapper[4897]: I1121 14:09:15.288954 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:15 crc kubenswrapper[4897]: I1121 14:09:15.289015 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:15 crc kubenswrapper[4897]: I1121 14:09:15.289031 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:15 crc kubenswrapper[4897]: I1121 14:09:15.289054 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:15 crc kubenswrapper[4897]: I1121 14:09:15.289073 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:15Z","lastTransitionTime":"2025-11-21T14:09:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:15 crc kubenswrapper[4897]: I1121 14:09:15.344051 4897 generic.go:334] "Generic (PLEG): container finished" podID="5a7d6101-d0d2-4828-b3ce-d01040c04640" containerID="89de56e32b796da11bf89dd598d04fbf2884745ddf0f5d87cccfcaf1c001d8d6" exitCode=0 Nov 21 14:09:15 crc kubenswrapper[4897]: I1121 14:09:15.344128 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-nl76g" event={"ID":"5a7d6101-d0d2-4828-b3ce-d01040c04640","Type":"ContainerDied","Data":"89de56e32b796da11bf89dd598d04fbf2884745ddf0f5d87cccfcaf1c001d8d6"} Nov 21 14:09:15 crc kubenswrapper[4897]: I1121 14:09:15.359905 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:15Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:15 crc kubenswrapper[4897]: I1121 14:09:15.383321 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c976d804a7ebeba2c879e920cbc389f96f7f5def4b4439e8c8980d5460ecdb65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:15Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:15 crc kubenswrapper[4897]: I1121 14:09:15.391641 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:15 crc kubenswrapper[4897]: I1121 14:09:15.391703 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:15 crc kubenswrapper[4897]: I1121 14:09:15.391721 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:15 crc kubenswrapper[4897]: I1121 14:09:15.391748 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:15 crc kubenswrapper[4897]: I1121 14:09:15.391765 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:15Z","lastTransitionTime":"2025-11-21T14:09:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:15 crc kubenswrapper[4897]: I1121 14:09:15.401133 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nl76g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a7d6101-d0d2-4828-b3ce-d01040c04640\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e56402f62821baf335f378084f4ed7c21b60ccbafe148419a0a9a403f106ad19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e56402f62821baf335f378084f4ed7c21b60ccbafe148419a0a9a403f106ad19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6022ff28e83dafcd846a7f03b090262aa3229a0da89f1984e4b6a054a04ac24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6022ff28e83dafcd846a7f03b090262aa3229a0da89f1984e4b6a054a04ac24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89de56e32b796da11bf89dd598d04fbf2884745ddf0f5d87cccfcaf1c001d8d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89de56e32b796da11bf89dd598d04fbf2884745ddf0f5d87cccfcaf1c001d8d6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nl76g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:15Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:15 crc kubenswrapper[4897]: I1121 14:09:15.422368 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48c6b1ea7497b33c00b956e400426c1d4f3c6c66e0225133434460800eef9314\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ea6068a6b6f8123b402de354e5374a0f02a7191a922d1718ab38f8898f47c52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:15Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:15 crc kubenswrapper[4897]: I1121 14:09:15.436665 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4qwqj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f46e2ba8-9ee0-406d-a668-2ba424618b69\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f43aa89d4d9defcb50d0dc6f90293b5f426b9cfc398b780c2e10f6a5c9f1e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbsgb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4qwqj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:15Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:15 crc kubenswrapper[4897]: I1121 14:09:15.461748 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"222f65f2-0b82-4760-bcf7-779244720f01\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7877db6f5c2bf0eb31218c64d67bca48f921ff7ac62b4e495f29024c086bf738\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7877db6f5c2bf0eb31218c64d67bca48f921ff7ac62b4e495f29024c086bf738\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tpfb6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:15Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:15 crc kubenswrapper[4897]: I1121 14:09:15.475382 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:15Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:15 crc kubenswrapper[4897]: I1121 14:09:15.487993 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7670227-d280-4847-b882-754429f56b0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://961961696b8750c248598f4cda032b18395e5c802275dcfe6d3e9f7ef3510f1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5854028821e6a360a259427b68e7bb4c2a7220e0c68f4cc91639b2ece0540a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-krv5b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:15Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:15 crc kubenswrapper[4897]: I1121 14:09:15.493720 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:15 crc kubenswrapper[4897]: I1121 14:09:15.493752 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:15 crc kubenswrapper[4897]: I1121 14:09:15.493762 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:15 crc kubenswrapper[4897]: I1121 14:09:15.493778 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:15 crc kubenswrapper[4897]: I1121 14:09:15.493790 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:15Z","lastTransitionTime":"2025-11-21T14:09:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:15 crc kubenswrapper[4897]: I1121 14:09:15.500775 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09877d9780caddd5f7f73e95fdf818a427ed645ed5997984113151f4b4759c13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:15Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:15 crc kubenswrapper[4897]: I1121 14:09:15.512621 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:15Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:15 crc kubenswrapper[4897]: I1121 14:09:15.524423 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6ljgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f3b70b5-9294-4f69-9528-500d28f34c89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://527cf28abdeef1b42d9c19fb88e3171dc05c4516c3a542a461b5b1daa85c6d25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78jhb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6ljgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:15Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:15 crc kubenswrapper[4897]: I1121 14:09:15.542219 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6cd330fd-e8a6-49d5-bfc9-5f6122ccd49e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1bd64b9b938bc80284bcee54141e90cad613d339c4f27ef2934b29bc2a318916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://936f59996d26e42099dcd055f87fff9e3ac4283b57e1956fe97463f2334152d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca74b28a069dc2e609b0779164c3aa470a15ed9848af0dd5fc3479aaf750157b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://270c53b9a342f26be077a01cd4e87e9ccd0eb5d2c84208ae38cbf720463245e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66c521c26d541bbddb064dd9710f1a7b41090d2f47343977922272dfc6dab7af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6eb2ac410b49098e5b3603de5e6bf5ed4484fab0b300021d0bed2e5ed8a42527\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6eb2ac410b49098e5b3603de5e6bf5ed4484fab0b300021d0bed2e5ed8a42527\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03385e0c78e4806f60c160227ef4905a11003ed3d70f0e5c0dd0a477979d1d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03385e0c78e4806f60c160227ef4905a11003ed3d70f0e5c0dd0a477979d1d4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://50aea2fa8b3b141b4330ddce2c8e0715230c6f06dab64b3abfd98a3067224bec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50aea2fa8b3b141b4330ddce2c8e0715230c6f06dab64b3abfd98a3067224bec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:15Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:15 crc kubenswrapper[4897]: I1121 14:09:15.554837 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a6c0a99-565b-41f9-a8b5-c8990a0c8eb3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://737553dd1d16e2019ea89a376e3a868f9792ed5239aa7d6cc40a05d3d4f286ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6682a8886d2b9bce069cf9261d4701ef6907f19574fd035bd89bd05626023e5c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de079956e3d8de0101f3f87dcc120ef23876c4972382639e419b3f34a38cdfca\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1907bfbad7eb34da642e30d6aa81820aab457f92347333a364ac12396b0e9dae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1907bfbad7eb34da642e30d6aa81820aab457f92347333a364ac12396b0e9dae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"le observer\\\\nW1121 14:09:07.994401 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1121 14:09:07.994601 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1121 14:09:07.995290 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1461162674/tls.crt::/tmp/serving-cert-1461162674/tls.key\\\\\\\"\\\\nI1121 14:09:08.162382 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1121 14:09:08.165301 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1121 14:09:08.165318 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1121 14:09:08.165343 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1121 14:09:08.165348 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1121 14:09:08.172015 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1121 14:09:08.172083 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1121 14:09:08.172089 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1121 14:09:08.172094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1121 14:09:08.172098 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1121 14:09:08.172102 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1121 14:09:08.172106 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1121 14:09:08.172023 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1121 14:09:08.174214 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d78c87c440c7d8e30075f6143111b3f504fb99686484b8f19a9c8d61a55eccc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://101fb592fef069af009c613a1e9c9825024ed205498ab72334211ba90d6f7397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://101fb592fef069af009c613a1e9c9825024ed205498ab72334211ba90d6f7397\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:15Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:15 crc kubenswrapper[4897]: I1121 14:09:15.568082 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b78c993a-9966-4e84-85a1-b4f73aaf17f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce3a2602338a1c1e9380ee55f78360f3d9cf6ccc62678886dc5ae574a72e3d9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c256823f2fb49ecc121f6ef564cd86196f9b4a473e9bdea670173596ef14a823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c65f71f8746044f5d41687c670a0308cd09dc41a6531f9b5720e852b3d240756\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1dbe88e7e6ad63899e107ebb884a16edb9a42b21ae7c86d7fbfaf89785b0d36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:15Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:15 crc kubenswrapper[4897]: I1121 14:09:15.596362 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:15 crc kubenswrapper[4897]: I1121 14:09:15.596399 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:15 crc kubenswrapper[4897]: I1121 14:09:15.596409 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:15 crc kubenswrapper[4897]: I1121 14:09:15.596427 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:15 crc kubenswrapper[4897]: I1121 14:09:15.596437 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:15Z","lastTransitionTime":"2025-11-21T14:09:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:15 crc kubenswrapper[4897]: I1121 14:09:15.698910 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:15 crc kubenswrapper[4897]: I1121 14:09:15.698954 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:15 crc kubenswrapper[4897]: I1121 14:09:15.698965 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:15 crc kubenswrapper[4897]: I1121 14:09:15.698982 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:15 crc kubenswrapper[4897]: I1121 14:09:15.698995 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:15Z","lastTransitionTime":"2025-11-21T14:09:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:15 crc kubenswrapper[4897]: I1121 14:09:15.801849 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:15 crc kubenswrapper[4897]: I1121 14:09:15.801903 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:15 crc kubenswrapper[4897]: I1121 14:09:15.801915 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:15 crc kubenswrapper[4897]: I1121 14:09:15.801935 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:15 crc kubenswrapper[4897]: I1121 14:09:15.802142 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:15Z","lastTransitionTime":"2025-11-21T14:09:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:15 crc kubenswrapper[4897]: I1121 14:09:15.905383 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:15 crc kubenswrapper[4897]: I1121 14:09:15.905433 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:15 crc kubenswrapper[4897]: I1121 14:09:15.905442 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:15 crc kubenswrapper[4897]: I1121 14:09:15.905458 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:15 crc kubenswrapper[4897]: I1121 14:09:15.905470 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:15Z","lastTransitionTime":"2025-11-21T14:09:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.008119 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.008160 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.008171 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.008188 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.008199 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:16Z","lastTransitionTime":"2025-11-21T14:09:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.110295 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.110327 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.110336 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.110347 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.110357 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:16Z","lastTransitionTime":"2025-11-21T14:09:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.212920 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.212952 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.212960 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.212974 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.212983 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:16Z","lastTransitionTime":"2025-11-21T14:09:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.316602 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.316936 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.316952 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.316974 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.316989 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:16Z","lastTransitionTime":"2025-11-21T14:09:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.352241 4897 generic.go:334] "Generic (PLEG): container finished" podID="5a7d6101-d0d2-4828-b3ce-d01040c04640" containerID="c2dff3cacaba7f9238f482e67b670ea775e0f96621434b99f4e76c2dfdde0321" exitCode=0 Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.352335 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-nl76g" event={"ID":"5a7d6101-d0d2-4828-b3ce-d01040c04640","Type":"ContainerDied","Data":"c2dff3cacaba7f9238f482e67b670ea775e0f96621434b99f4e76c2dfdde0321"} Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.365158 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48c6b1ea7497b33c00b956e400426c1d4f3c6c66e0225133434460800eef9314\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ea6068a6b6f8123b402de354e5374a0f02a7191a922d1718ab38f8898f47c52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:16Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.373912 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4qwqj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f46e2ba8-9ee0-406d-a668-2ba424618b69\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f43aa89d4d9defcb50d0dc6f90293b5f426b9cfc398b780c2e10f6a5c9f1e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbsgb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4qwqj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:16Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.388086 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:16Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.401978 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7670227-d280-4847-b882-754429f56b0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://961961696b8750c248598f4cda032b18395e5c802275dcfe6d3e9f7ef3510f1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5854028821e6a360a259427b68e7bb4c2a7220e0c68f4cc91639b2ece0540a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-krv5b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:16Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.419200 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.419254 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.419265 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.419285 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.419298 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:16Z","lastTransitionTime":"2025-11-21T14:09:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.420484 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"222f65f2-0b82-4760-bcf7-779244720f01\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7877db6f5c2bf0eb31218c64d67bca48f921ff7ac62b4e495f29024c086bf738\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7877db6f5c2bf0eb31218c64d67bca48f921ff7ac62b4e495f29024c086bf738\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tpfb6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:16Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.439996 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6cd330fd-e8a6-49d5-bfc9-5f6122ccd49e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1bd64b9b938bc80284bcee54141e90cad613d339c4f27ef2934b29bc2a318916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://936f59996d26e42099dcd055f87fff9e3ac4283b57e1956fe97463f2334152d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca74b28a069dc2e609b0779164c3aa470a15ed9848af0dd5fc3479aaf750157b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://270c53b9a342f26be077a01cd4e87e9ccd0eb5d2c84208ae38cbf720463245e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66c521c26d541bbddb064dd9710f1a7b41090d2f47343977922272dfc6dab7af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6eb2ac410b49098e5b3603de5e6bf5ed4484fab0b300021d0bed2e5ed8a42527\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6eb2ac410b49098e5b3603de5e6bf5ed4484fab0b300021d0bed2e5ed8a42527\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03385e0c78e4806f60c160227ef4905a11003ed3d70f0e5c0dd0a477979d1d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03385e0c78e4806f60c160227ef4905a11003ed3d70f0e5c0dd0a477979d1d4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://50aea2fa8b3b141b4330ddce2c8e0715230c6f06dab64b3abfd98a3067224bec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50aea2fa8b3b141b4330ddce2c8e0715230c6f06dab64b3abfd98a3067224bec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:16Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.452343 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a6c0a99-565b-41f9-a8b5-c8990a0c8eb3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://737553dd1d16e2019ea89a376e3a868f9792ed5239aa7d6cc40a05d3d4f286ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6682a8886d2b9bce069cf9261d4701ef6907f19574fd035bd89bd05626023e5c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de079956e3d8de0101f3f87dcc120ef23876c4972382639e419b3f34a38cdfca\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1907bfbad7eb34da642e30d6aa81820aab457f92347333a364ac12396b0e9dae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1907bfbad7eb34da642e30d6aa81820aab457f92347333a364ac12396b0e9dae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"le observer\\\\nW1121 14:09:07.994401 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1121 14:09:07.994601 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1121 14:09:07.995290 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1461162674/tls.crt::/tmp/serving-cert-1461162674/tls.key\\\\\\\"\\\\nI1121 14:09:08.162382 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1121 14:09:08.165301 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1121 14:09:08.165318 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1121 14:09:08.165343 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1121 14:09:08.165348 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1121 14:09:08.172015 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1121 14:09:08.172083 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1121 14:09:08.172089 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1121 14:09:08.172094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1121 14:09:08.172098 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1121 14:09:08.172102 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1121 14:09:08.172106 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1121 14:09:08.172023 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1121 14:09:08.174214 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d78c87c440c7d8e30075f6143111b3f504fb99686484b8f19a9c8d61a55eccc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://101fb592fef069af009c613a1e9c9825024ed205498ab72334211ba90d6f7397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://101fb592fef069af009c613a1e9c9825024ed205498ab72334211ba90d6f7397\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:16Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.463460 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b78c993a-9966-4e84-85a1-b4f73aaf17f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce3a2602338a1c1e9380ee55f78360f3d9cf6ccc62678886dc5ae574a72e3d9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c256823f2fb49ecc121f6ef564cd86196f9b4a473e9bdea670173596ef14a823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c65f71f8746044f5d41687c670a0308cd09dc41a6531f9b5720e852b3d240756\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1dbe88e7e6ad63899e107ebb884a16edb9a42b21ae7c86d7fbfaf89785b0d36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:16Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.475835 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09877d9780caddd5f7f73e95fdf818a427ed645ed5997984113151f4b4759c13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:16Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.503949 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:16Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.522314 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.522361 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.522371 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.522390 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.522403 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:16Z","lastTransitionTime":"2025-11-21T14:09:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.535708 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6ljgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f3b70b5-9294-4f69-9528-500d28f34c89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://527cf28abdeef1b42d9c19fb88e3171dc05c4516c3a542a461b5b1daa85c6d25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78jhb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6ljgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:16Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.554262 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:16Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.570044 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c976d804a7ebeba2c879e920cbc389f96f7f5def4b4439e8c8980d5460ecdb65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:16Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.584077 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nl76g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a7d6101-d0d2-4828-b3ce-d01040c04640\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e56402f62821baf335f378084f4ed7c21b60ccbafe148419a0a9a403f106ad19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e56402f62821baf335f378084f4ed7c21b60ccbafe148419a0a9a403f106ad19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6022ff28e83dafcd846a7f03b090262aa3229a0da89f1984e4b6a054a04ac24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6022ff28e83dafcd846a7f03b090262aa3229a0da89f1984e4b6a054a04ac24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89de56e32b796da11bf89dd598d04fbf2884745ddf0f5d87cccfcaf1c001d8d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89de56e32b796da11bf89dd598d04fbf2884745ddf0f5d87cccfcaf1c001d8d6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2dff3cacaba7f9238f482e67b670ea775e0f96621434b99f4e76c2dfdde0321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2dff3cacaba7f9238f482e67b670ea775e0f96621434b99f4e76c2dfdde0321\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nl76g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:16Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.625735 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.625788 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.625803 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.625821 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.625831 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:16Z","lastTransitionTime":"2025-11-21T14:09:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.711450 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-7qfmq"] Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.711908 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-7qfmq" Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.713798 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.713881 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.714175 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.714692 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.726997 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6ljgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f3b70b5-9294-4f69-9528-500d28f34c89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://527cf28abdeef1b42d9c19fb88e3171dc05c4516c3a542a461b5b1daa85c6d25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78jhb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6ljgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:16Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.728199 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.728243 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.728259 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.728280 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.728295 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:16Z","lastTransitionTime":"2025-11-21T14:09:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.739107 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 21 14:09:16 crc kubenswrapper[4897]: E1121 14:09:16.739308 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-21 14:09:24.739279132 +0000 UTC m=+42.023872617 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.739411 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.739455 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 21 14:09:16 crc kubenswrapper[4897]: E1121 14:09:16.739580 4897 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 21 14:09:16 crc kubenswrapper[4897]: E1121 14:09:16.739660 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-21 14:09:24.739636521 +0000 UTC m=+42.024229996 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 21 14:09:16 crc kubenswrapper[4897]: E1121 14:09:16.740039 4897 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 21 14:09:16 crc kubenswrapper[4897]: E1121 14:09:16.740075 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-21 14:09:24.740065283 +0000 UTC m=+42.024658818 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.745788 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6cd330fd-e8a6-49d5-bfc9-5f6122ccd49e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1bd64b9b938bc80284bcee54141e90cad613d339c4f27ef2934b29bc2a318916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://936f59996d26e42099dcd055f87fff9e3ac4283b57e1956fe97463f2334152d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca74b28a069dc2e609b0779164c3aa470a15ed9848af0dd5fc3479aaf750157b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://270c53b9a342f26be077a01cd4e87e9ccd0eb5d2c84208ae38cbf720463245e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66c521c26d541bbddb064dd9710f1a7b41090d2f47343977922272dfc6dab7af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6eb2ac410b49098e5b3603de5e6bf5ed4484fab0b300021d0bed2e5ed8a42527\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6eb2ac410b49098e5b3603de5e6bf5ed4484fab0b300021d0bed2e5ed8a42527\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03385e0c78e4806f60c160227ef4905a11003ed3d70f0e5c0dd0a477979d1d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03385e0c78e4806f60c160227ef4905a11003ed3d70f0e5c0dd0a477979d1d4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://50aea2fa8b3b141b4330ddce2c8e0715230c6f06dab64b3abfd98a3067224bec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50aea2fa8b3b141b4330ddce2c8e0715230c6f06dab64b3abfd98a3067224bec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:16Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.759426 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a6c0a99-565b-41f9-a8b5-c8990a0c8eb3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://737553dd1d16e2019ea89a376e3a868f9792ed5239aa7d6cc40a05d3d4f286ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6682a8886d2b9bce069cf9261d4701ef6907f19574fd035bd89bd05626023e5c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de079956e3d8de0101f3f87dcc120ef23876c4972382639e419b3f34a38cdfca\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1907bfbad7eb34da642e30d6aa81820aab457f92347333a364ac12396b0e9dae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1907bfbad7eb34da642e30d6aa81820aab457f92347333a364ac12396b0e9dae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"le observer\\\\nW1121 14:09:07.994401 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1121 14:09:07.994601 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1121 14:09:07.995290 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1461162674/tls.crt::/tmp/serving-cert-1461162674/tls.key\\\\\\\"\\\\nI1121 14:09:08.162382 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1121 14:09:08.165301 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1121 14:09:08.165318 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1121 14:09:08.165343 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1121 14:09:08.165348 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1121 14:09:08.172015 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1121 14:09:08.172083 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1121 14:09:08.172089 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1121 14:09:08.172094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1121 14:09:08.172098 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1121 14:09:08.172102 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1121 14:09:08.172106 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1121 14:09:08.172023 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1121 14:09:08.174214 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d78c87c440c7d8e30075f6143111b3f504fb99686484b8f19a9c8d61a55eccc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://101fb592fef069af009c613a1e9c9825024ed205498ab72334211ba90d6f7397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://101fb592fef069af009c613a1e9c9825024ed205498ab72334211ba90d6f7397\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:16Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.774293 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b78c993a-9966-4e84-85a1-b4f73aaf17f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce3a2602338a1c1e9380ee55f78360f3d9cf6ccc62678886dc5ae574a72e3d9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c256823f2fb49ecc121f6ef564cd86196f9b4a473e9bdea670173596ef14a823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c65f71f8746044f5d41687c670a0308cd09dc41a6531f9b5720e852b3d240756\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1dbe88e7e6ad63899e107ebb884a16edb9a42b21ae7c86d7fbfaf89785b0d36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:16Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.786495 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09877d9780caddd5f7f73e95fdf818a427ed645ed5997984113151f4b4759c13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:16Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.798923 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:16Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.815499 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:16Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.829137 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c976d804a7ebeba2c879e920cbc389f96f7f5def4b4439e8c8980d5460ecdb65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:16Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.830570 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.830606 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.830617 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.830635 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.830646 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:16Z","lastTransitionTime":"2025-11-21T14:09:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.839951 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/223b7002-3e67-4848-b1fb-db2921deb64e-serviceca\") pod \"node-ca-7qfmq\" (UID: \"223b7002-3e67-4848-b1fb-db2921deb64e\") " pod="openshift-image-registry/node-ca-7qfmq" Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.840005 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f2m27\" (UniqueName: \"kubernetes.io/projected/223b7002-3e67-4848-b1fb-db2921deb64e-kube-api-access-f2m27\") pod \"node-ca-7qfmq\" (UID: \"223b7002-3e67-4848-b1fb-db2921deb64e\") " pod="openshift-image-registry/node-ca-7qfmq" Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.840033 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/223b7002-3e67-4848-b1fb-db2921deb64e-host\") pod \"node-ca-7qfmq\" (UID: \"223b7002-3e67-4848-b1fb-db2921deb64e\") " pod="openshift-image-registry/node-ca-7qfmq" Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.840060 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.840092 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 21 14:09:16 crc kubenswrapper[4897]: E1121 14:09:16.840255 4897 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 21 14:09:16 crc kubenswrapper[4897]: E1121 14:09:16.840311 4897 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 21 14:09:16 crc kubenswrapper[4897]: E1121 14:09:16.840333 4897 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 21 14:09:16 crc kubenswrapper[4897]: E1121 14:09:16.840407 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-21 14:09:24.840378596 +0000 UTC m=+42.124972161 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 21 14:09:16 crc kubenswrapper[4897]: E1121 14:09:16.840258 4897 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 21 14:09:16 crc kubenswrapper[4897]: E1121 14:09:16.840445 4897 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 21 14:09:16 crc kubenswrapper[4897]: E1121 14:09:16.840460 4897 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 21 14:09:16 crc kubenswrapper[4897]: E1121 14:09:16.840508 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-21 14:09:24.840494028 +0000 UTC m=+42.125087523 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.844015 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nl76g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a7d6101-d0d2-4828-b3ce-d01040c04640\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e56402f62821baf335f378084f4ed7c21b60ccbafe148419a0a9a403f106ad19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e56402f62821baf335f378084f4ed7c21b60ccbafe148419a0a9a403f106ad19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6022ff28e83dafcd846a7f03b090262aa3229a0da89f1984e4b6a054a04ac24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6022ff28e83dafcd846a7f03b090262aa3229a0da89f1984e4b6a054a04ac24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89de56e32b796da11bf89dd598d04fbf2884745ddf0f5d87cccfcaf1c001d8d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89de56e32b796da11bf89dd598d04fbf2884745ddf0f5d87cccfcaf1c001d8d6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2dff3cacaba7f9238f482e67b670ea775e0f96621434b99f4e76c2dfdde0321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2dff3cacaba7f9238f482e67b670ea775e0f96621434b99f4e76c2dfdde0321\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nl76g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:16Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.858432 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48c6b1ea7497b33c00b956e400426c1d4f3c6c66e0225133434460800eef9314\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ea6068a6b6f8123b402de354e5374a0f02a7191a922d1718ab38f8898f47c52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:16Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.869152 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4qwqj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f46e2ba8-9ee0-406d-a668-2ba424618b69\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f43aa89d4d9defcb50d0dc6f90293b5f426b9cfc398b780c2e10f6a5c9f1e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbsgb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4qwqj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:16Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.882015 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:16Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.894851 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7670227-d280-4847-b882-754429f56b0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://961961696b8750c248598f4cda032b18395e5c802275dcfe6d3e9f7ef3510f1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5854028821e6a360a259427b68e7bb4c2a7220e0c68f4cc91639b2ece0540a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-krv5b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:16Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.914145 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"222f65f2-0b82-4760-bcf7-779244720f01\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7877db6f5c2bf0eb31218c64d67bca48f921ff7ac62b4e495f29024c086bf738\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7877db6f5c2bf0eb31218c64d67bca48f921ff7ac62b4e495f29024c086bf738\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tpfb6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:16Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.925866 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7qfmq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"223b7002-3e67-4848-b1fb-db2921deb64e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:16Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:16Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2m27\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7qfmq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:16Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.933405 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.933468 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.933481 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.933541 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.933556 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:16Z","lastTransitionTime":"2025-11-21T14:09:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.940899 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/223b7002-3e67-4848-b1fb-db2921deb64e-serviceca\") pod \"node-ca-7qfmq\" (UID: \"223b7002-3e67-4848-b1fb-db2921deb64e\") " pod="openshift-image-registry/node-ca-7qfmq" Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.940955 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f2m27\" (UniqueName: \"kubernetes.io/projected/223b7002-3e67-4848-b1fb-db2921deb64e-kube-api-access-f2m27\") pod \"node-ca-7qfmq\" (UID: \"223b7002-3e67-4848-b1fb-db2921deb64e\") " pod="openshift-image-registry/node-ca-7qfmq" Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.940985 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/223b7002-3e67-4848-b1fb-db2921deb64e-host\") pod \"node-ca-7qfmq\" (UID: \"223b7002-3e67-4848-b1fb-db2921deb64e\") " pod="openshift-image-registry/node-ca-7qfmq" Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.941038 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/223b7002-3e67-4848-b1fb-db2921deb64e-host\") pod \"node-ca-7qfmq\" (UID: \"223b7002-3e67-4848-b1fb-db2921deb64e\") " pod="openshift-image-registry/node-ca-7qfmq" Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.942575 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/223b7002-3e67-4848-b1fb-db2921deb64e-serviceca\") pod \"node-ca-7qfmq\" (UID: \"223b7002-3e67-4848-b1fb-db2921deb64e\") " pod="openshift-image-registry/node-ca-7qfmq" Nov 21 14:09:16 crc kubenswrapper[4897]: I1121 14:09:16.959682 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f2m27\" (UniqueName: \"kubernetes.io/projected/223b7002-3e67-4848-b1fb-db2921deb64e-kube-api-access-f2m27\") pod \"node-ca-7qfmq\" (UID: \"223b7002-3e67-4848-b1fb-db2921deb64e\") " pod="openshift-image-registry/node-ca-7qfmq" Nov 21 14:09:17 crc kubenswrapper[4897]: I1121 14:09:17.025094 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-7qfmq" Nov 21 14:09:17 crc kubenswrapper[4897]: I1121 14:09:17.035984 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:17 crc kubenswrapper[4897]: I1121 14:09:17.036030 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:17 crc kubenswrapper[4897]: I1121 14:09:17.036042 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:17 crc kubenswrapper[4897]: I1121 14:09:17.036060 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:17 crc kubenswrapper[4897]: I1121 14:09:17.036072 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:17Z","lastTransitionTime":"2025-11-21T14:09:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:17 crc kubenswrapper[4897]: W1121 14:09:17.040677 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod223b7002_3e67_4848_b1fb_db2921deb64e.slice/crio-50ca5ef6f6eb9472f22be9864ad997238c40f85124e7c368fd4ac7a9670151e6 WatchSource:0}: Error finding container 50ca5ef6f6eb9472f22be9864ad997238c40f85124e7c368fd4ac7a9670151e6: Status 404 returned error can't find the container with id 50ca5ef6f6eb9472f22be9864ad997238c40f85124e7c368fd4ac7a9670151e6 Nov 21 14:09:17 crc kubenswrapper[4897]: I1121 14:09:17.088940 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 21 14:09:17 crc kubenswrapper[4897]: E1121 14:09:17.089357 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 21 14:09:17 crc kubenswrapper[4897]: I1121 14:09:17.089438 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 21 14:09:17 crc kubenswrapper[4897]: I1121 14:09:17.089467 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 21 14:09:17 crc kubenswrapper[4897]: E1121 14:09:17.089611 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 21 14:09:17 crc kubenswrapper[4897]: E1121 14:09:17.089813 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 21 14:09:17 crc kubenswrapper[4897]: I1121 14:09:17.139100 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:17 crc kubenswrapper[4897]: I1121 14:09:17.139476 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:17 crc kubenswrapper[4897]: I1121 14:09:17.139486 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:17 crc kubenswrapper[4897]: I1121 14:09:17.139561 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:17 crc kubenswrapper[4897]: I1121 14:09:17.139578 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:17Z","lastTransitionTime":"2025-11-21T14:09:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:17 crc kubenswrapper[4897]: I1121 14:09:17.242034 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:17 crc kubenswrapper[4897]: I1121 14:09:17.242079 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:17 crc kubenswrapper[4897]: I1121 14:09:17.242099 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:17 crc kubenswrapper[4897]: I1121 14:09:17.242118 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:17 crc kubenswrapper[4897]: I1121 14:09:17.242129 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:17Z","lastTransitionTime":"2025-11-21T14:09:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:17 crc kubenswrapper[4897]: I1121 14:09:17.344840 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:17 crc kubenswrapper[4897]: I1121 14:09:17.344888 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:17 crc kubenswrapper[4897]: I1121 14:09:17.344898 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:17 crc kubenswrapper[4897]: I1121 14:09:17.344913 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:17 crc kubenswrapper[4897]: I1121 14:09:17.344924 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:17Z","lastTransitionTime":"2025-11-21T14:09:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:17 crc kubenswrapper[4897]: I1121 14:09:17.359431 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-nl76g" event={"ID":"5a7d6101-d0d2-4828-b3ce-d01040c04640","Type":"ContainerStarted","Data":"a153389a4137ee580e5118f3248400460f9d014c992a956f508815755f583afb"} Nov 21 14:09:17 crc kubenswrapper[4897]: I1121 14:09:17.361941 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-7qfmq" event={"ID":"223b7002-3e67-4848-b1fb-db2921deb64e","Type":"ContainerStarted","Data":"381a46c6290da2b36e572f46c16f1a50aa3e25359bcfbc02c7a5a18e1484860e"} Nov 21 14:09:17 crc kubenswrapper[4897]: I1121 14:09:17.361998 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-7qfmq" event={"ID":"223b7002-3e67-4848-b1fb-db2921deb64e","Type":"ContainerStarted","Data":"50ca5ef6f6eb9472f22be9864ad997238c40f85124e7c368fd4ac7a9670151e6"} Nov 21 14:09:17 crc kubenswrapper[4897]: I1121 14:09:17.366715 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" event={"ID":"222f65f2-0b82-4760-bcf7-779244720f01","Type":"ContainerStarted","Data":"089c9d4e6d0fb1a4e9a65a34a033e15c12508d2dbe3b9086b464914a1dac41e5"} Nov 21 14:09:17 crc kubenswrapper[4897]: I1121 14:09:17.367094 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" Nov 21 14:09:17 crc kubenswrapper[4897]: I1121 14:09:17.367140 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" Nov 21 14:09:17 crc kubenswrapper[4897]: I1121 14:09:17.367305 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" Nov 21 14:09:17 crc kubenswrapper[4897]: I1121 14:09:17.376965 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48c6b1ea7497b33c00b956e400426c1d4f3c6c66e0225133434460800eef9314\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ea6068a6b6f8123b402de354e5374a0f02a7191a922d1718ab38f8898f47c52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:17Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:17 crc kubenswrapper[4897]: I1121 14:09:17.390867 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4qwqj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f46e2ba8-9ee0-406d-a668-2ba424618b69\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f43aa89d4d9defcb50d0dc6f90293b5f426b9cfc398b780c2e10f6a5c9f1e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbsgb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4qwqj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:17Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:17 crc kubenswrapper[4897]: I1121 14:09:17.391607 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" Nov 21 14:09:17 crc kubenswrapper[4897]: I1121 14:09:17.392234 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" Nov 21 14:09:17 crc kubenswrapper[4897]: I1121 14:09:17.403727 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:17Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:17 crc kubenswrapper[4897]: I1121 14:09:17.414842 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7670227-d280-4847-b882-754429f56b0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://961961696b8750c248598f4cda032b18395e5c802275dcfe6d3e9f7ef3510f1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5854028821e6a360a259427b68e7bb4c2a7220e0c68f4cc91639b2ece0540a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-krv5b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:17Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:17 crc kubenswrapper[4897]: I1121 14:09:17.434991 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"222f65f2-0b82-4760-bcf7-779244720f01\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7877db6f5c2bf0eb31218c64d67bca48f921ff7ac62b4e495f29024c086bf738\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7877db6f5c2bf0eb31218c64d67bca48f921ff7ac62b4e495f29024c086bf738\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tpfb6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:17Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:17 crc kubenswrapper[4897]: I1121 14:09:17.446377 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7qfmq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"223b7002-3e67-4848-b1fb-db2921deb64e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:16Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:16Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2m27\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7qfmq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:17Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:17 crc kubenswrapper[4897]: I1121 14:09:17.447318 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:17 crc kubenswrapper[4897]: I1121 14:09:17.447356 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:17 crc kubenswrapper[4897]: I1121 14:09:17.447370 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:17 crc kubenswrapper[4897]: I1121 14:09:17.447389 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:17 crc kubenswrapper[4897]: I1121 14:09:17.447400 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:17Z","lastTransitionTime":"2025-11-21T14:09:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:17 crc kubenswrapper[4897]: I1121 14:09:17.459580 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6ljgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f3b70b5-9294-4f69-9528-500d28f34c89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://527cf28abdeef1b42d9c19fb88e3171dc05c4516c3a542a461b5b1daa85c6d25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78jhb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6ljgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:17Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:17 crc kubenswrapper[4897]: I1121 14:09:17.479829 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6cd330fd-e8a6-49d5-bfc9-5f6122ccd49e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1bd64b9b938bc80284bcee54141e90cad613d339c4f27ef2934b29bc2a318916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://936f59996d26e42099dcd055f87fff9e3ac4283b57e1956fe97463f2334152d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca74b28a069dc2e609b0779164c3aa470a15ed9848af0dd5fc3479aaf750157b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://270c53b9a342f26be077a01cd4e87e9ccd0eb5d2c84208ae38cbf720463245e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66c521c26d541bbddb064dd9710f1a7b41090d2f47343977922272dfc6dab7af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6eb2ac410b49098e5b3603de5e6bf5ed4484fab0b300021d0bed2e5ed8a42527\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6eb2ac410b49098e5b3603de5e6bf5ed4484fab0b300021d0bed2e5ed8a42527\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03385e0c78e4806f60c160227ef4905a11003ed3d70f0e5c0dd0a477979d1d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03385e0c78e4806f60c160227ef4905a11003ed3d70f0e5c0dd0a477979d1d4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://50aea2fa8b3b141b4330ddce2c8e0715230c6f06dab64b3abfd98a3067224bec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50aea2fa8b3b141b4330ddce2c8e0715230c6f06dab64b3abfd98a3067224bec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:17Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:17 crc kubenswrapper[4897]: I1121 14:09:17.493217 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a6c0a99-565b-41f9-a8b5-c8990a0c8eb3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://737553dd1d16e2019ea89a376e3a868f9792ed5239aa7d6cc40a05d3d4f286ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6682a8886d2b9bce069cf9261d4701ef6907f19574fd035bd89bd05626023e5c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de079956e3d8de0101f3f87dcc120ef23876c4972382639e419b3f34a38cdfca\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1907bfbad7eb34da642e30d6aa81820aab457f92347333a364ac12396b0e9dae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1907bfbad7eb34da642e30d6aa81820aab457f92347333a364ac12396b0e9dae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"le observer\\\\nW1121 14:09:07.994401 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1121 14:09:07.994601 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1121 14:09:07.995290 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1461162674/tls.crt::/tmp/serving-cert-1461162674/tls.key\\\\\\\"\\\\nI1121 14:09:08.162382 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1121 14:09:08.165301 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1121 14:09:08.165318 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1121 14:09:08.165343 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1121 14:09:08.165348 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1121 14:09:08.172015 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1121 14:09:08.172083 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1121 14:09:08.172089 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1121 14:09:08.172094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1121 14:09:08.172098 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1121 14:09:08.172102 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1121 14:09:08.172106 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1121 14:09:08.172023 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1121 14:09:08.174214 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d78c87c440c7d8e30075f6143111b3f504fb99686484b8f19a9c8d61a55eccc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://101fb592fef069af009c613a1e9c9825024ed205498ab72334211ba90d6f7397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://101fb592fef069af009c613a1e9c9825024ed205498ab72334211ba90d6f7397\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:17Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:17 crc kubenswrapper[4897]: I1121 14:09:17.507664 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b78c993a-9966-4e84-85a1-b4f73aaf17f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce3a2602338a1c1e9380ee55f78360f3d9cf6ccc62678886dc5ae574a72e3d9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c256823f2fb49ecc121f6ef564cd86196f9b4a473e9bdea670173596ef14a823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c65f71f8746044f5d41687c670a0308cd09dc41a6531f9b5720e852b3d240756\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1dbe88e7e6ad63899e107ebb884a16edb9a42b21ae7c86d7fbfaf89785b0d36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:17Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:17 crc kubenswrapper[4897]: I1121 14:09:17.520040 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09877d9780caddd5f7f73e95fdf818a427ed645ed5997984113151f4b4759c13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:17Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:17 crc kubenswrapper[4897]: I1121 14:09:17.531720 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:17Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:17 crc kubenswrapper[4897]: I1121 14:09:17.543775 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:17Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:17 crc kubenswrapper[4897]: I1121 14:09:17.549287 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:17 crc kubenswrapper[4897]: I1121 14:09:17.549330 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:17 crc kubenswrapper[4897]: I1121 14:09:17.549342 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:17 crc kubenswrapper[4897]: I1121 14:09:17.549358 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:17 crc kubenswrapper[4897]: I1121 14:09:17.549369 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:17Z","lastTransitionTime":"2025-11-21T14:09:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:17 crc kubenswrapper[4897]: I1121 14:09:17.559585 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c976d804a7ebeba2c879e920cbc389f96f7f5def4b4439e8c8980d5460ecdb65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:17Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:17 crc kubenswrapper[4897]: I1121 14:09:17.575502 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nl76g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a7d6101-d0d2-4828-b3ce-d01040c04640\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e56402f62821baf335f378084f4ed7c21b60ccbafe148419a0a9a403f106ad19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e56402f62821baf335f378084f4ed7c21b60ccbafe148419a0a9a403f106ad19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6022ff28e83dafcd846a7f03b090262aa3229a0da89f1984e4b6a054a04ac24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6022ff28e83dafcd846a7f03b090262aa3229a0da89f1984e4b6a054a04ac24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89de56e32b796da11bf89dd598d04fbf2884745ddf0f5d87cccfcaf1c001d8d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89de56e32b796da11bf89dd598d04fbf2884745ddf0f5d87cccfcaf1c001d8d6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2dff3cacaba7f9238f482e67b670ea775e0f96621434b99f4e76c2dfdde0321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2dff3cacaba7f9238f482e67b670ea775e0f96621434b99f4e76c2dfdde0321\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a153389a4137ee580e5118f3248400460f9d014c992a956f508815755f583afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nl76g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:17Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:17 crc kubenswrapper[4897]: I1121 14:09:17.589407 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c976d804a7ebeba2c879e920cbc389f96f7f5def4b4439e8c8980d5460ecdb65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:17Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:17 crc kubenswrapper[4897]: I1121 14:09:17.604576 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nl76g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a7d6101-d0d2-4828-b3ce-d01040c04640\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e56402f62821baf335f378084f4ed7c21b60ccbafe148419a0a9a403f106ad19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e56402f62821baf335f378084f4ed7c21b60ccbafe148419a0a9a403f106ad19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6022ff28e83dafcd846a7f03b090262aa3229a0da89f1984e4b6a054a04ac24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6022ff28e83dafcd846a7f03b090262aa3229a0da89f1984e4b6a054a04ac24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89de56e32b796da11bf89dd598d04fbf2884745ddf0f5d87cccfcaf1c001d8d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89de56e32b796da11bf89dd598d04fbf2884745ddf0f5d87cccfcaf1c001d8d6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2dff3cacaba7f9238f482e67b670ea775e0f96621434b99f4e76c2dfdde0321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2dff3cacaba7f9238f482e67b670ea775e0f96621434b99f4e76c2dfdde0321\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a153389a4137ee580e5118f3248400460f9d014c992a956f508815755f583afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nl76g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:17Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:17 crc kubenswrapper[4897]: I1121 14:09:17.618708 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:17Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:17 crc kubenswrapper[4897]: I1121 14:09:17.631679 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48c6b1ea7497b33c00b956e400426c1d4f3c6c66e0225133434460800eef9314\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ea6068a6b6f8123b402de354e5374a0f02a7191a922d1718ab38f8898f47c52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:17Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:17 crc kubenswrapper[4897]: I1121 14:09:17.642303 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4qwqj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f46e2ba8-9ee0-406d-a668-2ba424618b69\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f43aa89d4d9defcb50d0dc6f90293b5f426b9cfc398b780c2e10f6a5c9f1e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbsgb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4qwqj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:17Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:17 crc kubenswrapper[4897]: I1121 14:09:17.651910 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:17 crc kubenswrapper[4897]: I1121 14:09:17.651957 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:17 crc kubenswrapper[4897]: I1121 14:09:17.651970 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:17 crc kubenswrapper[4897]: I1121 14:09:17.651989 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:17 crc kubenswrapper[4897]: I1121 14:09:17.652002 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:17Z","lastTransitionTime":"2025-11-21T14:09:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:17 crc kubenswrapper[4897]: I1121 14:09:17.654674 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:17Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:17 crc kubenswrapper[4897]: I1121 14:09:17.668001 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7670227-d280-4847-b882-754429f56b0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://961961696b8750c248598f4cda032b18395e5c802275dcfe6d3e9f7ef3510f1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5854028821e6a360a259427b68e7bb4c2a7220e0c68f4cc91639b2ece0540a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-krv5b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:17Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:17 crc kubenswrapper[4897]: I1121 14:09:17.691233 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"222f65f2-0b82-4760-bcf7-779244720f01\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92f1d932a83d65ba6a9a906cc489ff51e76547d07d42656630c59421f833c4db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccd51aa0faf3b1327f207b355cddf86a7353c0051f7e5ce0396f3c4bc67e3b16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8971c38daad45bda3d61545f52120dad40992dbf2e1d6e1e776500ed58d73f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfee19f82e35869a641318a0c45bc7a2ae71d80039f5790e60e277be15c9d3ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec279f2f558a8ac653006321b18cbaf13185876fe303c5cc1c65988aef09edbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de095f3c760b13d5197f624eac3806a1f4186577f2b95eb522532d453eb266e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://089c9d4e6d0fb1a4e9a65a34a033e15c12508d2dbe3b9086b464914a1dac41e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b486a86e0e97ab409e590f30a1bf5cb9f584a7c67f96ff1236e42ca4f8675b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7877db6f5c2bf0eb31218c64d67bca48f921ff7ac62b4e495f29024c086bf738\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7877db6f5c2bf0eb31218c64d67bca48f921ff7ac62b4e495f29024c086bf738\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tpfb6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:17Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:17 crc kubenswrapper[4897]: I1121 14:09:17.701221 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7qfmq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"223b7002-3e67-4848-b1fb-db2921deb64e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381a46c6290da2b36e572f46c16f1a50aa3e25359bcfbc02c7a5a18e1484860e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2m27\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7qfmq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:17Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:17 crc kubenswrapper[4897]: I1121 14:09:17.713280 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a6c0a99-565b-41f9-a8b5-c8990a0c8eb3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://737553dd1d16e2019ea89a376e3a868f9792ed5239aa7d6cc40a05d3d4f286ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6682a8886d2b9bce069cf9261d4701ef6907f19574fd035bd89bd05626023e5c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de079956e3d8de0101f3f87dcc120ef23876c4972382639e419b3f34a38cdfca\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1907bfbad7eb34da642e30d6aa81820aab457f92347333a364ac12396b0e9dae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1907bfbad7eb34da642e30d6aa81820aab457f92347333a364ac12396b0e9dae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"le observer\\\\nW1121 14:09:07.994401 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1121 14:09:07.994601 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1121 14:09:07.995290 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1461162674/tls.crt::/tmp/serving-cert-1461162674/tls.key\\\\\\\"\\\\nI1121 14:09:08.162382 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1121 14:09:08.165301 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1121 14:09:08.165318 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1121 14:09:08.165343 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1121 14:09:08.165348 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1121 14:09:08.172015 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1121 14:09:08.172083 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1121 14:09:08.172089 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1121 14:09:08.172094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1121 14:09:08.172098 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1121 14:09:08.172102 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1121 14:09:08.172106 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1121 14:09:08.172023 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1121 14:09:08.174214 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d78c87c440c7d8e30075f6143111b3f504fb99686484b8f19a9c8d61a55eccc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://101fb592fef069af009c613a1e9c9825024ed205498ab72334211ba90d6f7397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://101fb592fef069af009c613a1e9c9825024ed205498ab72334211ba90d6f7397\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:17Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:17 crc kubenswrapper[4897]: I1121 14:09:17.726038 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b78c993a-9966-4e84-85a1-b4f73aaf17f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce3a2602338a1c1e9380ee55f78360f3d9cf6ccc62678886dc5ae574a72e3d9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c256823f2fb49ecc121f6ef564cd86196f9b4a473e9bdea670173596ef14a823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c65f71f8746044f5d41687c670a0308cd09dc41a6531f9b5720e852b3d240756\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1dbe88e7e6ad63899e107ebb884a16edb9a42b21ae7c86d7fbfaf89785b0d36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:17Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:17 crc kubenswrapper[4897]: I1121 14:09:17.739189 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09877d9780caddd5f7f73e95fdf818a427ed645ed5997984113151f4b4759c13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:17Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:17 crc kubenswrapper[4897]: I1121 14:09:17.754309 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:17 crc kubenswrapper[4897]: I1121 14:09:17.754348 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:17 crc kubenswrapper[4897]: I1121 14:09:17.754359 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:17 crc kubenswrapper[4897]: I1121 14:09:17.754375 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:17 crc kubenswrapper[4897]: I1121 14:09:17.754387 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:17Z","lastTransitionTime":"2025-11-21T14:09:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:17 crc kubenswrapper[4897]: I1121 14:09:17.756856 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:17Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:17 crc kubenswrapper[4897]: I1121 14:09:17.772157 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6ljgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f3b70b5-9294-4f69-9528-500d28f34c89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://527cf28abdeef1b42d9c19fb88e3171dc05c4516c3a542a461b5b1daa85c6d25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78jhb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6ljgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:17Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:17 crc kubenswrapper[4897]: I1121 14:09:17.834230 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6cd330fd-e8a6-49d5-bfc9-5f6122ccd49e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1bd64b9b938bc80284bcee54141e90cad613d339c4f27ef2934b29bc2a318916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://936f59996d26e42099dcd055f87fff9e3ac4283b57e1956fe97463f2334152d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca74b28a069dc2e609b0779164c3aa470a15ed9848af0dd5fc3479aaf750157b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://270c53b9a342f26be077a01cd4e87e9ccd0eb5d2c84208ae38cbf720463245e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66c521c26d541bbddb064dd9710f1a7b41090d2f47343977922272dfc6dab7af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6eb2ac410b49098e5b3603de5e6bf5ed4484fab0b300021d0bed2e5ed8a42527\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6eb2ac410b49098e5b3603de5e6bf5ed4484fab0b300021d0bed2e5ed8a42527\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03385e0c78e4806f60c160227ef4905a11003ed3d70f0e5c0dd0a477979d1d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03385e0c78e4806f60c160227ef4905a11003ed3d70f0e5c0dd0a477979d1d4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://50aea2fa8b3b141b4330ddce2c8e0715230c6f06dab64b3abfd98a3067224bec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50aea2fa8b3b141b4330ddce2c8e0715230c6f06dab64b3abfd98a3067224bec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:17Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:17 crc kubenswrapper[4897]: I1121 14:09:17.856828 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:17 crc kubenswrapper[4897]: I1121 14:09:17.856877 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:17 crc kubenswrapper[4897]: I1121 14:09:17.856889 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:17 crc kubenswrapper[4897]: I1121 14:09:17.856909 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:17 crc kubenswrapper[4897]: I1121 14:09:17.856924 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:17Z","lastTransitionTime":"2025-11-21T14:09:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:17 crc kubenswrapper[4897]: I1121 14:09:17.960410 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:17 crc kubenswrapper[4897]: I1121 14:09:17.960490 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:17 crc kubenswrapper[4897]: I1121 14:09:17.960527 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:17 crc kubenswrapper[4897]: I1121 14:09:17.960554 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:17 crc kubenswrapper[4897]: I1121 14:09:17.960573 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:17Z","lastTransitionTime":"2025-11-21T14:09:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:18 crc kubenswrapper[4897]: I1121 14:09:18.063674 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:18 crc kubenswrapper[4897]: I1121 14:09:18.063714 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:18 crc kubenswrapper[4897]: I1121 14:09:18.063726 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:18 crc kubenswrapper[4897]: I1121 14:09:18.063742 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:18 crc kubenswrapper[4897]: I1121 14:09:18.063754 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:18Z","lastTransitionTime":"2025-11-21T14:09:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:18 crc kubenswrapper[4897]: I1121 14:09:18.166298 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:18 crc kubenswrapper[4897]: I1121 14:09:18.166345 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:18 crc kubenswrapper[4897]: I1121 14:09:18.166355 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:18 crc kubenswrapper[4897]: I1121 14:09:18.166373 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:18 crc kubenswrapper[4897]: I1121 14:09:18.166386 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:18Z","lastTransitionTime":"2025-11-21T14:09:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:18 crc kubenswrapper[4897]: I1121 14:09:18.269903 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:18 crc kubenswrapper[4897]: I1121 14:09:18.269964 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:18 crc kubenswrapper[4897]: I1121 14:09:18.269981 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:18 crc kubenswrapper[4897]: I1121 14:09:18.270008 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:18 crc kubenswrapper[4897]: I1121 14:09:18.270032 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:18Z","lastTransitionTime":"2025-11-21T14:09:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:18 crc kubenswrapper[4897]: I1121 14:09:18.372818 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:18 crc kubenswrapper[4897]: I1121 14:09:18.372865 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:18 crc kubenswrapper[4897]: I1121 14:09:18.372878 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:18 crc kubenswrapper[4897]: I1121 14:09:18.372896 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:18 crc kubenswrapper[4897]: I1121 14:09:18.372907 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:18Z","lastTransitionTime":"2025-11-21T14:09:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:18 crc kubenswrapper[4897]: I1121 14:09:18.374430 4897 generic.go:334] "Generic (PLEG): container finished" podID="5a7d6101-d0d2-4828-b3ce-d01040c04640" containerID="a153389a4137ee580e5118f3248400460f9d014c992a956f508815755f583afb" exitCode=0 Nov 21 14:09:18 crc kubenswrapper[4897]: I1121 14:09:18.374559 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-nl76g" event={"ID":"5a7d6101-d0d2-4828-b3ce-d01040c04640","Type":"ContainerDied","Data":"a153389a4137ee580e5118f3248400460f9d014c992a956f508815755f583afb"} Nov 21 14:09:18 crc kubenswrapper[4897]: I1121 14:09:18.390885 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48c6b1ea7497b33c00b956e400426c1d4f3c6c66e0225133434460800eef9314\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ea6068a6b6f8123b402de354e5374a0f02a7191a922d1718ab38f8898f47c52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:18Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:18 crc kubenswrapper[4897]: I1121 14:09:18.405352 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4qwqj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f46e2ba8-9ee0-406d-a668-2ba424618b69\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f43aa89d4d9defcb50d0dc6f90293b5f426b9cfc398b780c2e10f6a5c9f1e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbsgb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4qwqj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:18Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:18 crc kubenswrapper[4897]: I1121 14:09:18.421653 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:18Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:18 crc kubenswrapper[4897]: I1121 14:09:18.434273 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7670227-d280-4847-b882-754429f56b0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://961961696b8750c248598f4cda032b18395e5c802275dcfe6d3e9f7ef3510f1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5854028821e6a360a259427b68e7bb4c2a7220e0c68f4cc91639b2ece0540a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-krv5b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:18Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:18 crc kubenswrapper[4897]: I1121 14:09:18.459399 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"222f65f2-0b82-4760-bcf7-779244720f01\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92f1d932a83d65ba6a9a906cc489ff51e76547d07d42656630c59421f833c4db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccd51aa0faf3b1327f207b355cddf86a7353c0051f7e5ce0396f3c4bc67e3b16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8971c38daad45bda3d61545f52120dad40992dbf2e1d6e1e776500ed58d73f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfee19f82e35869a641318a0c45bc7a2ae71d80039f5790e60e277be15c9d3ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec279f2f558a8ac653006321b18cbaf13185876fe303c5cc1c65988aef09edbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de095f3c760b13d5197f624eac3806a1f4186577f2b95eb522532d453eb266e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://089c9d4e6d0fb1a4e9a65a34a033e15c12508d2dbe3b9086b464914a1dac41e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b486a86e0e97ab409e590f30a1bf5cb9f584a7c67f96ff1236e42ca4f8675b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7877db6f5c2bf0eb31218c64d67bca48f921ff7ac62b4e495f29024c086bf738\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7877db6f5c2bf0eb31218c64d67bca48f921ff7ac62b4e495f29024c086bf738\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tpfb6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:18Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:18 crc kubenswrapper[4897]: I1121 14:09:18.473232 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7qfmq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"223b7002-3e67-4848-b1fb-db2921deb64e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381a46c6290da2b36e572f46c16f1a50aa3e25359bcfbc02c7a5a18e1484860e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2m27\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7qfmq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:18Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:18 crc kubenswrapper[4897]: I1121 14:09:18.475972 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:18 crc kubenswrapper[4897]: I1121 14:09:18.476004 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:18 crc kubenswrapper[4897]: I1121 14:09:18.476014 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:18 crc kubenswrapper[4897]: I1121 14:09:18.476030 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:18 crc kubenswrapper[4897]: I1121 14:09:18.476043 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:18Z","lastTransitionTime":"2025-11-21T14:09:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:18 crc kubenswrapper[4897]: I1121 14:09:18.488871 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a6c0a99-565b-41f9-a8b5-c8990a0c8eb3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://737553dd1d16e2019ea89a376e3a868f9792ed5239aa7d6cc40a05d3d4f286ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6682a8886d2b9bce069cf9261d4701ef6907f19574fd035bd89bd05626023e5c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de079956e3d8de0101f3f87dcc120ef23876c4972382639e419b3f34a38cdfca\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1907bfbad7eb34da642e30d6aa81820aab457f92347333a364ac12396b0e9dae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1907bfbad7eb34da642e30d6aa81820aab457f92347333a364ac12396b0e9dae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"le observer\\\\nW1121 14:09:07.994401 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1121 14:09:07.994601 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1121 14:09:07.995290 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1461162674/tls.crt::/tmp/serving-cert-1461162674/tls.key\\\\\\\"\\\\nI1121 14:09:08.162382 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1121 14:09:08.165301 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1121 14:09:08.165318 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1121 14:09:08.165343 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1121 14:09:08.165348 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1121 14:09:08.172015 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1121 14:09:08.172083 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1121 14:09:08.172089 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1121 14:09:08.172094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1121 14:09:08.172098 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1121 14:09:08.172102 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1121 14:09:08.172106 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1121 14:09:08.172023 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1121 14:09:08.174214 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d78c87c440c7d8e30075f6143111b3f504fb99686484b8f19a9c8d61a55eccc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://101fb592fef069af009c613a1e9c9825024ed205498ab72334211ba90d6f7397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://101fb592fef069af009c613a1e9c9825024ed205498ab72334211ba90d6f7397\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:18Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:18 crc kubenswrapper[4897]: I1121 14:09:18.501582 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b78c993a-9966-4e84-85a1-b4f73aaf17f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce3a2602338a1c1e9380ee55f78360f3d9cf6ccc62678886dc5ae574a72e3d9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c256823f2fb49ecc121f6ef564cd86196f9b4a473e9bdea670173596ef14a823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c65f71f8746044f5d41687c670a0308cd09dc41a6531f9b5720e852b3d240756\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1dbe88e7e6ad63899e107ebb884a16edb9a42b21ae7c86d7fbfaf89785b0d36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:18Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:18 crc kubenswrapper[4897]: I1121 14:09:18.512420 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09877d9780caddd5f7f73e95fdf818a427ed645ed5997984113151f4b4759c13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:18Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:18 crc kubenswrapper[4897]: I1121 14:09:18.524799 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:18Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:18 crc kubenswrapper[4897]: I1121 14:09:18.538778 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6ljgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f3b70b5-9294-4f69-9528-500d28f34c89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://527cf28abdeef1b42d9c19fb88e3171dc05c4516c3a542a461b5b1daa85c6d25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78jhb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6ljgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:18Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:18 crc kubenswrapper[4897]: I1121 14:09:18.560910 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6cd330fd-e8a6-49d5-bfc9-5f6122ccd49e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1bd64b9b938bc80284bcee54141e90cad613d339c4f27ef2934b29bc2a318916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://936f59996d26e42099dcd055f87fff9e3ac4283b57e1956fe97463f2334152d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca74b28a069dc2e609b0779164c3aa470a15ed9848af0dd5fc3479aaf750157b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://270c53b9a342f26be077a01cd4e87e9ccd0eb5d2c84208ae38cbf720463245e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66c521c26d541bbddb064dd9710f1a7b41090d2f47343977922272dfc6dab7af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6eb2ac410b49098e5b3603de5e6bf5ed4484fab0b300021d0bed2e5ed8a42527\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6eb2ac410b49098e5b3603de5e6bf5ed4484fab0b300021d0bed2e5ed8a42527\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03385e0c78e4806f60c160227ef4905a11003ed3d70f0e5c0dd0a477979d1d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03385e0c78e4806f60c160227ef4905a11003ed3d70f0e5c0dd0a477979d1d4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://50aea2fa8b3b141b4330ddce2c8e0715230c6f06dab64b3abfd98a3067224bec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50aea2fa8b3b141b4330ddce2c8e0715230c6f06dab64b3abfd98a3067224bec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:18Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:18 crc kubenswrapper[4897]: I1121 14:09:18.574992 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c976d804a7ebeba2c879e920cbc389f96f7f5def4b4439e8c8980d5460ecdb65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:18Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:18 crc kubenswrapper[4897]: I1121 14:09:18.582075 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:18 crc kubenswrapper[4897]: I1121 14:09:18.582125 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:18 crc kubenswrapper[4897]: I1121 14:09:18.582138 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:18 crc kubenswrapper[4897]: I1121 14:09:18.582156 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:18 crc kubenswrapper[4897]: I1121 14:09:18.582172 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:18Z","lastTransitionTime":"2025-11-21T14:09:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:18 crc kubenswrapper[4897]: I1121 14:09:18.594128 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nl76g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a7d6101-d0d2-4828-b3ce-d01040c04640\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e56402f62821baf335f378084f4ed7c21b60ccbafe148419a0a9a403f106ad19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e56402f62821baf335f378084f4ed7c21b60ccbafe148419a0a9a403f106ad19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6022ff28e83dafcd846a7f03b090262aa3229a0da89f1984e4b6a054a04ac24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6022ff28e83dafcd846a7f03b090262aa3229a0da89f1984e4b6a054a04ac24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89de56e32b796da11bf89dd598d04fbf2884745ddf0f5d87cccfcaf1c001d8d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89de56e32b796da11bf89dd598d04fbf2884745ddf0f5d87cccfcaf1c001d8d6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2dff3cacaba7f9238f482e67b670ea775e0f96621434b99f4e76c2dfdde0321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2dff3cacaba7f9238f482e67b670ea775e0f96621434b99f4e76c2dfdde0321\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a153389a4137ee580e5118f3248400460f9d014c992a956f508815755f583afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a153389a4137ee580e5118f3248400460f9d014c992a956f508815755f583afb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nl76g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:18Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:18 crc kubenswrapper[4897]: I1121 14:09:18.609293 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:18Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:18 crc kubenswrapper[4897]: I1121 14:09:18.685045 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:18 crc kubenswrapper[4897]: I1121 14:09:18.685081 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:18 crc kubenswrapper[4897]: I1121 14:09:18.685089 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:18 crc kubenswrapper[4897]: I1121 14:09:18.685104 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:18 crc kubenswrapper[4897]: I1121 14:09:18.685113 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:18Z","lastTransitionTime":"2025-11-21T14:09:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:18 crc kubenswrapper[4897]: I1121 14:09:18.787707 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:18 crc kubenswrapper[4897]: I1121 14:09:18.787753 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:18 crc kubenswrapper[4897]: I1121 14:09:18.787762 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:18 crc kubenswrapper[4897]: I1121 14:09:18.787778 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:18 crc kubenswrapper[4897]: I1121 14:09:18.787790 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:18Z","lastTransitionTime":"2025-11-21T14:09:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:18 crc kubenswrapper[4897]: I1121 14:09:18.889959 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:18 crc kubenswrapper[4897]: I1121 14:09:18.890227 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:18 crc kubenswrapper[4897]: I1121 14:09:18.890295 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:18 crc kubenswrapper[4897]: I1121 14:09:18.890400 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:18 crc kubenswrapper[4897]: I1121 14:09:18.890605 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:18Z","lastTransitionTime":"2025-11-21T14:09:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:18 crc kubenswrapper[4897]: I1121 14:09:18.993128 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:18 crc kubenswrapper[4897]: I1121 14:09:18.993178 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:18 crc kubenswrapper[4897]: I1121 14:09:18.993189 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:18 crc kubenswrapper[4897]: I1121 14:09:18.993204 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:18 crc kubenswrapper[4897]: I1121 14:09:18.993216 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:18Z","lastTransitionTime":"2025-11-21T14:09:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:19 crc kubenswrapper[4897]: I1121 14:09:19.088905 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 21 14:09:19 crc kubenswrapper[4897]: I1121 14:09:19.088943 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 21 14:09:19 crc kubenswrapper[4897]: I1121 14:09:19.088976 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 21 14:09:19 crc kubenswrapper[4897]: E1121 14:09:19.089077 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 21 14:09:19 crc kubenswrapper[4897]: E1121 14:09:19.089165 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 21 14:09:19 crc kubenswrapper[4897]: E1121 14:09:19.089336 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 21 14:09:19 crc kubenswrapper[4897]: I1121 14:09:19.095255 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:19 crc kubenswrapper[4897]: I1121 14:09:19.095287 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:19 crc kubenswrapper[4897]: I1121 14:09:19.095295 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:19 crc kubenswrapper[4897]: I1121 14:09:19.095308 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:19 crc kubenswrapper[4897]: I1121 14:09:19.095319 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:19Z","lastTransitionTime":"2025-11-21T14:09:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:19 crc kubenswrapper[4897]: I1121 14:09:19.198402 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:19 crc kubenswrapper[4897]: I1121 14:09:19.198439 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:19 crc kubenswrapper[4897]: I1121 14:09:19.198448 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:19 crc kubenswrapper[4897]: I1121 14:09:19.198462 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:19 crc kubenswrapper[4897]: I1121 14:09:19.198472 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:19Z","lastTransitionTime":"2025-11-21T14:09:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:19 crc kubenswrapper[4897]: I1121 14:09:19.301577 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:19 crc kubenswrapper[4897]: I1121 14:09:19.301632 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:19 crc kubenswrapper[4897]: I1121 14:09:19.301642 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:19 crc kubenswrapper[4897]: I1121 14:09:19.301659 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:19 crc kubenswrapper[4897]: I1121 14:09:19.301670 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:19Z","lastTransitionTime":"2025-11-21T14:09:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:19 crc kubenswrapper[4897]: I1121 14:09:19.384949 4897 generic.go:334] "Generic (PLEG): container finished" podID="5a7d6101-d0d2-4828-b3ce-d01040c04640" containerID="3a8c051ad1c1f2373f79eb0781fa6db489aeea3d52ed6c0bb43eec11520ec2b2" exitCode=0 Nov 21 14:09:19 crc kubenswrapper[4897]: I1121 14:09:19.385019 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-nl76g" event={"ID":"5a7d6101-d0d2-4828-b3ce-d01040c04640","Type":"ContainerDied","Data":"3a8c051ad1c1f2373f79eb0781fa6db489aeea3d52ed6c0bb43eec11520ec2b2"} Nov 21 14:09:19 crc kubenswrapper[4897]: I1121 14:09:19.404450 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:19 crc kubenswrapper[4897]: I1121 14:09:19.404818 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:19 crc kubenswrapper[4897]: I1121 14:09:19.404840 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:19 crc kubenswrapper[4897]: I1121 14:09:19.404868 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:19 crc kubenswrapper[4897]: I1121 14:09:19.404881 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:19Z","lastTransitionTime":"2025-11-21T14:09:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:19 crc kubenswrapper[4897]: I1121 14:09:19.407673 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:19Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:19 crc kubenswrapper[4897]: I1121 14:09:19.430434 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7670227-d280-4847-b882-754429f56b0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://961961696b8750c248598f4cda032b18395e5c802275dcfe6d3e9f7ef3510f1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5854028821e6a360a259427b68e7bb4c2a7220e0c68f4cc91639b2ece0540a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-krv5b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:19Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:19 crc kubenswrapper[4897]: I1121 14:09:19.450626 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"222f65f2-0b82-4760-bcf7-779244720f01\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92f1d932a83d65ba6a9a906cc489ff51e76547d07d42656630c59421f833c4db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccd51aa0faf3b1327f207b355cddf86a7353c0051f7e5ce0396f3c4bc67e3b16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8971c38daad45bda3d61545f52120dad40992dbf2e1d6e1e776500ed58d73f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfee19f82e35869a641318a0c45bc7a2ae71d80039f5790e60e277be15c9d3ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec279f2f558a8ac653006321b18cbaf13185876fe303c5cc1c65988aef09edbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de095f3c760b13d5197f624eac3806a1f4186577f2b95eb522532d453eb266e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://089c9d4e6d0fb1a4e9a65a34a033e15c12508d2dbe3b9086b464914a1dac41e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b486a86e0e97ab409e590f30a1bf5cb9f584a7c67f96ff1236e42ca4f8675b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7877db6f5c2bf0eb31218c64d67bca48f921ff7ac62b4e495f29024c086bf738\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7877db6f5c2bf0eb31218c64d67bca48f921ff7ac62b4e495f29024c086bf738\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tpfb6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:19Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:19 crc kubenswrapper[4897]: I1121 14:09:19.462018 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7qfmq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"223b7002-3e67-4848-b1fb-db2921deb64e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381a46c6290da2b36e572f46c16f1a50aa3e25359bcfbc02c7a5a18e1484860e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2m27\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7qfmq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:19Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:19 crc kubenswrapper[4897]: I1121 14:09:19.489550 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6cd330fd-e8a6-49d5-bfc9-5f6122ccd49e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1bd64b9b938bc80284bcee54141e90cad613d339c4f27ef2934b29bc2a318916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://936f59996d26e42099dcd055f87fff9e3ac4283b57e1956fe97463f2334152d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca74b28a069dc2e609b0779164c3aa470a15ed9848af0dd5fc3479aaf750157b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://270c53b9a342f26be077a01cd4e87e9ccd0eb5d2c84208ae38cbf720463245e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66c521c26d541bbddb064dd9710f1a7b41090d2f47343977922272dfc6dab7af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6eb2ac410b49098e5b3603de5e6bf5ed4484fab0b300021d0bed2e5ed8a42527\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6eb2ac410b49098e5b3603de5e6bf5ed4484fab0b300021d0bed2e5ed8a42527\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03385e0c78e4806f60c160227ef4905a11003ed3d70f0e5c0dd0a477979d1d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03385e0c78e4806f60c160227ef4905a11003ed3d70f0e5c0dd0a477979d1d4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://50aea2fa8b3b141b4330ddce2c8e0715230c6f06dab64b3abfd98a3067224bec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50aea2fa8b3b141b4330ddce2c8e0715230c6f06dab64b3abfd98a3067224bec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:19Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:19 crc kubenswrapper[4897]: I1121 14:09:19.504832 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a6c0a99-565b-41f9-a8b5-c8990a0c8eb3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://737553dd1d16e2019ea89a376e3a868f9792ed5239aa7d6cc40a05d3d4f286ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6682a8886d2b9bce069cf9261d4701ef6907f19574fd035bd89bd05626023e5c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de079956e3d8de0101f3f87dcc120ef23876c4972382639e419b3f34a38cdfca\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1907bfbad7eb34da642e30d6aa81820aab457f92347333a364ac12396b0e9dae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1907bfbad7eb34da642e30d6aa81820aab457f92347333a364ac12396b0e9dae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"le observer\\\\nW1121 14:09:07.994401 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1121 14:09:07.994601 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1121 14:09:07.995290 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1461162674/tls.crt::/tmp/serving-cert-1461162674/tls.key\\\\\\\"\\\\nI1121 14:09:08.162382 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1121 14:09:08.165301 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1121 14:09:08.165318 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1121 14:09:08.165343 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1121 14:09:08.165348 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1121 14:09:08.172015 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1121 14:09:08.172083 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1121 14:09:08.172089 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1121 14:09:08.172094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1121 14:09:08.172098 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1121 14:09:08.172102 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1121 14:09:08.172106 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1121 14:09:08.172023 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1121 14:09:08.174214 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d78c87c440c7d8e30075f6143111b3f504fb99686484b8f19a9c8d61a55eccc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://101fb592fef069af009c613a1e9c9825024ed205498ab72334211ba90d6f7397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://101fb592fef069af009c613a1e9c9825024ed205498ab72334211ba90d6f7397\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:19Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:19 crc kubenswrapper[4897]: I1121 14:09:19.507880 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:19 crc kubenswrapper[4897]: I1121 14:09:19.507944 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:19 crc kubenswrapper[4897]: I1121 14:09:19.507956 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:19 crc kubenswrapper[4897]: I1121 14:09:19.507970 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:19 crc kubenswrapper[4897]: I1121 14:09:19.507982 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:19Z","lastTransitionTime":"2025-11-21T14:09:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:19 crc kubenswrapper[4897]: I1121 14:09:19.516973 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b78c993a-9966-4e84-85a1-b4f73aaf17f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce3a2602338a1c1e9380ee55f78360f3d9cf6ccc62678886dc5ae574a72e3d9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c256823f2fb49ecc121f6ef564cd86196f9b4a473e9bdea670173596ef14a823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c65f71f8746044f5d41687c670a0308cd09dc41a6531f9b5720e852b3d240756\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1dbe88e7e6ad63899e107ebb884a16edb9a42b21ae7c86d7fbfaf89785b0d36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:19Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:19 crc kubenswrapper[4897]: I1121 14:09:19.530935 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09877d9780caddd5f7f73e95fdf818a427ed645ed5997984113151f4b4759c13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:19Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:19 crc kubenswrapper[4897]: I1121 14:09:19.543817 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:19Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:19 crc kubenswrapper[4897]: I1121 14:09:19.559307 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6ljgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f3b70b5-9294-4f69-9528-500d28f34c89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://527cf28abdeef1b42d9c19fb88e3171dc05c4516c3a542a461b5b1daa85c6d25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78jhb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6ljgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:19Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:19 crc kubenswrapper[4897]: I1121 14:09:19.573785 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:19Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:19 crc kubenswrapper[4897]: I1121 14:09:19.588167 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c976d804a7ebeba2c879e920cbc389f96f7f5def4b4439e8c8980d5460ecdb65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:19Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:19 crc kubenswrapper[4897]: I1121 14:09:19.603432 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nl76g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a7d6101-d0d2-4828-b3ce-d01040c04640\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e56402f62821baf335f378084f4ed7c21b60ccbafe148419a0a9a403f106ad19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e56402f62821baf335f378084f4ed7c21b60ccbafe148419a0a9a403f106ad19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6022ff28e83dafcd846a7f03b090262aa3229a0da89f1984e4b6a054a04ac24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6022ff28e83dafcd846a7f03b090262aa3229a0da89f1984e4b6a054a04ac24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89de56e32b796da11bf89dd598d04fbf2884745ddf0f5d87cccfcaf1c001d8d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89de56e32b796da11bf89dd598d04fbf2884745ddf0f5d87cccfcaf1c001d8d6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2dff3cacaba7f9238f482e67b670ea775e0f96621434b99f4e76c2dfdde0321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2dff3cacaba7f9238f482e67b670ea775e0f96621434b99f4e76c2dfdde0321\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a153389a4137ee580e5118f3248400460f9d014c992a956f508815755f583afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a153389a4137ee580e5118f3248400460f9d014c992a956f508815755f583afb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a8c051ad1c1f2373f79eb0781fa6db489aeea3d52ed6c0bb43eec11520ec2b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a8c051ad1c1f2373f79eb0781fa6db489aeea3d52ed6c0bb43eec11520ec2b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nl76g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:19Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:19 crc kubenswrapper[4897]: I1121 14:09:19.610612 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:19 crc kubenswrapper[4897]: I1121 14:09:19.610664 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:19 crc kubenswrapper[4897]: I1121 14:09:19.610677 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:19 crc kubenswrapper[4897]: I1121 14:09:19.610699 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:19 crc kubenswrapper[4897]: I1121 14:09:19.610713 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:19Z","lastTransitionTime":"2025-11-21T14:09:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:19 crc kubenswrapper[4897]: I1121 14:09:19.618828 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48c6b1ea7497b33c00b956e400426c1d4f3c6c66e0225133434460800eef9314\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ea6068a6b6f8123b402de354e5374a0f02a7191a922d1718ab38f8898f47c52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:19Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:19 crc kubenswrapper[4897]: I1121 14:09:19.632066 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4qwqj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f46e2ba8-9ee0-406d-a668-2ba424618b69\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f43aa89d4d9defcb50d0dc6f90293b5f426b9cfc398b780c2e10f6a5c9f1e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbsgb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4qwqj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:19Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:19 crc kubenswrapper[4897]: I1121 14:09:19.714774 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:19 crc kubenswrapper[4897]: I1121 14:09:19.714841 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:19 crc kubenswrapper[4897]: I1121 14:09:19.714855 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:19 crc kubenswrapper[4897]: I1121 14:09:19.714878 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:19 crc kubenswrapper[4897]: I1121 14:09:19.714899 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:19Z","lastTransitionTime":"2025-11-21T14:09:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:19 crc kubenswrapper[4897]: I1121 14:09:19.821677 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:19 crc kubenswrapper[4897]: I1121 14:09:19.821744 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:19 crc kubenswrapper[4897]: I1121 14:09:19.821757 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:19 crc kubenswrapper[4897]: I1121 14:09:19.821775 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:19 crc kubenswrapper[4897]: I1121 14:09:19.821805 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:19Z","lastTransitionTime":"2025-11-21T14:09:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:19 crc kubenswrapper[4897]: I1121 14:09:19.924538 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:19 crc kubenswrapper[4897]: I1121 14:09:19.924572 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:19 crc kubenswrapper[4897]: I1121 14:09:19.924580 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:19 crc kubenswrapper[4897]: I1121 14:09:19.924594 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:19 crc kubenswrapper[4897]: I1121 14:09:19.924603 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:19Z","lastTransitionTime":"2025-11-21T14:09:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:20 crc kubenswrapper[4897]: I1121 14:09:20.027357 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:20 crc kubenswrapper[4897]: I1121 14:09:20.027398 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:20 crc kubenswrapper[4897]: I1121 14:09:20.027410 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:20 crc kubenswrapper[4897]: I1121 14:09:20.027426 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:20 crc kubenswrapper[4897]: I1121 14:09:20.027436 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:20Z","lastTransitionTime":"2025-11-21T14:09:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:20 crc kubenswrapper[4897]: I1121 14:09:20.129662 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:20 crc kubenswrapper[4897]: I1121 14:09:20.129706 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:20 crc kubenswrapper[4897]: I1121 14:09:20.129714 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:20 crc kubenswrapper[4897]: I1121 14:09:20.129730 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:20 crc kubenswrapper[4897]: I1121 14:09:20.129741 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:20Z","lastTransitionTime":"2025-11-21T14:09:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:20 crc kubenswrapper[4897]: I1121 14:09:20.231845 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:20 crc kubenswrapper[4897]: I1121 14:09:20.231883 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:20 crc kubenswrapper[4897]: I1121 14:09:20.231893 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:20 crc kubenswrapper[4897]: I1121 14:09:20.231906 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:20 crc kubenswrapper[4897]: I1121 14:09:20.231915 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:20Z","lastTransitionTime":"2025-11-21T14:09:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:20 crc kubenswrapper[4897]: I1121 14:09:20.334384 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:20 crc kubenswrapper[4897]: I1121 14:09:20.334428 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:20 crc kubenswrapper[4897]: I1121 14:09:20.334439 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:20 crc kubenswrapper[4897]: I1121 14:09:20.334456 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:20 crc kubenswrapper[4897]: I1121 14:09:20.334469 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:20Z","lastTransitionTime":"2025-11-21T14:09:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:20 crc kubenswrapper[4897]: I1121 14:09:20.392832 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-nl76g" event={"ID":"5a7d6101-d0d2-4828-b3ce-d01040c04640","Type":"ContainerStarted","Data":"ac58b6ef1bd08e5f950260c18120c72e50b90a0de182c6ab0a5dbbbe45cf7328"} Nov 21 14:09:20 crc kubenswrapper[4897]: I1121 14:09:20.408946 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48c6b1ea7497b33c00b956e400426c1d4f3c6c66e0225133434460800eef9314\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ea6068a6b6f8123b402de354e5374a0f02a7191a922d1718ab38f8898f47c52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:20Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:20 crc kubenswrapper[4897]: I1121 14:09:20.424705 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4qwqj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f46e2ba8-9ee0-406d-a668-2ba424618b69\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f43aa89d4d9defcb50d0dc6f90293b5f426b9cfc398b780c2e10f6a5c9f1e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbsgb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4qwqj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:20Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:20 crc kubenswrapper[4897]: I1121 14:09:20.436772 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:20 crc kubenswrapper[4897]: I1121 14:09:20.436819 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:20 crc kubenswrapper[4897]: I1121 14:09:20.436829 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:20 crc kubenswrapper[4897]: I1121 14:09:20.436851 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:20 crc kubenswrapper[4897]: I1121 14:09:20.436864 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:20Z","lastTransitionTime":"2025-11-21T14:09:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:20 crc kubenswrapper[4897]: I1121 14:09:20.438744 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:20Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:20 crc kubenswrapper[4897]: I1121 14:09:20.452821 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7670227-d280-4847-b882-754429f56b0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://961961696b8750c248598f4cda032b18395e5c802275dcfe6d3e9f7ef3510f1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5854028821e6a360a259427b68e7bb4c2a7220e0c68f4cc91639b2ece0540a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-krv5b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:20Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:20 crc kubenswrapper[4897]: I1121 14:09:20.471118 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"222f65f2-0b82-4760-bcf7-779244720f01\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92f1d932a83d65ba6a9a906cc489ff51e76547d07d42656630c59421f833c4db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccd51aa0faf3b1327f207b355cddf86a7353c0051f7e5ce0396f3c4bc67e3b16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8971c38daad45bda3d61545f52120dad40992dbf2e1d6e1e776500ed58d73f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfee19f82e35869a641318a0c45bc7a2ae71d80039f5790e60e277be15c9d3ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec279f2f558a8ac653006321b18cbaf13185876fe303c5cc1c65988aef09edbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de095f3c760b13d5197f624eac3806a1f4186577f2b95eb522532d453eb266e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://089c9d4e6d0fb1a4e9a65a34a033e15c12508d2dbe3b9086b464914a1dac41e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b486a86e0e97ab409e590f30a1bf5cb9f584a7c67f96ff1236e42ca4f8675b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7877db6f5c2bf0eb31218c64d67bca48f921ff7ac62b4e495f29024c086bf738\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7877db6f5c2bf0eb31218c64d67bca48f921ff7ac62b4e495f29024c086bf738\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tpfb6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:20Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:20 crc kubenswrapper[4897]: I1121 14:09:20.481982 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7qfmq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"223b7002-3e67-4848-b1fb-db2921deb64e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381a46c6290da2b36e572f46c16f1a50aa3e25359bcfbc02c7a5a18e1484860e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2m27\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7qfmq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:20Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:20 crc kubenswrapper[4897]: I1121 14:09:20.500411 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6cd330fd-e8a6-49d5-bfc9-5f6122ccd49e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1bd64b9b938bc80284bcee54141e90cad613d339c4f27ef2934b29bc2a318916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://936f59996d26e42099dcd055f87fff9e3ac4283b57e1956fe97463f2334152d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca74b28a069dc2e609b0779164c3aa470a15ed9848af0dd5fc3479aaf750157b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://270c53b9a342f26be077a01cd4e87e9ccd0eb5d2c84208ae38cbf720463245e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66c521c26d541bbddb064dd9710f1a7b41090d2f47343977922272dfc6dab7af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6eb2ac410b49098e5b3603de5e6bf5ed4484fab0b300021d0bed2e5ed8a42527\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6eb2ac410b49098e5b3603de5e6bf5ed4484fab0b300021d0bed2e5ed8a42527\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03385e0c78e4806f60c160227ef4905a11003ed3d70f0e5c0dd0a477979d1d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03385e0c78e4806f60c160227ef4905a11003ed3d70f0e5c0dd0a477979d1d4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://50aea2fa8b3b141b4330ddce2c8e0715230c6f06dab64b3abfd98a3067224bec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50aea2fa8b3b141b4330ddce2c8e0715230c6f06dab64b3abfd98a3067224bec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:20Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:20 crc kubenswrapper[4897]: I1121 14:09:20.514343 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a6c0a99-565b-41f9-a8b5-c8990a0c8eb3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://737553dd1d16e2019ea89a376e3a868f9792ed5239aa7d6cc40a05d3d4f286ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6682a8886d2b9bce069cf9261d4701ef6907f19574fd035bd89bd05626023e5c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de079956e3d8de0101f3f87dcc120ef23876c4972382639e419b3f34a38cdfca\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1907bfbad7eb34da642e30d6aa81820aab457f92347333a364ac12396b0e9dae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1907bfbad7eb34da642e30d6aa81820aab457f92347333a364ac12396b0e9dae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"le observer\\\\nW1121 14:09:07.994401 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1121 14:09:07.994601 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1121 14:09:07.995290 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1461162674/tls.crt::/tmp/serving-cert-1461162674/tls.key\\\\\\\"\\\\nI1121 14:09:08.162382 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1121 14:09:08.165301 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1121 14:09:08.165318 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1121 14:09:08.165343 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1121 14:09:08.165348 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1121 14:09:08.172015 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1121 14:09:08.172083 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1121 14:09:08.172089 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1121 14:09:08.172094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1121 14:09:08.172098 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1121 14:09:08.172102 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1121 14:09:08.172106 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1121 14:09:08.172023 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1121 14:09:08.174214 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d78c87c440c7d8e30075f6143111b3f504fb99686484b8f19a9c8d61a55eccc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://101fb592fef069af009c613a1e9c9825024ed205498ab72334211ba90d6f7397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://101fb592fef069af009c613a1e9c9825024ed205498ab72334211ba90d6f7397\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:20Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:20 crc kubenswrapper[4897]: I1121 14:09:20.526169 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b78c993a-9966-4e84-85a1-b4f73aaf17f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce3a2602338a1c1e9380ee55f78360f3d9cf6ccc62678886dc5ae574a72e3d9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c256823f2fb49ecc121f6ef564cd86196f9b4a473e9bdea670173596ef14a823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c65f71f8746044f5d41687c670a0308cd09dc41a6531f9b5720e852b3d240756\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1dbe88e7e6ad63899e107ebb884a16edb9a42b21ae7c86d7fbfaf89785b0d36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:20Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:20 crc kubenswrapper[4897]: I1121 14:09:20.539035 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:20 crc kubenswrapper[4897]: I1121 14:09:20.539083 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:20 crc kubenswrapper[4897]: I1121 14:09:20.539095 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:20 crc kubenswrapper[4897]: I1121 14:09:20.539113 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:20 crc kubenswrapper[4897]: I1121 14:09:20.539128 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:20Z","lastTransitionTime":"2025-11-21T14:09:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:20 crc kubenswrapper[4897]: I1121 14:09:20.540128 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09877d9780caddd5f7f73e95fdf818a427ed645ed5997984113151f4b4759c13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:20Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:20 crc kubenswrapper[4897]: I1121 14:09:20.552118 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:20Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:20 crc kubenswrapper[4897]: I1121 14:09:20.564769 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6ljgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f3b70b5-9294-4f69-9528-500d28f34c89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://527cf28abdeef1b42d9c19fb88e3171dc05c4516c3a542a461b5b1daa85c6d25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78jhb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6ljgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:20Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:20 crc kubenswrapper[4897]: I1121 14:09:20.580062 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:20Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:20 crc kubenswrapper[4897]: I1121 14:09:20.595039 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c976d804a7ebeba2c879e920cbc389f96f7f5def4b4439e8c8980d5460ecdb65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:20Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:20 crc kubenswrapper[4897]: I1121 14:09:20.611534 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nl76g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a7d6101-d0d2-4828-b3ce-d01040c04640\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac58b6ef1bd08e5f950260c18120c72e50b90a0de182c6ab0a5dbbbe45cf7328\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e56402f62821baf335f378084f4ed7c21b60ccbafe148419a0a9a403f106ad19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e56402f62821baf335f378084f4ed7c21b60ccbafe148419a0a9a403f106ad19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6022ff28e83dafcd846a7f03b090262aa3229a0da89f1984e4b6a054a04ac24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6022ff28e83dafcd846a7f03b090262aa3229a0da89f1984e4b6a054a04ac24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89de56e32b796da11bf89dd598d04fbf2884745ddf0f5d87cccfcaf1c001d8d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89de56e32b796da11bf89dd598d04fbf2884745ddf0f5d87cccfcaf1c001d8d6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2dff3cacaba7f9238f482e67b670ea775e0f96621434b99f4e76c2dfdde0321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2dff3cacaba7f9238f482e67b670ea775e0f96621434b99f4e76c2dfdde0321\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a153389a4137ee580e5118f3248400460f9d014c992a956f508815755f583afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a153389a4137ee580e5118f3248400460f9d014c992a956f508815755f583afb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a8c051ad1c1f2373f79eb0781fa6db489aeea3d52ed6c0bb43eec11520ec2b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a8c051ad1c1f2373f79eb0781fa6db489aeea3d52ed6c0bb43eec11520ec2b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nl76g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:20Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:20 crc kubenswrapper[4897]: I1121 14:09:20.641094 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:20 crc kubenswrapper[4897]: I1121 14:09:20.641135 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:20 crc kubenswrapper[4897]: I1121 14:09:20.641144 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:20 crc kubenswrapper[4897]: I1121 14:09:20.641161 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:20 crc kubenswrapper[4897]: I1121 14:09:20.641175 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:20Z","lastTransitionTime":"2025-11-21T14:09:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:20 crc kubenswrapper[4897]: I1121 14:09:20.743686 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:20 crc kubenswrapper[4897]: I1121 14:09:20.743764 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:20 crc kubenswrapper[4897]: I1121 14:09:20.743777 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:20 crc kubenswrapper[4897]: I1121 14:09:20.743796 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:20 crc kubenswrapper[4897]: I1121 14:09:20.743810 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:20Z","lastTransitionTime":"2025-11-21T14:09:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:20 crc kubenswrapper[4897]: I1121 14:09:20.846562 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:20 crc kubenswrapper[4897]: I1121 14:09:20.846613 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:20 crc kubenswrapper[4897]: I1121 14:09:20.846624 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:20 crc kubenswrapper[4897]: I1121 14:09:20.846644 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:20 crc kubenswrapper[4897]: I1121 14:09:20.846656 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:20Z","lastTransitionTime":"2025-11-21T14:09:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:20 crc kubenswrapper[4897]: I1121 14:09:20.948829 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:20 crc kubenswrapper[4897]: I1121 14:09:20.948899 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:20 crc kubenswrapper[4897]: I1121 14:09:20.948920 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:20 crc kubenswrapper[4897]: I1121 14:09:20.948937 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:20 crc kubenswrapper[4897]: I1121 14:09:20.948947 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:20Z","lastTransitionTime":"2025-11-21T14:09:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:21 crc kubenswrapper[4897]: I1121 14:09:21.051962 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:21 crc kubenswrapper[4897]: I1121 14:09:21.052012 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:21 crc kubenswrapper[4897]: I1121 14:09:21.052021 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:21 crc kubenswrapper[4897]: I1121 14:09:21.052037 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:21 crc kubenswrapper[4897]: I1121 14:09:21.052048 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:21Z","lastTransitionTime":"2025-11-21T14:09:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:21 crc kubenswrapper[4897]: I1121 14:09:21.087989 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 21 14:09:21 crc kubenswrapper[4897]: I1121 14:09:21.088053 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 21 14:09:21 crc kubenswrapper[4897]: I1121 14:09:21.088128 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 21 14:09:21 crc kubenswrapper[4897]: E1121 14:09:21.088246 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 21 14:09:21 crc kubenswrapper[4897]: E1121 14:09:21.088303 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 21 14:09:21 crc kubenswrapper[4897]: E1121 14:09:21.088424 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 21 14:09:21 crc kubenswrapper[4897]: I1121 14:09:21.155006 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:21 crc kubenswrapper[4897]: I1121 14:09:21.155055 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:21 crc kubenswrapper[4897]: I1121 14:09:21.155065 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:21 crc kubenswrapper[4897]: I1121 14:09:21.155081 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:21 crc kubenswrapper[4897]: I1121 14:09:21.155091 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:21Z","lastTransitionTime":"2025-11-21T14:09:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:21 crc kubenswrapper[4897]: I1121 14:09:21.257726 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:21 crc kubenswrapper[4897]: I1121 14:09:21.257769 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:21 crc kubenswrapper[4897]: I1121 14:09:21.257777 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:21 crc kubenswrapper[4897]: I1121 14:09:21.257793 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:21 crc kubenswrapper[4897]: I1121 14:09:21.257803 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:21Z","lastTransitionTime":"2025-11-21T14:09:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:21 crc kubenswrapper[4897]: I1121 14:09:21.360668 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:21 crc kubenswrapper[4897]: I1121 14:09:21.360721 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:21 crc kubenswrapper[4897]: I1121 14:09:21.360736 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:21 crc kubenswrapper[4897]: I1121 14:09:21.360759 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:21 crc kubenswrapper[4897]: I1121 14:09:21.360769 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:21Z","lastTransitionTime":"2025-11-21T14:09:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:21 crc kubenswrapper[4897]: I1121 14:09:21.398129 4897 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tpfb6_222f65f2-0b82-4760-bcf7-779244720f01/ovnkube-controller/0.log" Nov 21 14:09:21 crc kubenswrapper[4897]: I1121 14:09:21.401283 4897 generic.go:334] "Generic (PLEG): container finished" podID="222f65f2-0b82-4760-bcf7-779244720f01" containerID="089c9d4e6d0fb1a4e9a65a34a033e15c12508d2dbe3b9086b464914a1dac41e5" exitCode=1 Nov 21 14:09:21 crc kubenswrapper[4897]: I1121 14:09:21.401321 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" event={"ID":"222f65f2-0b82-4760-bcf7-779244720f01","Type":"ContainerDied","Data":"089c9d4e6d0fb1a4e9a65a34a033e15c12508d2dbe3b9086b464914a1dac41e5"} Nov 21 14:09:21 crc kubenswrapper[4897]: I1121 14:09:21.402026 4897 scope.go:117] "RemoveContainer" containerID="089c9d4e6d0fb1a4e9a65a34a033e15c12508d2dbe3b9086b464914a1dac41e5" Nov 21 14:09:21 crc kubenswrapper[4897]: I1121 14:09:21.418320 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7qfmq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"223b7002-3e67-4848-b1fb-db2921deb64e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381a46c6290da2b36e572f46c16f1a50aa3e25359bcfbc02c7a5a18e1484860e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2m27\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7qfmq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:21Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:21 crc kubenswrapper[4897]: I1121 14:09:21.430100 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:21Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:21 crc kubenswrapper[4897]: I1121 14:09:21.441597 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7670227-d280-4847-b882-754429f56b0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://961961696b8750c248598f4cda032b18395e5c802275dcfe6d3e9f7ef3510f1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5854028821e6a360a259427b68e7bb4c2a7220e0c68f4cc91639b2ece0540a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-krv5b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:21Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:21 crc kubenswrapper[4897]: I1121 14:09:21.460709 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"222f65f2-0b82-4760-bcf7-779244720f01\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92f1d932a83d65ba6a9a906cc489ff51e76547d07d42656630c59421f833c4db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccd51aa0faf3b1327f207b355cddf86a7353c0051f7e5ce0396f3c4bc67e3b16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8971c38daad45bda3d61545f52120dad40992dbf2e1d6e1e776500ed58d73f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfee19f82e35869a641318a0c45bc7a2ae71d80039f5790e60e277be15c9d3ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec279f2f558a8ac653006321b18cbaf13185876fe303c5cc1c65988aef09edbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de095f3c760b13d5197f624eac3806a1f4186577f2b95eb522532d453eb266e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://089c9d4e6d0fb1a4e9a65a34a033e15c12508d2dbe3b9086b464914a1dac41e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://089c9d4e6d0fb1a4e9a65a34a033e15c12508d2dbe3b9086b464914a1dac41e5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-21T14:09:20Z\\\",\\\"message\\\":\\\"60\\\\nI1121 14:09:20.759274 6099 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1121 14:09:20.759278 6099 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1121 14:09:20.759284 6099 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1121 14:09:20.759289 6099 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1121 14:09:20.759296 6099 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1121 14:09:20.759303 6099 handler.go:208] Removed *v1.Node event handler 2\\\\nI1121 14:09:20.759302 6099 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1121 14:09:20.759313 6099 handler.go:208] Removed *v1.Node event handler 7\\\\nI1121 14:09:20.759343 6099 factory.go:656] Stopping watch factory\\\\nI1121 14:09:20.759354 6099 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1121 14:09:20.759360 6099 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1121 14:09:20.759390 6099 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1121 14:09:20.759533 6099 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1121 14:09:20.759593 6099 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b486a86e0e97ab409e590f30a1bf5cb9f584a7c67f96ff1236e42ca4f8675b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7877db6f5c2bf0eb31218c64d67bca48f921ff7ac62b4e495f29024c086bf738\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7877db6f5c2bf0eb31218c64d67bca48f921ff7ac62b4e495f29024c086bf738\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tpfb6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:21Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:21 crc kubenswrapper[4897]: I1121 14:09:21.462484 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:21 crc kubenswrapper[4897]: I1121 14:09:21.462529 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:21 crc kubenswrapper[4897]: I1121 14:09:21.462540 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:21 crc kubenswrapper[4897]: I1121 14:09:21.462554 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:21 crc kubenswrapper[4897]: I1121 14:09:21.462564 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:21Z","lastTransitionTime":"2025-11-21T14:09:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:21 crc kubenswrapper[4897]: I1121 14:09:21.472621 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:21Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:21 crc kubenswrapper[4897]: I1121 14:09:21.485268 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6ljgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f3b70b5-9294-4f69-9528-500d28f34c89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://527cf28abdeef1b42d9c19fb88e3171dc05c4516c3a542a461b5b1daa85c6d25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78jhb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6ljgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:21Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:21 crc kubenswrapper[4897]: I1121 14:09:21.510114 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6cd330fd-e8a6-49d5-bfc9-5f6122ccd49e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1bd64b9b938bc80284bcee54141e90cad613d339c4f27ef2934b29bc2a318916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://936f59996d26e42099dcd055f87fff9e3ac4283b57e1956fe97463f2334152d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca74b28a069dc2e609b0779164c3aa470a15ed9848af0dd5fc3479aaf750157b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://270c53b9a342f26be077a01cd4e87e9ccd0eb5d2c84208ae38cbf720463245e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66c521c26d541bbddb064dd9710f1a7b41090d2f47343977922272dfc6dab7af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6eb2ac410b49098e5b3603de5e6bf5ed4484fab0b300021d0bed2e5ed8a42527\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6eb2ac410b49098e5b3603de5e6bf5ed4484fab0b300021d0bed2e5ed8a42527\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03385e0c78e4806f60c160227ef4905a11003ed3d70f0e5c0dd0a477979d1d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03385e0c78e4806f60c160227ef4905a11003ed3d70f0e5c0dd0a477979d1d4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://50aea2fa8b3b141b4330ddce2c8e0715230c6f06dab64b3abfd98a3067224bec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50aea2fa8b3b141b4330ddce2c8e0715230c6f06dab64b3abfd98a3067224bec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:21Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:21 crc kubenswrapper[4897]: I1121 14:09:21.540396 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a6c0a99-565b-41f9-a8b5-c8990a0c8eb3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://737553dd1d16e2019ea89a376e3a868f9792ed5239aa7d6cc40a05d3d4f286ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6682a8886d2b9bce069cf9261d4701ef6907f19574fd035bd89bd05626023e5c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de079956e3d8de0101f3f87dcc120ef23876c4972382639e419b3f34a38cdfca\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1907bfbad7eb34da642e30d6aa81820aab457f92347333a364ac12396b0e9dae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1907bfbad7eb34da642e30d6aa81820aab457f92347333a364ac12396b0e9dae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"le observer\\\\nW1121 14:09:07.994401 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1121 14:09:07.994601 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1121 14:09:07.995290 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1461162674/tls.crt::/tmp/serving-cert-1461162674/tls.key\\\\\\\"\\\\nI1121 14:09:08.162382 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1121 14:09:08.165301 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1121 14:09:08.165318 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1121 14:09:08.165343 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1121 14:09:08.165348 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1121 14:09:08.172015 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1121 14:09:08.172083 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1121 14:09:08.172089 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1121 14:09:08.172094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1121 14:09:08.172098 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1121 14:09:08.172102 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1121 14:09:08.172106 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1121 14:09:08.172023 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1121 14:09:08.174214 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d78c87c440c7d8e30075f6143111b3f504fb99686484b8f19a9c8d61a55eccc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://101fb592fef069af009c613a1e9c9825024ed205498ab72334211ba90d6f7397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://101fb592fef069af009c613a1e9c9825024ed205498ab72334211ba90d6f7397\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:21Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:21 crc kubenswrapper[4897]: I1121 14:09:21.561855 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b78c993a-9966-4e84-85a1-b4f73aaf17f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce3a2602338a1c1e9380ee55f78360f3d9cf6ccc62678886dc5ae574a72e3d9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c256823f2fb49ecc121f6ef564cd86196f9b4a473e9bdea670173596ef14a823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c65f71f8746044f5d41687c670a0308cd09dc41a6531f9b5720e852b3d240756\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1dbe88e7e6ad63899e107ebb884a16edb9a42b21ae7c86d7fbfaf89785b0d36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:21Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:21 crc kubenswrapper[4897]: I1121 14:09:21.564838 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:21 crc kubenswrapper[4897]: I1121 14:09:21.564876 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:21 crc kubenswrapper[4897]: I1121 14:09:21.564886 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:21 crc kubenswrapper[4897]: I1121 14:09:21.564900 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:21 crc kubenswrapper[4897]: I1121 14:09:21.564910 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:21Z","lastTransitionTime":"2025-11-21T14:09:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:21 crc kubenswrapper[4897]: I1121 14:09:21.583863 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09877d9780caddd5f7f73e95fdf818a427ed645ed5997984113151f4b4759c13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:21Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:21 crc kubenswrapper[4897]: I1121 14:09:21.601102 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:21Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:21 crc kubenswrapper[4897]: I1121 14:09:21.616504 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c976d804a7ebeba2c879e920cbc389f96f7f5def4b4439e8c8980d5460ecdb65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:21Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:21 crc kubenswrapper[4897]: I1121 14:09:21.636891 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nl76g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a7d6101-d0d2-4828-b3ce-d01040c04640\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac58b6ef1bd08e5f950260c18120c72e50b90a0de182c6ab0a5dbbbe45cf7328\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e56402f62821baf335f378084f4ed7c21b60ccbafe148419a0a9a403f106ad19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e56402f62821baf335f378084f4ed7c21b60ccbafe148419a0a9a403f106ad19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6022ff28e83dafcd846a7f03b090262aa3229a0da89f1984e4b6a054a04ac24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6022ff28e83dafcd846a7f03b090262aa3229a0da89f1984e4b6a054a04ac24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89de56e32b796da11bf89dd598d04fbf2884745ddf0f5d87cccfcaf1c001d8d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89de56e32b796da11bf89dd598d04fbf2884745ddf0f5d87cccfcaf1c001d8d6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2dff3cacaba7f9238f482e67b670ea775e0f96621434b99f4e76c2dfdde0321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2dff3cacaba7f9238f482e67b670ea775e0f96621434b99f4e76c2dfdde0321\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a153389a4137ee580e5118f3248400460f9d014c992a956f508815755f583afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a153389a4137ee580e5118f3248400460f9d014c992a956f508815755f583afb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a8c051ad1c1f2373f79eb0781fa6db489aeea3d52ed6c0bb43eec11520ec2b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a8c051ad1c1f2373f79eb0781fa6db489aeea3d52ed6c0bb43eec11520ec2b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nl76g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:21Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:21 crc kubenswrapper[4897]: I1121 14:09:21.652839 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48c6b1ea7497b33c00b956e400426c1d4f3c6c66e0225133434460800eef9314\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ea6068a6b6f8123b402de354e5374a0f02a7191a922d1718ab38f8898f47c52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:21Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:21 crc kubenswrapper[4897]: I1121 14:09:21.662010 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4qwqj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f46e2ba8-9ee0-406d-a668-2ba424618b69\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f43aa89d4d9defcb50d0dc6f90293b5f426b9cfc398b780c2e10f6a5c9f1e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbsgb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4qwqj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:21Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:21 crc kubenswrapper[4897]: I1121 14:09:21.667441 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:21 crc kubenswrapper[4897]: I1121 14:09:21.667469 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:21 crc kubenswrapper[4897]: I1121 14:09:21.667480 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:21 crc kubenswrapper[4897]: I1121 14:09:21.667495 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:21 crc kubenswrapper[4897]: I1121 14:09:21.667524 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:21Z","lastTransitionTime":"2025-11-21T14:09:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:21 crc kubenswrapper[4897]: I1121 14:09:21.770416 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:21 crc kubenswrapper[4897]: I1121 14:09:21.770446 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:21 crc kubenswrapper[4897]: I1121 14:09:21.770454 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:21 crc kubenswrapper[4897]: I1121 14:09:21.770468 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:21 crc kubenswrapper[4897]: I1121 14:09:21.770479 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:21Z","lastTransitionTime":"2025-11-21T14:09:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:21 crc kubenswrapper[4897]: I1121 14:09:21.873486 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:21 crc kubenswrapper[4897]: I1121 14:09:21.873555 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:21 crc kubenswrapper[4897]: I1121 14:09:21.873568 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:21 crc kubenswrapper[4897]: I1121 14:09:21.873584 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:21 crc kubenswrapper[4897]: I1121 14:09:21.873595 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:21Z","lastTransitionTime":"2025-11-21T14:09:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:21 crc kubenswrapper[4897]: I1121 14:09:21.976299 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:21 crc kubenswrapper[4897]: I1121 14:09:21.976347 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:21 crc kubenswrapper[4897]: I1121 14:09:21.976360 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:21 crc kubenswrapper[4897]: I1121 14:09:21.976377 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:21 crc kubenswrapper[4897]: I1121 14:09:21.976391 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:21Z","lastTransitionTime":"2025-11-21T14:09:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:22 crc kubenswrapper[4897]: I1121 14:09:22.078950 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:22 crc kubenswrapper[4897]: I1121 14:09:22.078989 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:22 crc kubenswrapper[4897]: I1121 14:09:22.078997 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:22 crc kubenswrapper[4897]: I1121 14:09:22.079011 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:22 crc kubenswrapper[4897]: I1121 14:09:22.079021 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:22Z","lastTransitionTime":"2025-11-21T14:09:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:22 crc kubenswrapper[4897]: I1121 14:09:22.181573 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:22 crc kubenswrapper[4897]: I1121 14:09:22.181624 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:22 crc kubenswrapper[4897]: I1121 14:09:22.181634 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:22 crc kubenswrapper[4897]: I1121 14:09:22.181651 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:22 crc kubenswrapper[4897]: I1121 14:09:22.181662 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:22Z","lastTransitionTime":"2025-11-21T14:09:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:22 crc kubenswrapper[4897]: I1121 14:09:22.211715 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x5pws"] Nov 21 14:09:22 crc kubenswrapper[4897]: I1121 14:09:22.212271 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x5pws" Nov 21 14:09:22 crc kubenswrapper[4897]: I1121 14:09:22.217841 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Nov 21 14:09:22 crc kubenswrapper[4897]: I1121 14:09:22.218136 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Nov 21 14:09:22 crc kubenswrapper[4897]: I1121 14:09:22.233654 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:22Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:22 crc kubenswrapper[4897]: I1121 14:09:22.250069 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7670227-d280-4847-b882-754429f56b0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://961961696b8750c248598f4cda032b18395e5c802275dcfe6d3e9f7ef3510f1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5854028821e6a360a259427b68e7bb4c2a7220e0c68f4cc91639b2ece0540a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-krv5b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:22Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:22 crc kubenswrapper[4897]: I1121 14:09:22.268038 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"222f65f2-0b82-4760-bcf7-779244720f01\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92f1d932a83d65ba6a9a906cc489ff51e76547d07d42656630c59421f833c4db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccd51aa0faf3b1327f207b355cddf86a7353c0051f7e5ce0396f3c4bc67e3b16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8971c38daad45bda3d61545f52120dad40992dbf2e1d6e1e776500ed58d73f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfee19f82e35869a641318a0c45bc7a2ae71d80039f5790e60e277be15c9d3ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec279f2f558a8ac653006321b18cbaf13185876fe303c5cc1c65988aef09edbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de095f3c760b13d5197f624eac3806a1f4186577f2b95eb522532d453eb266e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://089c9d4e6d0fb1a4e9a65a34a033e15c12508d2dbe3b9086b464914a1dac41e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://089c9d4e6d0fb1a4e9a65a34a033e15c12508d2dbe3b9086b464914a1dac41e5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-21T14:09:20Z\\\",\\\"message\\\":\\\"60\\\\nI1121 14:09:20.759274 6099 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1121 14:09:20.759278 6099 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1121 14:09:20.759284 6099 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1121 14:09:20.759289 6099 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1121 14:09:20.759296 6099 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1121 14:09:20.759303 6099 handler.go:208] Removed *v1.Node event handler 2\\\\nI1121 14:09:20.759302 6099 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1121 14:09:20.759313 6099 handler.go:208] Removed *v1.Node event handler 7\\\\nI1121 14:09:20.759343 6099 factory.go:656] Stopping watch factory\\\\nI1121 14:09:20.759354 6099 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1121 14:09:20.759360 6099 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1121 14:09:20.759390 6099 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1121 14:09:20.759533 6099 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1121 14:09:20.759593 6099 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b486a86e0e97ab409e590f30a1bf5cb9f584a7c67f96ff1236e42ca4f8675b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7877db6f5c2bf0eb31218c64d67bca48f921ff7ac62b4e495f29024c086bf738\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7877db6f5c2bf0eb31218c64d67bca48f921ff7ac62b4e495f29024c086bf738\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tpfb6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:22Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:22 crc kubenswrapper[4897]: I1121 14:09:22.279684 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7qfmq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"223b7002-3e67-4848-b1fb-db2921deb64e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381a46c6290da2b36e572f46c16f1a50aa3e25359bcfbc02c7a5a18e1484860e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2m27\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7qfmq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:22Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:22 crc kubenswrapper[4897]: I1121 14:09:22.284082 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:22 crc kubenswrapper[4897]: I1121 14:09:22.284128 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:22 crc kubenswrapper[4897]: I1121 14:09:22.284140 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:22 crc kubenswrapper[4897]: I1121 14:09:22.284158 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:22 crc kubenswrapper[4897]: I1121 14:09:22.284171 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:22Z","lastTransitionTime":"2025-11-21T14:09:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:22 crc kubenswrapper[4897]: I1121 14:09:22.292556 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x5pws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc9e717e-af03-48bf-b005-62b3d084c6fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9c4f5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9c4f5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-x5pws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:22Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:22 crc kubenswrapper[4897]: I1121 14:09:22.301674 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/dc9e717e-af03-48bf-b005-62b3d084c6fc-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-x5pws\" (UID: \"dc9e717e-af03-48bf-b005-62b3d084c6fc\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x5pws" Nov 21 14:09:22 crc kubenswrapper[4897]: I1121 14:09:22.301725 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/dc9e717e-af03-48bf-b005-62b3d084c6fc-env-overrides\") pod \"ovnkube-control-plane-749d76644c-x5pws\" (UID: \"dc9e717e-af03-48bf-b005-62b3d084c6fc\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x5pws" Nov 21 14:09:22 crc kubenswrapper[4897]: I1121 14:09:22.301771 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/dc9e717e-af03-48bf-b005-62b3d084c6fc-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-x5pws\" (UID: \"dc9e717e-af03-48bf-b005-62b3d084c6fc\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x5pws" Nov 21 14:09:22 crc kubenswrapper[4897]: I1121 14:09:22.301807 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9c4f5\" (UniqueName: \"kubernetes.io/projected/dc9e717e-af03-48bf-b005-62b3d084c6fc-kube-api-access-9c4f5\") pod \"ovnkube-control-plane-749d76644c-x5pws\" (UID: \"dc9e717e-af03-48bf-b005-62b3d084c6fc\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x5pws" Nov 21 14:09:22 crc kubenswrapper[4897]: I1121 14:09:22.313608 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6cd330fd-e8a6-49d5-bfc9-5f6122ccd49e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1bd64b9b938bc80284bcee54141e90cad613d339c4f27ef2934b29bc2a318916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://936f59996d26e42099dcd055f87fff9e3ac4283b57e1956fe97463f2334152d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca74b28a069dc2e609b0779164c3aa470a15ed9848af0dd5fc3479aaf750157b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://270c53b9a342f26be077a01cd4e87e9ccd0eb5d2c84208ae38cbf720463245e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66c521c26d541bbddb064dd9710f1a7b41090d2f47343977922272dfc6dab7af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6eb2ac410b49098e5b3603de5e6bf5ed4484fab0b300021d0bed2e5ed8a42527\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6eb2ac410b49098e5b3603de5e6bf5ed4484fab0b300021d0bed2e5ed8a42527\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03385e0c78e4806f60c160227ef4905a11003ed3d70f0e5c0dd0a477979d1d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03385e0c78e4806f60c160227ef4905a11003ed3d70f0e5c0dd0a477979d1d4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://50aea2fa8b3b141b4330ddce2c8e0715230c6f06dab64b3abfd98a3067224bec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50aea2fa8b3b141b4330ddce2c8e0715230c6f06dab64b3abfd98a3067224bec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:22Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:22 crc kubenswrapper[4897]: I1121 14:09:22.329360 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a6c0a99-565b-41f9-a8b5-c8990a0c8eb3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://737553dd1d16e2019ea89a376e3a868f9792ed5239aa7d6cc40a05d3d4f286ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6682a8886d2b9bce069cf9261d4701ef6907f19574fd035bd89bd05626023e5c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de079956e3d8de0101f3f87dcc120ef23876c4972382639e419b3f34a38cdfca\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1907bfbad7eb34da642e30d6aa81820aab457f92347333a364ac12396b0e9dae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1907bfbad7eb34da642e30d6aa81820aab457f92347333a364ac12396b0e9dae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"le observer\\\\nW1121 14:09:07.994401 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1121 14:09:07.994601 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1121 14:09:07.995290 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1461162674/tls.crt::/tmp/serving-cert-1461162674/tls.key\\\\\\\"\\\\nI1121 14:09:08.162382 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1121 14:09:08.165301 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1121 14:09:08.165318 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1121 14:09:08.165343 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1121 14:09:08.165348 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1121 14:09:08.172015 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1121 14:09:08.172083 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1121 14:09:08.172089 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1121 14:09:08.172094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1121 14:09:08.172098 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1121 14:09:08.172102 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1121 14:09:08.172106 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1121 14:09:08.172023 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1121 14:09:08.174214 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d78c87c440c7d8e30075f6143111b3f504fb99686484b8f19a9c8d61a55eccc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://101fb592fef069af009c613a1e9c9825024ed205498ab72334211ba90d6f7397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://101fb592fef069af009c613a1e9c9825024ed205498ab72334211ba90d6f7397\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:22Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:22 crc kubenswrapper[4897]: I1121 14:09:22.344208 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b78c993a-9966-4e84-85a1-b4f73aaf17f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce3a2602338a1c1e9380ee55f78360f3d9cf6ccc62678886dc5ae574a72e3d9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c256823f2fb49ecc121f6ef564cd86196f9b4a473e9bdea670173596ef14a823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c65f71f8746044f5d41687c670a0308cd09dc41a6531f9b5720e852b3d240756\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1dbe88e7e6ad63899e107ebb884a16edb9a42b21ae7c86d7fbfaf89785b0d36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:22Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:22 crc kubenswrapper[4897]: I1121 14:09:22.358895 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09877d9780caddd5f7f73e95fdf818a427ed645ed5997984113151f4b4759c13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:22Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:22 crc kubenswrapper[4897]: I1121 14:09:22.373163 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:22Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:22 crc kubenswrapper[4897]: I1121 14:09:22.386553 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:22 crc kubenswrapper[4897]: I1121 14:09:22.386634 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:22 crc kubenswrapper[4897]: I1121 14:09:22.386647 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:22 crc kubenswrapper[4897]: I1121 14:09:22.386677 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:22 crc kubenswrapper[4897]: I1121 14:09:22.386694 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:22Z","lastTransitionTime":"2025-11-21T14:09:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:22 crc kubenswrapper[4897]: I1121 14:09:22.387427 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6ljgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f3b70b5-9294-4f69-9528-500d28f34c89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://527cf28abdeef1b42d9c19fb88e3171dc05c4516c3a542a461b5b1daa85c6d25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78jhb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6ljgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:22Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:22 crc kubenswrapper[4897]: I1121 14:09:22.401042 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:22Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:22 crc kubenswrapper[4897]: I1121 14:09:22.402494 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/dc9e717e-af03-48bf-b005-62b3d084c6fc-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-x5pws\" (UID: \"dc9e717e-af03-48bf-b005-62b3d084c6fc\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x5pws" Nov 21 14:09:22 crc kubenswrapper[4897]: I1121 14:09:22.402912 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9c4f5\" (UniqueName: \"kubernetes.io/projected/dc9e717e-af03-48bf-b005-62b3d084c6fc-kube-api-access-9c4f5\") pod \"ovnkube-control-plane-749d76644c-x5pws\" (UID: \"dc9e717e-af03-48bf-b005-62b3d084c6fc\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x5pws" Nov 21 14:09:22 crc kubenswrapper[4897]: I1121 14:09:22.402950 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/dc9e717e-af03-48bf-b005-62b3d084c6fc-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-x5pws\" (UID: \"dc9e717e-af03-48bf-b005-62b3d084c6fc\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x5pws" Nov 21 14:09:22 crc kubenswrapper[4897]: I1121 14:09:22.402971 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/dc9e717e-af03-48bf-b005-62b3d084c6fc-env-overrides\") pod \"ovnkube-control-plane-749d76644c-x5pws\" (UID: \"dc9e717e-af03-48bf-b005-62b3d084c6fc\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x5pws" Nov 21 14:09:22 crc kubenswrapper[4897]: I1121 14:09:22.403568 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/dc9e717e-af03-48bf-b005-62b3d084c6fc-env-overrides\") pod \"ovnkube-control-plane-749d76644c-x5pws\" (UID: \"dc9e717e-af03-48bf-b005-62b3d084c6fc\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x5pws" Nov 21 14:09:22 crc kubenswrapper[4897]: I1121 14:09:22.403832 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/dc9e717e-af03-48bf-b005-62b3d084c6fc-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-x5pws\" (UID: \"dc9e717e-af03-48bf-b005-62b3d084c6fc\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x5pws" Nov 21 14:09:22 crc kubenswrapper[4897]: I1121 14:09:22.407315 4897 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tpfb6_222f65f2-0b82-4760-bcf7-779244720f01/ovnkube-controller/0.log" Nov 21 14:09:22 crc kubenswrapper[4897]: I1121 14:09:22.410168 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" event={"ID":"222f65f2-0b82-4760-bcf7-779244720f01","Type":"ContainerStarted","Data":"814079abfa7d1b029825da6c6c5f1e57d439ff21ccb626c5d4e52e41d059ad3a"} Nov 21 14:09:22 crc kubenswrapper[4897]: I1121 14:09:22.410687 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" Nov 21 14:09:22 crc kubenswrapper[4897]: I1121 14:09:22.410800 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/dc9e717e-af03-48bf-b005-62b3d084c6fc-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-x5pws\" (UID: \"dc9e717e-af03-48bf-b005-62b3d084c6fc\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x5pws" Nov 21 14:09:22 crc kubenswrapper[4897]: I1121 14:09:22.420190 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9c4f5\" (UniqueName: \"kubernetes.io/projected/dc9e717e-af03-48bf-b005-62b3d084c6fc-kube-api-access-9c4f5\") pod \"ovnkube-control-plane-749d76644c-x5pws\" (UID: \"dc9e717e-af03-48bf-b005-62b3d084c6fc\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x5pws" Nov 21 14:09:22 crc kubenswrapper[4897]: I1121 14:09:22.422823 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c976d804a7ebeba2c879e920cbc389f96f7f5def4b4439e8c8980d5460ecdb65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:22Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:22 crc kubenswrapper[4897]: I1121 14:09:22.440969 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nl76g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a7d6101-d0d2-4828-b3ce-d01040c04640\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac58b6ef1bd08e5f950260c18120c72e50b90a0de182c6ab0a5dbbbe45cf7328\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e56402f62821baf335f378084f4ed7c21b60ccbafe148419a0a9a403f106ad19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e56402f62821baf335f378084f4ed7c21b60ccbafe148419a0a9a403f106ad19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6022ff28e83dafcd846a7f03b090262aa3229a0da89f1984e4b6a054a04ac24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6022ff28e83dafcd846a7f03b090262aa3229a0da89f1984e4b6a054a04ac24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89de56e32b796da11bf89dd598d04fbf2884745ddf0f5d87cccfcaf1c001d8d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89de56e32b796da11bf89dd598d04fbf2884745ddf0f5d87cccfcaf1c001d8d6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2dff3cacaba7f9238f482e67b670ea775e0f96621434b99f4e76c2dfdde0321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2dff3cacaba7f9238f482e67b670ea775e0f96621434b99f4e76c2dfdde0321\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a153389a4137ee580e5118f3248400460f9d014c992a956f508815755f583afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a153389a4137ee580e5118f3248400460f9d014c992a956f508815755f583afb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a8c051ad1c1f2373f79eb0781fa6db489aeea3d52ed6c0bb43eec11520ec2b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a8c051ad1c1f2373f79eb0781fa6db489aeea3d52ed6c0bb43eec11520ec2b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nl76g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:22Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:22 crc kubenswrapper[4897]: I1121 14:09:22.457196 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48c6b1ea7497b33c00b956e400426c1d4f3c6c66e0225133434460800eef9314\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ea6068a6b6f8123b402de354e5374a0f02a7191a922d1718ab38f8898f47c52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:22Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:22 crc kubenswrapper[4897]: I1121 14:09:22.470859 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4qwqj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f46e2ba8-9ee0-406d-a668-2ba424618b69\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f43aa89d4d9defcb50d0dc6f90293b5f426b9cfc398b780c2e10f6a5c9f1e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbsgb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4qwqj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:22Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:22 crc kubenswrapper[4897]: I1121 14:09:22.485019 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7qfmq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"223b7002-3e67-4848-b1fb-db2921deb64e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381a46c6290da2b36e572f46c16f1a50aa3e25359bcfbc02c7a5a18e1484860e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2m27\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7qfmq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:22Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:22 crc kubenswrapper[4897]: I1121 14:09:22.488846 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:22 crc kubenswrapper[4897]: I1121 14:09:22.488888 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:22 crc kubenswrapper[4897]: I1121 14:09:22.488898 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:22 crc kubenswrapper[4897]: I1121 14:09:22.488914 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:22 crc kubenswrapper[4897]: I1121 14:09:22.488924 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:22Z","lastTransitionTime":"2025-11-21T14:09:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:22 crc kubenswrapper[4897]: I1121 14:09:22.502125 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:22Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:22 crc kubenswrapper[4897]: I1121 14:09:22.514580 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7670227-d280-4847-b882-754429f56b0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://961961696b8750c248598f4cda032b18395e5c802275dcfe6d3e9f7ef3510f1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5854028821e6a360a259427b68e7bb4c2a7220e0c68f4cc91639b2ece0540a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-krv5b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:22Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:22 crc kubenswrapper[4897]: I1121 14:09:22.527769 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x5pws" Nov 21 14:09:22 crc kubenswrapper[4897]: I1121 14:09:22.540718 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"222f65f2-0b82-4760-bcf7-779244720f01\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92f1d932a83d65ba6a9a906cc489ff51e76547d07d42656630c59421f833c4db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccd51aa0faf3b1327f207b355cddf86a7353c0051f7e5ce0396f3c4bc67e3b16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8971c38daad45bda3d61545f52120dad40992dbf2e1d6e1e776500ed58d73f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfee19f82e35869a641318a0c45bc7a2ae71d80039f5790e60e277be15c9d3ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec279f2f558a8ac653006321b18cbaf13185876fe303c5cc1c65988aef09edbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de095f3c760b13d5197f624eac3806a1f4186577f2b95eb522532d453eb266e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://814079abfa7d1b029825da6c6c5f1e57d439ff21ccb626c5d4e52e41d059ad3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://089c9d4e6d0fb1a4e9a65a34a033e15c12508d2dbe3b9086b464914a1dac41e5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-21T14:09:20Z\\\",\\\"message\\\":\\\"60\\\\nI1121 14:09:20.759274 6099 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1121 14:09:20.759278 6099 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1121 14:09:20.759284 6099 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1121 14:09:20.759289 6099 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1121 14:09:20.759296 6099 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1121 14:09:20.759303 6099 handler.go:208] Removed *v1.Node event handler 2\\\\nI1121 14:09:20.759302 6099 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1121 14:09:20.759313 6099 handler.go:208] Removed *v1.Node event handler 7\\\\nI1121 14:09:20.759343 6099 factory.go:656] Stopping watch factory\\\\nI1121 14:09:20.759354 6099 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1121 14:09:20.759360 6099 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1121 14:09:20.759390 6099 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1121 14:09:20.759533 6099 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1121 14:09:20.759593 6099 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:16Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b486a86e0e97ab409e590f30a1bf5cb9f584a7c67f96ff1236e42ca4f8675b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7877db6f5c2bf0eb31218c64d67bca48f921ff7ac62b4e495f29024c086bf738\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7877db6f5c2bf0eb31218c64d67bca48f921ff7ac62b4e495f29024c086bf738\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tpfb6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:22Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:22 crc kubenswrapper[4897]: I1121 14:09:22.560527 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:22Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:22 crc kubenswrapper[4897]: I1121 14:09:22.572239 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6ljgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f3b70b5-9294-4f69-9528-500d28f34c89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://527cf28abdeef1b42d9c19fb88e3171dc05c4516c3a542a461b5b1daa85c6d25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78jhb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6ljgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:22Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:22 crc kubenswrapper[4897]: I1121 14:09:22.584173 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x5pws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc9e717e-af03-48bf-b005-62b3d084c6fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9c4f5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9c4f5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-x5pws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:22Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:22 crc kubenswrapper[4897]: I1121 14:09:22.591061 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:22 crc kubenswrapper[4897]: I1121 14:09:22.591098 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:22 crc kubenswrapper[4897]: I1121 14:09:22.591110 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:22 crc kubenswrapper[4897]: I1121 14:09:22.591126 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:22 crc kubenswrapper[4897]: I1121 14:09:22.591138 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:22Z","lastTransitionTime":"2025-11-21T14:09:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:22 crc kubenswrapper[4897]: I1121 14:09:22.605950 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6cd330fd-e8a6-49d5-bfc9-5f6122ccd49e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1bd64b9b938bc80284bcee54141e90cad613d339c4f27ef2934b29bc2a318916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://936f59996d26e42099dcd055f87fff9e3ac4283b57e1956fe97463f2334152d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca74b28a069dc2e609b0779164c3aa470a15ed9848af0dd5fc3479aaf750157b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://270c53b9a342f26be077a01cd4e87e9ccd0eb5d2c84208ae38cbf720463245e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66c521c26d541bbddb064dd9710f1a7b41090d2f47343977922272dfc6dab7af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6eb2ac410b49098e5b3603de5e6bf5ed4484fab0b300021d0bed2e5ed8a42527\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6eb2ac410b49098e5b3603de5e6bf5ed4484fab0b300021d0bed2e5ed8a42527\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03385e0c78e4806f60c160227ef4905a11003ed3d70f0e5c0dd0a477979d1d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03385e0c78e4806f60c160227ef4905a11003ed3d70f0e5c0dd0a477979d1d4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://50aea2fa8b3b141b4330ddce2c8e0715230c6f06dab64b3abfd98a3067224bec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50aea2fa8b3b141b4330ddce2c8e0715230c6f06dab64b3abfd98a3067224bec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:22Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:22 crc kubenswrapper[4897]: I1121 14:09:22.619834 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a6c0a99-565b-41f9-a8b5-c8990a0c8eb3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://737553dd1d16e2019ea89a376e3a868f9792ed5239aa7d6cc40a05d3d4f286ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6682a8886d2b9bce069cf9261d4701ef6907f19574fd035bd89bd05626023e5c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de079956e3d8de0101f3f87dcc120ef23876c4972382639e419b3f34a38cdfca\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1907bfbad7eb34da642e30d6aa81820aab457f92347333a364ac12396b0e9dae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1907bfbad7eb34da642e30d6aa81820aab457f92347333a364ac12396b0e9dae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"le observer\\\\nW1121 14:09:07.994401 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1121 14:09:07.994601 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1121 14:09:07.995290 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1461162674/tls.crt::/tmp/serving-cert-1461162674/tls.key\\\\\\\"\\\\nI1121 14:09:08.162382 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1121 14:09:08.165301 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1121 14:09:08.165318 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1121 14:09:08.165343 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1121 14:09:08.165348 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1121 14:09:08.172015 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1121 14:09:08.172083 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1121 14:09:08.172089 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1121 14:09:08.172094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1121 14:09:08.172098 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1121 14:09:08.172102 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1121 14:09:08.172106 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1121 14:09:08.172023 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1121 14:09:08.174214 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d78c87c440c7d8e30075f6143111b3f504fb99686484b8f19a9c8d61a55eccc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://101fb592fef069af009c613a1e9c9825024ed205498ab72334211ba90d6f7397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://101fb592fef069af009c613a1e9c9825024ed205498ab72334211ba90d6f7397\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:22Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:22 crc kubenswrapper[4897]: I1121 14:09:22.631119 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b78c993a-9966-4e84-85a1-b4f73aaf17f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce3a2602338a1c1e9380ee55f78360f3d9cf6ccc62678886dc5ae574a72e3d9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c256823f2fb49ecc121f6ef564cd86196f9b4a473e9bdea670173596ef14a823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c65f71f8746044f5d41687c670a0308cd09dc41a6531f9b5720e852b3d240756\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1dbe88e7e6ad63899e107ebb884a16edb9a42b21ae7c86d7fbfaf89785b0d36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:22Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:22 crc kubenswrapper[4897]: I1121 14:09:22.643992 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09877d9780caddd5f7f73e95fdf818a427ed645ed5997984113151f4b4759c13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:22Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:22 crc kubenswrapper[4897]: I1121 14:09:22.657739 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:22Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:22 crc kubenswrapper[4897]: I1121 14:09:22.671150 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c976d804a7ebeba2c879e920cbc389f96f7f5def4b4439e8c8980d5460ecdb65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:22Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:22 crc kubenswrapper[4897]: I1121 14:09:22.686121 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nl76g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a7d6101-d0d2-4828-b3ce-d01040c04640\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac58b6ef1bd08e5f950260c18120c72e50b90a0de182c6ab0a5dbbbe45cf7328\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e56402f62821baf335f378084f4ed7c21b60ccbafe148419a0a9a403f106ad19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e56402f62821baf335f378084f4ed7c21b60ccbafe148419a0a9a403f106ad19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6022ff28e83dafcd846a7f03b090262aa3229a0da89f1984e4b6a054a04ac24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6022ff28e83dafcd846a7f03b090262aa3229a0da89f1984e4b6a054a04ac24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89de56e32b796da11bf89dd598d04fbf2884745ddf0f5d87cccfcaf1c001d8d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89de56e32b796da11bf89dd598d04fbf2884745ddf0f5d87cccfcaf1c001d8d6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2dff3cacaba7f9238f482e67b670ea775e0f96621434b99f4e76c2dfdde0321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2dff3cacaba7f9238f482e67b670ea775e0f96621434b99f4e76c2dfdde0321\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a153389a4137ee580e5118f3248400460f9d014c992a956f508815755f583afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a153389a4137ee580e5118f3248400460f9d014c992a956f508815755f583afb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a8c051ad1c1f2373f79eb0781fa6db489aeea3d52ed6c0bb43eec11520ec2b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a8c051ad1c1f2373f79eb0781fa6db489aeea3d52ed6c0bb43eec11520ec2b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nl76g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:22Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:22 crc kubenswrapper[4897]: I1121 14:09:22.693830 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:22 crc kubenswrapper[4897]: I1121 14:09:22.693932 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:22 crc kubenswrapper[4897]: I1121 14:09:22.693947 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:22 crc kubenswrapper[4897]: I1121 14:09:22.693967 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:22 crc kubenswrapper[4897]: I1121 14:09:22.693981 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:22Z","lastTransitionTime":"2025-11-21T14:09:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:22 crc kubenswrapper[4897]: I1121 14:09:22.699735 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48c6b1ea7497b33c00b956e400426c1d4f3c6c66e0225133434460800eef9314\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ea6068a6b6f8123b402de354e5374a0f02a7191a922d1718ab38f8898f47c52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:22Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:22 crc kubenswrapper[4897]: I1121 14:09:22.711531 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4qwqj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f46e2ba8-9ee0-406d-a668-2ba424618b69\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f43aa89d4d9defcb50d0dc6f90293b5f426b9cfc398b780c2e10f6a5c9f1e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbsgb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4qwqj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:22Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:22 crc kubenswrapper[4897]: I1121 14:09:22.796636 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:22 crc kubenswrapper[4897]: I1121 14:09:22.796683 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:22 crc kubenswrapper[4897]: I1121 14:09:22.796695 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:22 crc kubenswrapper[4897]: I1121 14:09:22.796712 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:22 crc kubenswrapper[4897]: I1121 14:09:22.796724 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:22Z","lastTransitionTime":"2025-11-21T14:09:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:22 crc kubenswrapper[4897]: I1121 14:09:22.899135 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:22 crc kubenswrapper[4897]: I1121 14:09:22.899186 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:22 crc kubenswrapper[4897]: I1121 14:09:22.899195 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:22 crc kubenswrapper[4897]: I1121 14:09:22.899212 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:22 crc kubenswrapper[4897]: I1121 14:09:22.899224 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:22Z","lastTransitionTime":"2025-11-21T14:09:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:23 crc kubenswrapper[4897]: I1121 14:09:23.001642 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:23 crc kubenswrapper[4897]: I1121 14:09:23.001689 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:23 crc kubenswrapper[4897]: I1121 14:09:23.001701 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:23 crc kubenswrapper[4897]: I1121 14:09:23.001717 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:23 crc kubenswrapper[4897]: I1121 14:09:23.001729 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:23Z","lastTransitionTime":"2025-11-21T14:09:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:23 crc kubenswrapper[4897]: I1121 14:09:23.088581 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 21 14:09:23 crc kubenswrapper[4897]: I1121 14:09:23.088633 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 21 14:09:23 crc kubenswrapper[4897]: I1121 14:09:23.088690 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 21 14:09:23 crc kubenswrapper[4897]: E1121 14:09:23.088727 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 21 14:09:23 crc kubenswrapper[4897]: E1121 14:09:23.088775 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 21 14:09:23 crc kubenswrapper[4897]: E1121 14:09:23.088830 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 21 14:09:23 crc kubenswrapper[4897]: I1121 14:09:23.104485 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:23 crc kubenswrapper[4897]: I1121 14:09:23.104568 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:23 crc kubenswrapper[4897]: I1121 14:09:23.104581 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:23 crc kubenswrapper[4897]: I1121 14:09:23.104598 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:23 crc kubenswrapper[4897]: I1121 14:09:23.104609 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:23Z","lastTransitionTime":"2025-11-21T14:09:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:23 crc kubenswrapper[4897]: I1121 14:09:23.207642 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:23 crc kubenswrapper[4897]: I1121 14:09:23.207678 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:23 crc kubenswrapper[4897]: I1121 14:09:23.207687 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:23 crc kubenswrapper[4897]: I1121 14:09:23.207701 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:23 crc kubenswrapper[4897]: I1121 14:09:23.207710 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:23Z","lastTransitionTime":"2025-11-21T14:09:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:23 crc kubenswrapper[4897]: I1121 14:09:23.311772 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:23 crc kubenswrapper[4897]: I1121 14:09:23.311821 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:23 crc kubenswrapper[4897]: I1121 14:09:23.311832 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:23 crc kubenswrapper[4897]: I1121 14:09:23.311850 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:23 crc kubenswrapper[4897]: I1121 14:09:23.311862 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:23Z","lastTransitionTime":"2025-11-21T14:09:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:23 crc kubenswrapper[4897]: I1121 14:09:23.414258 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:23 crc kubenswrapper[4897]: I1121 14:09:23.414305 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:23 crc kubenswrapper[4897]: I1121 14:09:23.414320 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:23 crc kubenswrapper[4897]: I1121 14:09:23.414373 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:23 crc kubenswrapper[4897]: I1121 14:09:23.414294 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x5pws" event={"ID":"dc9e717e-af03-48bf-b005-62b3d084c6fc","Type":"ContainerStarted","Data":"436d6d6ffd960d562af581a8d3f5295af7eafc3c013a6e3b89ab197256dd29ca"} Nov 21 14:09:23 crc kubenswrapper[4897]: I1121 14:09:23.414387 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:23Z","lastTransitionTime":"2025-11-21T14:09:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:23 crc kubenswrapper[4897]: I1121 14:09:23.516920 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:23 crc kubenswrapper[4897]: I1121 14:09:23.516971 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:23 crc kubenswrapper[4897]: I1121 14:09:23.516983 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:23 crc kubenswrapper[4897]: I1121 14:09:23.517002 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:23 crc kubenswrapper[4897]: I1121 14:09:23.517014 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:23Z","lastTransitionTime":"2025-11-21T14:09:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:23 crc kubenswrapper[4897]: I1121 14:09:23.619393 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:23 crc kubenswrapper[4897]: I1121 14:09:23.619655 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:23 crc kubenswrapper[4897]: I1121 14:09:23.619765 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:23 crc kubenswrapper[4897]: I1121 14:09:23.619860 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:23 crc kubenswrapper[4897]: I1121 14:09:23.619948 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:23Z","lastTransitionTime":"2025-11-21T14:09:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:23 crc kubenswrapper[4897]: I1121 14:09:23.690994 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-rs4c8"] Nov 21 14:09:23 crc kubenswrapper[4897]: I1121 14:09:23.691454 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs4c8" Nov 21 14:09:23 crc kubenswrapper[4897]: E1121 14:09:23.691644 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs4c8" podUID="01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd" Nov 21 14:09:23 crc kubenswrapper[4897]: I1121 14:09:23.711851 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6cd330fd-e8a6-49d5-bfc9-5f6122ccd49e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1bd64b9b938bc80284bcee54141e90cad613d339c4f27ef2934b29bc2a318916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://936f59996d26e42099dcd055f87fff9e3ac4283b57e1956fe97463f2334152d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca74b28a069dc2e609b0779164c3aa470a15ed9848af0dd5fc3479aaf750157b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://270c53b9a342f26be077a01cd4e87e9ccd0eb5d2c84208ae38cbf720463245e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66c521c26d541bbddb064dd9710f1a7b41090d2f47343977922272dfc6dab7af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6eb2ac410b49098e5b3603de5e6bf5ed4484fab0b300021d0bed2e5ed8a42527\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6eb2ac410b49098e5b3603de5e6bf5ed4484fab0b300021d0bed2e5ed8a42527\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03385e0c78e4806f60c160227ef4905a11003ed3d70f0e5c0dd0a477979d1d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03385e0c78e4806f60c160227ef4905a11003ed3d70f0e5c0dd0a477979d1d4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://50aea2fa8b3b141b4330ddce2c8e0715230c6f06dab64b3abfd98a3067224bec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50aea2fa8b3b141b4330ddce2c8e0715230c6f06dab64b3abfd98a3067224bec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:23Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:23 crc kubenswrapper[4897]: I1121 14:09:23.722286 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:23 crc kubenswrapper[4897]: I1121 14:09:23.722326 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:23 crc kubenswrapper[4897]: I1121 14:09:23.722336 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:23 crc kubenswrapper[4897]: I1121 14:09:23.722350 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:23 crc kubenswrapper[4897]: I1121 14:09:23.722362 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:23Z","lastTransitionTime":"2025-11-21T14:09:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:23 crc kubenswrapper[4897]: I1121 14:09:23.726453 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a6c0a99-565b-41f9-a8b5-c8990a0c8eb3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://737553dd1d16e2019ea89a376e3a868f9792ed5239aa7d6cc40a05d3d4f286ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6682a8886d2b9bce069cf9261d4701ef6907f19574fd035bd89bd05626023e5c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de079956e3d8de0101f3f87dcc120ef23876c4972382639e419b3f34a38cdfca\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1907bfbad7eb34da642e30d6aa81820aab457f92347333a364ac12396b0e9dae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1907bfbad7eb34da642e30d6aa81820aab457f92347333a364ac12396b0e9dae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"le observer\\\\nW1121 14:09:07.994401 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1121 14:09:07.994601 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1121 14:09:07.995290 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1461162674/tls.crt::/tmp/serving-cert-1461162674/tls.key\\\\\\\"\\\\nI1121 14:09:08.162382 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1121 14:09:08.165301 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1121 14:09:08.165318 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1121 14:09:08.165343 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1121 14:09:08.165348 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1121 14:09:08.172015 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1121 14:09:08.172083 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1121 14:09:08.172089 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1121 14:09:08.172094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1121 14:09:08.172098 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1121 14:09:08.172102 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1121 14:09:08.172106 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1121 14:09:08.172023 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1121 14:09:08.174214 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d78c87c440c7d8e30075f6143111b3f504fb99686484b8f19a9c8d61a55eccc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://101fb592fef069af009c613a1e9c9825024ed205498ab72334211ba90d6f7397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://101fb592fef069af009c613a1e9c9825024ed205498ab72334211ba90d6f7397\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:23Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:23 crc kubenswrapper[4897]: I1121 14:09:23.738458 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b78c993a-9966-4e84-85a1-b4f73aaf17f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce3a2602338a1c1e9380ee55f78360f3d9cf6ccc62678886dc5ae574a72e3d9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c256823f2fb49ecc121f6ef564cd86196f9b4a473e9bdea670173596ef14a823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c65f71f8746044f5d41687c670a0308cd09dc41a6531f9b5720e852b3d240756\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1dbe88e7e6ad63899e107ebb884a16edb9a42b21ae7c86d7fbfaf89785b0d36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:23Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:23 crc kubenswrapper[4897]: I1121 14:09:23.749533 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09877d9780caddd5f7f73e95fdf818a427ed645ed5997984113151f4b4759c13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:23Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:23 crc kubenswrapper[4897]: I1121 14:09:23.760704 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:23Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:23 crc kubenswrapper[4897]: I1121 14:09:23.773000 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6ljgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f3b70b5-9294-4f69-9528-500d28f34c89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://527cf28abdeef1b42d9c19fb88e3171dc05c4516c3a542a461b5b1daa85c6d25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78jhb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6ljgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:23Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:23 crc kubenswrapper[4897]: I1121 14:09:23.783920 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x5pws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc9e717e-af03-48bf-b005-62b3d084c6fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9c4f5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9c4f5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-x5pws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:23Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:23 crc kubenswrapper[4897]: I1121 14:09:23.796864 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:23Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:23 crc kubenswrapper[4897]: I1121 14:09:23.809447 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c976d804a7ebeba2c879e920cbc389f96f7f5def4b4439e8c8980d5460ecdb65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:23Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:23 crc kubenswrapper[4897]: I1121 14:09:23.817716 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd-metrics-certs\") pod \"network-metrics-daemon-rs4c8\" (UID: \"01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd\") " pod="openshift-multus/network-metrics-daemon-rs4c8" Nov 21 14:09:23 crc kubenswrapper[4897]: I1121 14:09:23.817778 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7xkdf\" (UniqueName: \"kubernetes.io/projected/01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd-kube-api-access-7xkdf\") pod \"network-metrics-daemon-rs4c8\" (UID: \"01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd\") " pod="openshift-multus/network-metrics-daemon-rs4c8" Nov 21 14:09:23 crc kubenswrapper[4897]: I1121 14:09:23.822777 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nl76g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a7d6101-d0d2-4828-b3ce-d01040c04640\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac58b6ef1bd08e5f950260c18120c72e50b90a0de182c6ab0a5dbbbe45cf7328\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e56402f62821baf335f378084f4ed7c21b60ccbafe148419a0a9a403f106ad19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e56402f62821baf335f378084f4ed7c21b60ccbafe148419a0a9a403f106ad19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6022ff28e83dafcd846a7f03b090262aa3229a0da89f1984e4b6a054a04ac24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6022ff28e83dafcd846a7f03b090262aa3229a0da89f1984e4b6a054a04ac24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89de56e32b796da11bf89dd598d04fbf2884745ddf0f5d87cccfcaf1c001d8d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89de56e32b796da11bf89dd598d04fbf2884745ddf0f5d87cccfcaf1c001d8d6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2dff3cacaba7f9238f482e67b670ea775e0f96621434b99f4e76c2dfdde0321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2dff3cacaba7f9238f482e67b670ea775e0f96621434b99f4e76c2dfdde0321\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a153389a4137ee580e5118f3248400460f9d014c992a956f508815755f583afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a153389a4137ee580e5118f3248400460f9d014c992a956f508815755f583afb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a8c051ad1c1f2373f79eb0781fa6db489aeea3d52ed6c0bb43eec11520ec2b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a8c051ad1c1f2373f79eb0781fa6db489aeea3d52ed6c0bb43eec11520ec2b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nl76g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:23Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:23 crc kubenswrapper[4897]: I1121 14:09:23.824438 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:23 crc kubenswrapper[4897]: I1121 14:09:23.824486 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:23 crc kubenswrapper[4897]: I1121 14:09:23.824524 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:23 crc kubenswrapper[4897]: I1121 14:09:23.824553 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:23 crc kubenswrapper[4897]: I1121 14:09:23.824571 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:23Z","lastTransitionTime":"2025-11-21T14:09:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:23 crc kubenswrapper[4897]: I1121 14:09:23.834733 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48c6b1ea7497b33c00b956e400426c1d4f3c6c66e0225133434460800eef9314\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ea6068a6b6f8123b402de354e5374a0f02a7191a922d1718ab38f8898f47c52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:23Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:23 crc kubenswrapper[4897]: I1121 14:09:23.843257 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4qwqj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f46e2ba8-9ee0-406d-a668-2ba424618b69\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f43aa89d4d9defcb50d0dc6f90293b5f426b9cfc398b780c2e10f6a5c9f1e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbsgb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4qwqj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:23Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:23 crc kubenswrapper[4897]: I1121 14:09:23.853819 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-rs4c8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7xkdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7xkdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:23Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-rs4c8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:23Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:23 crc kubenswrapper[4897]: I1121 14:09:23.868660 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:23Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:23 crc kubenswrapper[4897]: I1121 14:09:23.878908 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7670227-d280-4847-b882-754429f56b0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://961961696b8750c248598f4cda032b18395e5c802275dcfe6d3e9f7ef3510f1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5854028821e6a360a259427b68e7bb4c2a7220e0c68f4cc91639b2ece0540a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-krv5b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:23Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:23 crc kubenswrapper[4897]: I1121 14:09:23.899484 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"222f65f2-0b82-4760-bcf7-779244720f01\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92f1d932a83d65ba6a9a906cc489ff51e76547d07d42656630c59421f833c4db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccd51aa0faf3b1327f207b355cddf86a7353c0051f7e5ce0396f3c4bc67e3b16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8971c38daad45bda3d61545f52120dad40992dbf2e1d6e1e776500ed58d73f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfee19f82e35869a641318a0c45bc7a2ae71d80039f5790e60e277be15c9d3ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec279f2f558a8ac653006321b18cbaf13185876fe303c5cc1c65988aef09edbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de095f3c760b13d5197f624eac3806a1f4186577f2b95eb522532d453eb266e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://814079abfa7d1b029825da6c6c5f1e57d439ff21ccb626c5d4e52e41d059ad3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://089c9d4e6d0fb1a4e9a65a34a033e15c12508d2dbe3b9086b464914a1dac41e5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-21T14:09:20Z\\\",\\\"message\\\":\\\"60\\\\nI1121 14:09:20.759274 6099 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1121 14:09:20.759278 6099 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1121 14:09:20.759284 6099 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1121 14:09:20.759289 6099 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1121 14:09:20.759296 6099 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1121 14:09:20.759303 6099 handler.go:208] Removed *v1.Node event handler 2\\\\nI1121 14:09:20.759302 6099 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1121 14:09:20.759313 6099 handler.go:208] Removed *v1.Node event handler 7\\\\nI1121 14:09:20.759343 6099 factory.go:656] Stopping watch factory\\\\nI1121 14:09:20.759354 6099 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1121 14:09:20.759360 6099 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1121 14:09:20.759390 6099 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1121 14:09:20.759533 6099 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1121 14:09:20.759593 6099 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:16Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b486a86e0e97ab409e590f30a1bf5cb9f584a7c67f96ff1236e42ca4f8675b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7877db6f5c2bf0eb31218c64d67bca48f921ff7ac62b4e495f29024c086bf738\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7877db6f5c2bf0eb31218c64d67bca48f921ff7ac62b4e495f29024c086bf738\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tpfb6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:23Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:23 crc kubenswrapper[4897]: I1121 14:09:23.911526 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7qfmq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"223b7002-3e67-4848-b1fb-db2921deb64e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381a46c6290da2b36e572f46c16f1a50aa3e25359bcfbc02c7a5a18e1484860e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2m27\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7qfmq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:23Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:23 crc kubenswrapper[4897]: I1121 14:09:23.919092 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd-metrics-certs\") pod \"network-metrics-daemon-rs4c8\" (UID: \"01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd\") " pod="openshift-multus/network-metrics-daemon-rs4c8" Nov 21 14:09:23 crc kubenswrapper[4897]: I1121 14:09:23.919137 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7xkdf\" (UniqueName: \"kubernetes.io/projected/01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd-kube-api-access-7xkdf\") pod \"network-metrics-daemon-rs4c8\" (UID: \"01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd\") " pod="openshift-multus/network-metrics-daemon-rs4c8" Nov 21 14:09:23 crc kubenswrapper[4897]: E1121 14:09:23.919305 4897 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 21 14:09:23 crc kubenswrapper[4897]: E1121 14:09:23.919428 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd-metrics-certs podName:01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd nodeName:}" failed. No retries permitted until 2025-11-21 14:09:24.419396164 +0000 UTC m=+41.703989709 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd-metrics-certs") pod "network-metrics-daemon-rs4c8" (UID: "01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 21 14:09:23 crc kubenswrapper[4897]: I1121 14:09:23.926963 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:23 crc kubenswrapper[4897]: I1121 14:09:23.927005 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:23 crc kubenswrapper[4897]: I1121 14:09:23.927014 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:23 crc kubenswrapper[4897]: I1121 14:09:23.927031 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:23 crc kubenswrapper[4897]: I1121 14:09:23.927045 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:23Z","lastTransitionTime":"2025-11-21T14:09:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:23 crc kubenswrapper[4897]: I1121 14:09:23.938897 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7xkdf\" (UniqueName: \"kubernetes.io/projected/01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd-kube-api-access-7xkdf\") pod \"network-metrics-daemon-rs4c8\" (UID: \"01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd\") " pod="openshift-multus/network-metrics-daemon-rs4c8" Nov 21 14:09:24 crc kubenswrapper[4897]: I1121 14:09:24.029687 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:24 crc kubenswrapper[4897]: I1121 14:09:24.029728 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:24 crc kubenswrapper[4897]: I1121 14:09:24.029736 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:24 crc kubenswrapper[4897]: I1121 14:09:24.029751 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:24 crc kubenswrapper[4897]: I1121 14:09:24.029761 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:24Z","lastTransitionTime":"2025-11-21T14:09:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:24 crc kubenswrapper[4897]: I1121 14:09:24.103928 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48c6b1ea7497b33c00b956e400426c1d4f3c6c66e0225133434460800eef9314\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ea6068a6b6f8123b402de354e5374a0f02a7191a922d1718ab38f8898f47c52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:24Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:24 crc kubenswrapper[4897]: I1121 14:09:24.118155 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4qwqj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f46e2ba8-9ee0-406d-a668-2ba424618b69\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f43aa89d4d9defcb50d0dc6f90293b5f426b9cfc398b780c2e10f6a5c9f1e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbsgb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4qwqj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:24Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:24 crc kubenswrapper[4897]: I1121 14:09:24.132368 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:24 crc kubenswrapper[4897]: I1121 14:09:24.132424 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:24 crc kubenswrapper[4897]: I1121 14:09:24.132436 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:24 crc kubenswrapper[4897]: I1121 14:09:24.132458 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:24 crc kubenswrapper[4897]: I1121 14:09:24.132469 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:24Z","lastTransitionTime":"2025-11-21T14:09:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:24 crc kubenswrapper[4897]: I1121 14:09:24.134837 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-rs4c8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7xkdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7xkdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:23Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-rs4c8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:24Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:24 crc kubenswrapper[4897]: I1121 14:09:24.150367 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:24Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:24 crc kubenswrapper[4897]: I1121 14:09:24.166102 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7670227-d280-4847-b882-754429f56b0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://961961696b8750c248598f4cda032b18395e5c802275dcfe6d3e9f7ef3510f1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5854028821e6a360a259427b68e7bb4c2a7220e0c68f4cc91639b2ece0540a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-krv5b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:24Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:24 crc kubenswrapper[4897]: I1121 14:09:24.186054 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"222f65f2-0b82-4760-bcf7-779244720f01\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92f1d932a83d65ba6a9a906cc489ff51e76547d07d42656630c59421f833c4db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccd51aa0faf3b1327f207b355cddf86a7353c0051f7e5ce0396f3c4bc67e3b16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8971c38daad45bda3d61545f52120dad40992dbf2e1d6e1e776500ed58d73f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfee19f82e35869a641318a0c45bc7a2ae71d80039f5790e60e277be15c9d3ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec279f2f558a8ac653006321b18cbaf13185876fe303c5cc1c65988aef09edbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de095f3c760b13d5197f624eac3806a1f4186577f2b95eb522532d453eb266e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://814079abfa7d1b029825da6c6c5f1e57d439ff21ccb626c5d4e52e41d059ad3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://089c9d4e6d0fb1a4e9a65a34a033e15c12508d2dbe3b9086b464914a1dac41e5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-21T14:09:20Z\\\",\\\"message\\\":\\\"60\\\\nI1121 14:09:20.759274 6099 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1121 14:09:20.759278 6099 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1121 14:09:20.759284 6099 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1121 14:09:20.759289 6099 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1121 14:09:20.759296 6099 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1121 14:09:20.759303 6099 handler.go:208] Removed *v1.Node event handler 2\\\\nI1121 14:09:20.759302 6099 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1121 14:09:20.759313 6099 handler.go:208] Removed *v1.Node event handler 7\\\\nI1121 14:09:20.759343 6099 factory.go:656] Stopping watch factory\\\\nI1121 14:09:20.759354 6099 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1121 14:09:20.759360 6099 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1121 14:09:20.759390 6099 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1121 14:09:20.759533 6099 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1121 14:09:20.759593 6099 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:16Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b486a86e0e97ab409e590f30a1bf5cb9f584a7c67f96ff1236e42ca4f8675b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7877db6f5c2bf0eb31218c64d67bca48f921ff7ac62b4e495f29024c086bf738\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7877db6f5c2bf0eb31218c64d67bca48f921ff7ac62b4e495f29024c086bf738\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tpfb6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:24Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:24 crc kubenswrapper[4897]: I1121 14:09:24.197717 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7qfmq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"223b7002-3e67-4848-b1fb-db2921deb64e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381a46c6290da2b36e572f46c16f1a50aa3e25359bcfbc02c7a5a18e1484860e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2m27\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7qfmq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:24Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:24 crc kubenswrapper[4897]: I1121 14:09:24.219468 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6cd330fd-e8a6-49d5-bfc9-5f6122ccd49e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1bd64b9b938bc80284bcee54141e90cad613d339c4f27ef2934b29bc2a318916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://936f59996d26e42099dcd055f87fff9e3ac4283b57e1956fe97463f2334152d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca74b28a069dc2e609b0779164c3aa470a15ed9848af0dd5fc3479aaf750157b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://270c53b9a342f26be077a01cd4e87e9ccd0eb5d2c84208ae38cbf720463245e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66c521c26d541bbddb064dd9710f1a7b41090d2f47343977922272dfc6dab7af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6eb2ac410b49098e5b3603de5e6bf5ed4484fab0b300021d0bed2e5ed8a42527\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6eb2ac410b49098e5b3603de5e6bf5ed4484fab0b300021d0bed2e5ed8a42527\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03385e0c78e4806f60c160227ef4905a11003ed3d70f0e5c0dd0a477979d1d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03385e0c78e4806f60c160227ef4905a11003ed3d70f0e5c0dd0a477979d1d4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://50aea2fa8b3b141b4330ddce2c8e0715230c6f06dab64b3abfd98a3067224bec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50aea2fa8b3b141b4330ddce2c8e0715230c6f06dab64b3abfd98a3067224bec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:24Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:24 crc kubenswrapper[4897]: I1121 14:09:24.233960 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:24 crc kubenswrapper[4897]: I1121 14:09:24.234001 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:24 crc kubenswrapper[4897]: I1121 14:09:24.234014 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:24 crc kubenswrapper[4897]: I1121 14:09:24.234033 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:24 crc kubenswrapper[4897]: I1121 14:09:24.234047 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:24Z","lastTransitionTime":"2025-11-21T14:09:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:24 crc kubenswrapper[4897]: I1121 14:09:24.236247 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a6c0a99-565b-41f9-a8b5-c8990a0c8eb3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://737553dd1d16e2019ea89a376e3a868f9792ed5239aa7d6cc40a05d3d4f286ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6682a8886d2b9bce069cf9261d4701ef6907f19574fd035bd89bd05626023e5c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de079956e3d8de0101f3f87dcc120ef23876c4972382639e419b3f34a38cdfca\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1907bfbad7eb34da642e30d6aa81820aab457f92347333a364ac12396b0e9dae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1907bfbad7eb34da642e30d6aa81820aab457f92347333a364ac12396b0e9dae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"le observer\\\\nW1121 14:09:07.994401 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1121 14:09:07.994601 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1121 14:09:07.995290 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1461162674/tls.crt::/tmp/serving-cert-1461162674/tls.key\\\\\\\"\\\\nI1121 14:09:08.162382 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1121 14:09:08.165301 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1121 14:09:08.165318 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1121 14:09:08.165343 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1121 14:09:08.165348 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1121 14:09:08.172015 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1121 14:09:08.172083 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1121 14:09:08.172089 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1121 14:09:08.172094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1121 14:09:08.172098 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1121 14:09:08.172102 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1121 14:09:08.172106 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1121 14:09:08.172023 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1121 14:09:08.174214 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d78c87c440c7d8e30075f6143111b3f504fb99686484b8f19a9c8d61a55eccc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://101fb592fef069af009c613a1e9c9825024ed205498ab72334211ba90d6f7397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://101fb592fef069af009c613a1e9c9825024ed205498ab72334211ba90d6f7397\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:24Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:24 crc kubenswrapper[4897]: I1121 14:09:24.247979 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b78c993a-9966-4e84-85a1-b4f73aaf17f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce3a2602338a1c1e9380ee55f78360f3d9cf6ccc62678886dc5ae574a72e3d9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c256823f2fb49ecc121f6ef564cd86196f9b4a473e9bdea670173596ef14a823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c65f71f8746044f5d41687c670a0308cd09dc41a6531f9b5720e852b3d240756\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1dbe88e7e6ad63899e107ebb884a16edb9a42b21ae7c86d7fbfaf89785b0d36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:24Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:24 crc kubenswrapper[4897]: I1121 14:09:24.258526 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09877d9780caddd5f7f73e95fdf818a427ed645ed5997984113151f4b4759c13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:24Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:24 crc kubenswrapper[4897]: I1121 14:09:24.271787 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:24Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:24 crc kubenswrapper[4897]: I1121 14:09:24.289260 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6ljgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f3b70b5-9294-4f69-9528-500d28f34c89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://527cf28abdeef1b42d9c19fb88e3171dc05c4516c3a542a461b5b1daa85c6d25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78jhb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6ljgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:24Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:24 crc kubenswrapper[4897]: I1121 14:09:24.303163 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x5pws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc9e717e-af03-48bf-b005-62b3d084c6fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9c4f5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9c4f5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-x5pws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:24Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:24 crc kubenswrapper[4897]: I1121 14:09:24.317653 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:24Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:24 crc kubenswrapper[4897]: I1121 14:09:24.336863 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:24 crc kubenswrapper[4897]: I1121 14:09:24.336906 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:24 crc kubenswrapper[4897]: I1121 14:09:24.336916 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:24 crc kubenswrapper[4897]: I1121 14:09:24.336936 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:24 crc kubenswrapper[4897]: I1121 14:09:24.336947 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:24Z","lastTransitionTime":"2025-11-21T14:09:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:24 crc kubenswrapper[4897]: I1121 14:09:24.337760 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c976d804a7ebeba2c879e920cbc389f96f7f5def4b4439e8c8980d5460ecdb65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:24Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:24 crc kubenswrapper[4897]: I1121 14:09:24.353486 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nl76g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a7d6101-d0d2-4828-b3ce-d01040c04640\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac58b6ef1bd08e5f950260c18120c72e50b90a0de182c6ab0a5dbbbe45cf7328\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e56402f62821baf335f378084f4ed7c21b60ccbafe148419a0a9a403f106ad19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e56402f62821baf335f378084f4ed7c21b60ccbafe148419a0a9a403f106ad19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6022ff28e83dafcd846a7f03b090262aa3229a0da89f1984e4b6a054a04ac24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6022ff28e83dafcd846a7f03b090262aa3229a0da89f1984e4b6a054a04ac24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89de56e32b796da11bf89dd598d04fbf2884745ddf0f5d87cccfcaf1c001d8d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89de56e32b796da11bf89dd598d04fbf2884745ddf0f5d87cccfcaf1c001d8d6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2dff3cacaba7f9238f482e67b670ea775e0f96621434b99f4e76c2dfdde0321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2dff3cacaba7f9238f482e67b670ea775e0f96621434b99f4e76c2dfdde0321\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a153389a4137ee580e5118f3248400460f9d014c992a956f508815755f583afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a153389a4137ee580e5118f3248400460f9d014c992a956f508815755f583afb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a8c051ad1c1f2373f79eb0781fa6db489aeea3d52ed6c0bb43eec11520ec2b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a8c051ad1c1f2373f79eb0781fa6db489aeea3d52ed6c0bb43eec11520ec2b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nl76g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:24Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:24 crc kubenswrapper[4897]: I1121 14:09:24.419714 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x5pws" event={"ID":"dc9e717e-af03-48bf-b005-62b3d084c6fc","Type":"ContainerStarted","Data":"b0f99b9213902c003e6ca754134d2482404ace39c184fb3b431d2b600481d5b4"} Nov 21 14:09:24 crc kubenswrapper[4897]: I1121 14:09:24.419778 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x5pws" event={"ID":"dc9e717e-af03-48bf-b005-62b3d084c6fc","Type":"ContainerStarted","Data":"f968eee1d8a1888229e23abda59fd58fa5e86f02ca96ff80fbe093904b27e8df"} Nov 21 14:09:24 crc kubenswrapper[4897]: I1121 14:09:24.421703 4897 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tpfb6_222f65f2-0b82-4760-bcf7-779244720f01/ovnkube-controller/1.log" Nov 21 14:09:24 crc kubenswrapper[4897]: I1121 14:09:24.422194 4897 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tpfb6_222f65f2-0b82-4760-bcf7-779244720f01/ovnkube-controller/0.log" Nov 21 14:09:24 crc kubenswrapper[4897]: I1121 14:09:24.423567 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd-metrics-certs\") pod \"network-metrics-daemon-rs4c8\" (UID: \"01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd\") " pod="openshift-multus/network-metrics-daemon-rs4c8" Nov 21 14:09:24 crc kubenswrapper[4897]: E1121 14:09:24.423694 4897 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 21 14:09:24 crc kubenswrapper[4897]: E1121 14:09:24.423782 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd-metrics-certs podName:01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd nodeName:}" failed. No retries permitted until 2025-11-21 14:09:25.423755543 +0000 UTC m=+42.708349098 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd-metrics-certs") pod "network-metrics-daemon-rs4c8" (UID: "01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 21 14:09:24 crc kubenswrapper[4897]: I1121 14:09:24.429200 4897 generic.go:334] "Generic (PLEG): container finished" podID="222f65f2-0b82-4760-bcf7-779244720f01" containerID="814079abfa7d1b029825da6c6c5f1e57d439ff21ccb626c5d4e52e41d059ad3a" exitCode=1 Nov 21 14:09:24 crc kubenswrapper[4897]: I1121 14:09:24.429278 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" event={"ID":"222f65f2-0b82-4760-bcf7-779244720f01","Type":"ContainerDied","Data":"814079abfa7d1b029825da6c6c5f1e57d439ff21ccb626c5d4e52e41d059ad3a"} Nov 21 14:09:24 crc kubenswrapper[4897]: I1121 14:09:24.429439 4897 scope.go:117] "RemoveContainer" containerID="089c9d4e6d0fb1a4e9a65a34a033e15c12508d2dbe3b9086b464914a1dac41e5" Nov 21 14:09:24 crc kubenswrapper[4897]: I1121 14:09:24.430243 4897 scope.go:117] "RemoveContainer" containerID="814079abfa7d1b029825da6c6c5f1e57d439ff21ccb626c5d4e52e41d059ad3a" Nov 21 14:09:24 crc kubenswrapper[4897]: E1121 14:09:24.430425 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-tpfb6_openshift-ovn-kubernetes(222f65f2-0b82-4760-bcf7-779244720f01)\"" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" podUID="222f65f2-0b82-4760-bcf7-779244720f01" Nov 21 14:09:24 crc kubenswrapper[4897]: I1121 14:09:24.439997 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:24 crc kubenswrapper[4897]: I1121 14:09:24.440041 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:24 crc kubenswrapper[4897]: I1121 14:09:24.440054 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:24 crc kubenswrapper[4897]: I1121 14:09:24.440070 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:24 crc kubenswrapper[4897]: I1121 14:09:24.440080 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:24Z","lastTransitionTime":"2025-11-21T14:09:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:24 crc kubenswrapper[4897]: I1121 14:09:24.448714 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b78c993a-9966-4e84-85a1-b4f73aaf17f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce3a2602338a1c1e9380ee55f78360f3d9cf6ccc62678886dc5ae574a72e3d9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c256823f2fb49ecc121f6ef564cd86196f9b4a473e9bdea670173596ef14a823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c65f71f8746044f5d41687c670a0308cd09dc41a6531f9b5720e852b3d240756\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1dbe88e7e6ad63899e107ebb884a16edb9a42b21ae7c86d7fbfaf89785b0d36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:24Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:24 crc kubenswrapper[4897]: I1121 14:09:24.461385 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09877d9780caddd5f7f73e95fdf818a427ed645ed5997984113151f4b4759c13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:24Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:24 crc kubenswrapper[4897]: I1121 14:09:24.475409 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:24Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:24 crc kubenswrapper[4897]: I1121 14:09:24.493294 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6ljgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f3b70b5-9294-4f69-9528-500d28f34c89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://527cf28abdeef1b42d9c19fb88e3171dc05c4516c3a542a461b5b1daa85c6d25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78jhb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6ljgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:24Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:24 crc kubenswrapper[4897]: I1121 14:09:24.506616 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x5pws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc9e717e-af03-48bf-b005-62b3d084c6fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9c4f5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9c4f5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-x5pws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:24Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:24 crc kubenswrapper[4897]: I1121 14:09:24.529969 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6cd330fd-e8a6-49d5-bfc9-5f6122ccd49e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1bd64b9b938bc80284bcee54141e90cad613d339c4f27ef2934b29bc2a318916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://936f59996d26e42099dcd055f87fff9e3ac4283b57e1956fe97463f2334152d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca74b28a069dc2e609b0779164c3aa470a15ed9848af0dd5fc3479aaf750157b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://270c53b9a342f26be077a01cd4e87e9ccd0eb5d2c84208ae38cbf720463245e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66c521c26d541bbddb064dd9710f1a7b41090d2f47343977922272dfc6dab7af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6eb2ac410b49098e5b3603de5e6bf5ed4484fab0b300021d0bed2e5ed8a42527\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6eb2ac410b49098e5b3603de5e6bf5ed4484fab0b300021d0bed2e5ed8a42527\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03385e0c78e4806f60c160227ef4905a11003ed3d70f0e5c0dd0a477979d1d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03385e0c78e4806f60c160227ef4905a11003ed3d70f0e5c0dd0a477979d1d4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://50aea2fa8b3b141b4330ddce2c8e0715230c6f06dab64b3abfd98a3067224bec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50aea2fa8b3b141b4330ddce2c8e0715230c6f06dab64b3abfd98a3067224bec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:24Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:24 crc kubenswrapper[4897]: I1121 14:09:24.542560 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:24 crc kubenswrapper[4897]: I1121 14:09:24.542595 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:24 crc kubenswrapper[4897]: I1121 14:09:24.542604 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:24 crc kubenswrapper[4897]: I1121 14:09:24.542621 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:24 crc kubenswrapper[4897]: I1121 14:09:24.542632 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:24Z","lastTransitionTime":"2025-11-21T14:09:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:24 crc kubenswrapper[4897]: I1121 14:09:24.546469 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a6c0a99-565b-41f9-a8b5-c8990a0c8eb3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://737553dd1d16e2019ea89a376e3a868f9792ed5239aa7d6cc40a05d3d4f286ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6682a8886d2b9bce069cf9261d4701ef6907f19574fd035bd89bd05626023e5c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de079956e3d8de0101f3f87dcc120ef23876c4972382639e419b3f34a38cdfca\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1907bfbad7eb34da642e30d6aa81820aab457f92347333a364ac12396b0e9dae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1907bfbad7eb34da642e30d6aa81820aab457f92347333a364ac12396b0e9dae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"le observer\\\\nW1121 14:09:07.994401 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1121 14:09:07.994601 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1121 14:09:07.995290 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1461162674/tls.crt::/tmp/serving-cert-1461162674/tls.key\\\\\\\"\\\\nI1121 14:09:08.162382 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1121 14:09:08.165301 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1121 14:09:08.165318 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1121 14:09:08.165343 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1121 14:09:08.165348 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1121 14:09:08.172015 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1121 14:09:08.172083 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1121 14:09:08.172089 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1121 14:09:08.172094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1121 14:09:08.172098 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1121 14:09:08.172102 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1121 14:09:08.172106 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1121 14:09:08.172023 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1121 14:09:08.174214 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d78c87c440c7d8e30075f6143111b3f504fb99686484b8f19a9c8d61a55eccc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://101fb592fef069af009c613a1e9c9825024ed205498ab72334211ba90d6f7397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://101fb592fef069af009c613a1e9c9825024ed205498ab72334211ba90d6f7397\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:24Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:24 crc kubenswrapper[4897]: I1121 14:09:24.568868 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nl76g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a7d6101-d0d2-4828-b3ce-d01040c04640\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac58b6ef1bd08e5f950260c18120c72e50b90a0de182c6ab0a5dbbbe45cf7328\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e56402f62821baf335f378084f4ed7c21b60ccbafe148419a0a9a403f106ad19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e56402f62821baf335f378084f4ed7c21b60ccbafe148419a0a9a403f106ad19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6022ff28e83dafcd846a7f03b090262aa3229a0da89f1984e4b6a054a04ac24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6022ff28e83dafcd846a7f03b090262aa3229a0da89f1984e4b6a054a04ac24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89de56e32b796da11bf89dd598d04fbf2884745ddf0f5d87cccfcaf1c001d8d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89de56e32b796da11bf89dd598d04fbf2884745ddf0f5d87cccfcaf1c001d8d6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2dff3cacaba7f9238f482e67b670ea775e0f96621434b99f4e76c2dfdde0321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2dff3cacaba7f9238f482e67b670ea775e0f96621434b99f4e76c2dfdde0321\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a153389a4137ee580e5118f3248400460f9d014c992a956f508815755f583afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a153389a4137ee580e5118f3248400460f9d014c992a956f508815755f583afb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a8c051ad1c1f2373f79eb0781fa6db489aeea3d52ed6c0bb43eec11520ec2b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a8c051ad1c1f2373f79eb0781fa6db489aeea3d52ed6c0bb43eec11520ec2b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nl76g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:24Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:24 crc kubenswrapper[4897]: I1121 14:09:24.581651 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:24Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:24 crc kubenswrapper[4897]: I1121 14:09:24.593742 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c976d804a7ebeba2c879e920cbc389f96f7f5def4b4439e8c8980d5460ecdb65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:24Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:24 crc kubenswrapper[4897]: I1121 14:09:24.603621 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4qwqj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f46e2ba8-9ee0-406d-a668-2ba424618b69\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f43aa89d4d9defcb50d0dc6f90293b5f426b9cfc398b780c2e10f6a5c9f1e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbsgb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4qwqj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:24Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:24 crc kubenswrapper[4897]: I1121 14:09:24.613204 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-rs4c8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7xkdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7xkdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:23Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-rs4c8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:24Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:24 crc kubenswrapper[4897]: I1121 14:09:24.623556 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48c6b1ea7497b33c00b956e400426c1d4f3c6c66e0225133434460800eef9314\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ea6068a6b6f8123b402de354e5374a0f02a7191a922d1718ab38f8898f47c52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:24Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:24 crc kubenswrapper[4897]: I1121 14:09:24.634162 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7670227-d280-4847-b882-754429f56b0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://961961696b8750c248598f4cda032b18395e5c802275dcfe6d3e9f7ef3510f1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5854028821e6a360a259427b68e7bb4c2a7220e0c68f4cc91639b2ece0540a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-krv5b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:24Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:24 crc kubenswrapper[4897]: I1121 14:09:24.645052 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:24 crc kubenswrapper[4897]: I1121 14:09:24.645091 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:24 crc kubenswrapper[4897]: I1121 14:09:24.645101 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:24 crc kubenswrapper[4897]: I1121 14:09:24.645115 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:24 crc kubenswrapper[4897]: I1121 14:09:24.645125 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:24Z","lastTransitionTime":"2025-11-21T14:09:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:24 crc kubenswrapper[4897]: I1121 14:09:24.651348 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"222f65f2-0b82-4760-bcf7-779244720f01\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92f1d932a83d65ba6a9a906cc489ff51e76547d07d42656630c59421f833c4db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccd51aa0faf3b1327f207b355cddf86a7353c0051f7e5ce0396f3c4bc67e3b16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8971c38daad45bda3d61545f52120dad40992dbf2e1d6e1e776500ed58d73f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfee19f82e35869a641318a0c45bc7a2ae71d80039f5790e60e277be15c9d3ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec279f2f558a8ac653006321b18cbaf13185876fe303c5cc1c65988aef09edbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de095f3c760b13d5197f624eac3806a1f4186577f2b95eb522532d453eb266e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://814079abfa7d1b029825da6c6c5f1e57d439ff21ccb626c5d4e52e41d059ad3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://089c9d4e6d0fb1a4e9a65a34a033e15c12508d2dbe3b9086b464914a1dac41e5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-21T14:09:20Z\\\",\\\"message\\\":\\\"60\\\\nI1121 14:09:20.759274 6099 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1121 14:09:20.759278 6099 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1121 14:09:20.759284 6099 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1121 14:09:20.759289 6099 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1121 14:09:20.759296 6099 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1121 14:09:20.759303 6099 handler.go:208] Removed *v1.Node event handler 2\\\\nI1121 14:09:20.759302 6099 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1121 14:09:20.759313 6099 handler.go:208] Removed *v1.Node event handler 7\\\\nI1121 14:09:20.759343 6099 factory.go:656] Stopping watch factory\\\\nI1121 14:09:20.759354 6099 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1121 14:09:20.759360 6099 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1121 14:09:20.759390 6099 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1121 14:09:20.759533 6099 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1121 14:09:20.759593 6099 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:16Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://814079abfa7d1b029825da6c6c5f1e57d439ff21ccb626c5d4e52e41d059ad3a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-21T14:09:23Z\\\",\\\"message\\\":\\\"luster\\\\\\\", UUID:\\\\\\\"63b1440a-0908-4cab-8799-012fa1cf0b07\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-operator-lifecycle-manager/olm-operator-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-operator-lifecycle-manager/olm-operator-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-operator-lifecycle-manager/olm-operator-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.168\\\\\\\", Port:8443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1121 14:09:22.444620 6334 ovnkube.go:599] Stopped ovnkube\\\\nI1121 14:09:22.444715 6334 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1121 14:09:22.444771 6334 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b486a86e0e97ab409e590f30a1bf5cb9f584a7c67f96ff1236e42ca4f8675b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7877db6f5c2bf0eb31218c64d67bca48f921ff7ac62b4e495f29024c086bf738\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7877db6f5c2bf0eb31218c64d67bca48f921ff7ac62b4e495f29024c086bf738\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tpfb6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:24Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:24 crc kubenswrapper[4897]: I1121 14:09:24.660473 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7qfmq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"223b7002-3e67-4848-b1fb-db2921deb64e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381a46c6290da2b36e572f46c16f1a50aa3e25359bcfbc02c7a5a18e1484860e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2m27\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7qfmq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:24Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:24 crc kubenswrapper[4897]: I1121 14:09:24.671245 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:24Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:24 crc kubenswrapper[4897]: I1121 14:09:24.747980 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:24 crc kubenswrapper[4897]: I1121 14:09:24.748032 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:24 crc kubenswrapper[4897]: I1121 14:09:24.748043 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:24 crc kubenswrapper[4897]: I1121 14:09:24.748059 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:24 crc kubenswrapper[4897]: I1121 14:09:24.748084 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:24Z","lastTransitionTime":"2025-11-21T14:09:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:24 crc kubenswrapper[4897]: I1121 14:09:24.827869 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 21 14:09:24 crc kubenswrapper[4897]: I1121 14:09:24.828116 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 21 14:09:24 crc kubenswrapper[4897]: I1121 14:09:24.828152 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 21 14:09:24 crc kubenswrapper[4897]: E1121 14:09:24.828253 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-21 14:09:40.828181859 +0000 UTC m=+58.112775344 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:09:24 crc kubenswrapper[4897]: E1121 14:09:24.828270 4897 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 21 14:09:24 crc kubenswrapper[4897]: E1121 14:09:24.828377 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-21 14:09:40.828367754 +0000 UTC m=+58.112961229 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 21 14:09:24 crc kubenswrapper[4897]: E1121 14:09:24.828270 4897 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 21 14:09:24 crc kubenswrapper[4897]: E1121 14:09:24.828492 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-21 14:09:40.828469707 +0000 UTC m=+58.113063292 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 21 14:09:24 crc kubenswrapper[4897]: I1121 14:09:24.851619 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:24 crc kubenswrapper[4897]: I1121 14:09:24.851681 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:24 crc kubenswrapper[4897]: I1121 14:09:24.851692 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:24 crc kubenswrapper[4897]: I1121 14:09:24.851713 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:24 crc kubenswrapper[4897]: I1121 14:09:24.851726 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:24Z","lastTransitionTime":"2025-11-21T14:09:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:24 crc kubenswrapper[4897]: I1121 14:09:24.929452 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 21 14:09:24 crc kubenswrapper[4897]: I1121 14:09:24.929556 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 21 14:09:24 crc kubenswrapper[4897]: E1121 14:09:24.929675 4897 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 21 14:09:24 crc kubenswrapper[4897]: E1121 14:09:24.929694 4897 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 21 14:09:24 crc kubenswrapper[4897]: E1121 14:09:24.929705 4897 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 21 14:09:24 crc kubenswrapper[4897]: E1121 14:09:24.929751 4897 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 21 14:09:24 crc kubenswrapper[4897]: E1121 14:09:24.929810 4897 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 21 14:09:24 crc kubenswrapper[4897]: E1121 14:09:24.929828 4897 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 21 14:09:24 crc kubenswrapper[4897]: E1121 14:09:24.929759 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-21 14:09:40.929744086 +0000 UTC m=+58.214337561 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 21 14:09:24 crc kubenswrapper[4897]: E1121 14:09:24.929923 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-21 14:09:40.929897459 +0000 UTC m=+58.214490974 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 21 14:09:24 crc kubenswrapper[4897]: I1121 14:09:24.954173 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:24 crc kubenswrapper[4897]: I1121 14:09:24.954237 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:24 crc kubenswrapper[4897]: I1121 14:09:24.954254 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:24 crc kubenswrapper[4897]: I1121 14:09:24.954274 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:24 crc kubenswrapper[4897]: I1121 14:09:24.954285 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:24Z","lastTransitionTime":"2025-11-21T14:09:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.057056 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.057100 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.057110 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.057124 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.057133 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:25Z","lastTransitionTime":"2025-11-21T14:09:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.088528 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs4c8" Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.088570 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.088636 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 21 14:09:25 crc kubenswrapper[4897]: E1121 14:09:25.088690 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs4c8" podUID="01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd" Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.088716 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 21 14:09:25 crc kubenswrapper[4897]: E1121 14:09:25.088794 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 21 14:09:25 crc kubenswrapper[4897]: E1121 14:09:25.088924 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 21 14:09:25 crc kubenswrapper[4897]: E1121 14:09:25.089162 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.089479 4897 scope.go:117] "RemoveContainer" containerID="1907bfbad7eb34da642e30d6aa81820aab457f92347333a364ac12396b0e9dae" Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.159335 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.159371 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.159380 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.159395 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.159406 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:25Z","lastTransitionTime":"2025-11-21T14:09:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.262206 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.262278 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.262288 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.262306 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.262318 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:25Z","lastTransitionTime":"2025-11-21T14:09:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.364843 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.364881 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.364890 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.364904 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.364914 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:25Z","lastTransitionTime":"2025-11-21T14:09:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.408756 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.408798 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.408811 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.408827 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.408838 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:25Z","lastTransitionTime":"2025-11-21T14:09:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:25 crc kubenswrapper[4897]: E1121 14:09:25.428053 4897 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8b8bdf9e-0fef-41ac-98ad-710dc2137183\\\",\\\"systemUUID\\\":\\\"8a210af6-a71f-4ea9-90a3-aac32fb868b7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:25Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.432310 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.432350 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.432381 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.432400 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.432412 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:25Z","lastTransitionTime":"2025-11-21T14:09:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.434684 4897 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.434784 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd-metrics-certs\") pod \"network-metrics-daemon-rs4c8\" (UID: \"01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd\") " pod="openshift-multus/network-metrics-daemon-rs4c8" Nov 21 14:09:25 crc kubenswrapper[4897]: E1121 14:09:25.434982 4897 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 21 14:09:25 crc kubenswrapper[4897]: E1121 14:09:25.435053 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd-metrics-certs podName:01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd nodeName:}" failed. No retries permitted until 2025-11-21 14:09:27.435029369 +0000 UTC m=+44.719622904 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd-metrics-certs") pod "network-metrics-daemon-rs4c8" (UID: "01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.436556 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"469b5be7467d4a58d3776ec0a967ed911be31b770b49965da4f5d2fb264d8a3d"} Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.437180 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.439869 4897 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tpfb6_222f65f2-0b82-4760-bcf7-779244720f01/ovnkube-controller/1.log" Nov 21 14:09:25 crc kubenswrapper[4897]: E1121 14:09:25.444348 4897 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8b8bdf9e-0fef-41ac-98ad-710dc2137183\\\",\\\"systemUUID\\\":\\\"8a210af6-a71f-4ea9-90a3-aac32fb868b7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:25Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.447721 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.447768 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.447782 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.447797 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.447809 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:25Z","lastTransitionTime":"2025-11-21T14:09:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.455173 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48c6b1ea7497b33c00b956e400426c1d4f3c6c66e0225133434460800eef9314\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ea6068a6b6f8123b402de354e5374a0f02a7191a922d1718ab38f8898f47c52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:25Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:25 crc kubenswrapper[4897]: E1121 14:09:25.462439 4897 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8b8bdf9e-0fef-41ac-98ad-710dc2137183\\\",\\\"systemUUID\\\":\\\"8a210af6-a71f-4ea9-90a3-aac32fb868b7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:25Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.466442 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.466486 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.466497 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.466538 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.466551 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:25Z","lastTransitionTime":"2025-11-21T14:09:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.468424 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4qwqj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f46e2ba8-9ee0-406d-a668-2ba424618b69\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f43aa89d4d9defcb50d0dc6f90293b5f426b9cfc398b780c2e10f6a5c9f1e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbsgb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4qwqj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:25Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:25 crc kubenswrapper[4897]: E1121 14:09:25.478408 4897 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8b8bdf9e-0fef-41ac-98ad-710dc2137183\\\",\\\"systemUUID\\\":\\\"8a210af6-a71f-4ea9-90a3-aac32fb868b7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:25Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.478752 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-rs4c8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7xkdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7xkdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:23Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-rs4c8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:25Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.482671 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.482705 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.482725 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.482747 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.482760 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:25Z","lastTransitionTime":"2025-11-21T14:09:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.494438 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:25Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:25 crc kubenswrapper[4897]: E1121 14:09:25.496622 4897 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8b8bdf9e-0fef-41ac-98ad-710dc2137183\\\",\\\"systemUUID\\\":\\\"8a210af6-a71f-4ea9-90a3-aac32fb868b7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:25Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:25 crc kubenswrapper[4897]: E1121 14:09:25.496750 4897 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.498986 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.499027 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.499044 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.499064 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.499078 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:25Z","lastTransitionTime":"2025-11-21T14:09:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.507337 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7670227-d280-4847-b882-754429f56b0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://961961696b8750c248598f4cda032b18395e5c802275dcfe6d3e9f7ef3510f1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5854028821e6a360a259427b68e7bb4c2a7220e0c68f4cc91639b2ece0540a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-krv5b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:25Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.526822 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"222f65f2-0b82-4760-bcf7-779244720f01\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92f1d932a83d65ba6a9a906cc489ff51e76547d07d42656630c59421f833c4db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccd51aa0faf3b1327f207b355cddf86a7353c0051f7e5ce0396f3c4bc67e3b16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8971c38daad45bda3d61545f52120dad40992dbf2e1d6e1e776500ed58d73f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfee19f82e35869a641318a0c45bc7a2ae71d80039f5790e60e277be15c9d3ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec279f2f558a8ac653006321b18cbaf13185876fe303c5cc1c65988aef09edbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de095f3c760b13d5197f624eac3806a1f4186577f2b95eb522532d453eb266e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://814079abfa7d1b029825da6c6c5f1e57d439ff21ccb626c5d4e52e41d059ad3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://089c9d4e6d0fb1a4e9a65a34a033e15c12508d2dbe3b9086b464914a1dac41e5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-21T14:09:20Z\\\",\\\"message\\\":\\\"60\\\\nI1121 14:09:20.759274 6099 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1121 14:09:20.759278 6099 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1121 14:09:20.759284 6099 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1121 14:09:20.759289 6099 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1121 14:09:20.759296 6099 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1121 14:09:20.759303 6099 handler.go:208] Removed *v1.Node event handler 2\\\\nI1121 14:09:20.759302 6099 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1121 14:09:20.759313 6099 handler.go:208] Removed *v1.Node event handler 7\\\\nI1121 14:09:20.759343 6099 factory.go:656] Stopping watch factory\\\\nI1121 14:09:20.759354 6099 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1121 14:09:20.759360 6099 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1121 14:09:20.759390 6099 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1121 14:09:20.759533 6099 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1121 14:09:20.759593 6099 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:16Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://814079abfa7d1b029825da6c6c5f1e57d439ff21ccb626c5d4e52e41d059ad3a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-21T14:09:23Z\\\",\\\"message\\\":\\\"luster\\\\\\\", UUID:\\\\\\\"63b1440a-0908-4cab-8799-012fa1cf0b07\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-operator-lifecycle-manager/olm-operator-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-operator-lifecycle-manager/olm-operator-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-operator-lifecycle-manager/olm-operator-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.168\\\\\\\", Port:8443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1121 14:09:22.444620 6334 ovnkube.go:599] Stopped ovnkube\\\\nI1121 14:09:22.444715 6334 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1121 14:09:22.444771 6334 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b486a86e0e97ab409e590f30a1bf5cb9f584a7c67f96ff1236e42ca4f8675b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7877db6f5c2bf0eb31218c64d67bca48f921ff7ac62b4e495f29024c086bf738\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7877db6f5c2bf0eb31218c64d67bca48f921ff7ac62b4e495f29024c086bf738\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tpfb6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:25Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.537494 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7qfmq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"223b7002-3e67-4848-b1fb-db2921deb64e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381a46c6290da2b36e572f46c16f1a50aa3e25359bcfbc02c7a5a18e1484860e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2m27\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7qfmq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:25Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.554188 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6ljgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f3b70b5-9294-4f69-9528-500d28f34c89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://527cf28abdeef1b42d9c19fb88e3171dc05c4516c3a542a461b5b1daa85c6d25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78jhb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6ljgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:25Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.567414 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x5pws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc9e717e-af03-48bf-b005-62b3d084c6fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9c4f5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9c4f5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-x5pws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:25Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.586096 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6cd330fd-e8a6-49d5-bfc9-5f6122ccd49e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1bd64b9b938bc80284bcee54141e90cad613d339c4f27ef2934b29bc2a318916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://936f59996d26e42099dcd055f87fff9e3ac4283b57e1956fe97463f2334152d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca74b28a069dc2e609b0779164c3aa470a15ed9848af0dd5fc3479aaf750157b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://270c53b9a342f26be077a01cd4e87e9ccd0eb5d2c84208ae38cbf720463245e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66c521c26d541bbddb064dd9710f1a7b41090d2f47343977922272dfc6dab7af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6eb2ac410b49098e5b3603de5e6bf5ed4484fab0b300021d0bed2e5ed8a42527\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6eb2ac410b49098e5b3603de5e6bf5ed4484fab0b300021d0bed2e5ed8a42527\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03385e0c78e4806f60c160227ef4905a11003ed3d70f0e5c0dd0a477979d1d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03385e0c78e4806f60c160227ef4905a11003ed3d70f0e5c0dd0a477979d1d4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://50aea2fa8b3b141b4330ddce2c8e0715230c6f06dab64b3abfd98a3067224bec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50aea2fa8b3b141b4330ddce2c8e0715230c6f06dab64b3abfd98a3067224bec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:25Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.599241 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a6c0a99-565b-41f9-a8b5-c8990a0c8eb3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://737553dd1d16e2019ea89a376e3a868f9792ed5239aa7d6cc40a05d3d4f286ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6682a8886d2b9bce069cf9261d4701ef6907f19574fd035bd89bd05626023e5c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de079956e3d8de0101f3f87dcc120ef23876c4972382639e419b3f34a38cdfca\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://469b5be7467d4a58d3776ec0a967ed911be31b770b49965da4f5d2fb264d8a3d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1907bfbad7eb34da642e30d6aa81820aab457f92347333a364ac12396b0e9dae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"le observer\\\\nW1121 14:09:07.994401 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1121 14:09:07.994601 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1121 14:09:07.995290 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1461162674/tls.crt::/tmp/serving-cert-1461162674/tls.key\\\\\\\"\\\\nI1121 14:09:08.162382 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1121 14:09:08.165301 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1121 14:09:08.165318 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1121 14:09:08.165343 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1121 14:09:08.165348 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1121 14:09:08.172015 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1121 14:09:08.172083 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1121 14:09:08.172089 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1121 14:09:08.172094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1121 14:09:08.172098 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1121 14:09:08.172102 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1121 14:09:08.172106 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1121 14:09:08.172023 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1121 14:09:08.174214 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d78c87c440c7d8e30075f6143111b3f504fb99686484b8f19a9c8d61a55eccc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://101fb592fef069af009c613a1e9c9825024ed205498ab72334211ba90d6f7397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://101fb592fef069af009c613a1e9c9825024ed205498ab72334211ba90d6f7397\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:25Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.601271 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.601335 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.601347 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.601367 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.601379 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:25Z","lastTransitionTime":"2025-11-21T14:09:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.614021 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b78c993a-9966-4e84-85a1-b4f73aaf17f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce3a2602338a1c1e9380ee55f78360f3d9cf6ccc62678886dc5ae574a72e3d9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c256823f2fb49ecc121f6ef564cd86196f9b4a473e9bdea670173596ef14a823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c65f71f8746044f5d41687c670a0308cd09dc41a6531f9b5720e852b3d240756\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1dbe88e7e6ad63899e107ebb884a16edb9a42b21ae7c86d7fbfaf89785b0d36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:25Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.627389 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09877d9780caddd5f7f73e95fdf818a427ed645ed5997984113151f4b4759c13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:25Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.639936 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:25Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.657281 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:25Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.670132 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c976d804a7ebeba2c879e920cbc389f96f7f5def4b4439e8c8980d5460ecdb65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:25Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.691534 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nl76g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a7d6101-d0d2-4828-b3ce-d01040c04640\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac58b6ef1bd08e5f950260c18120c72e50b90a0de182c6ab0a5dbbbe45cf7328\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e56402f62821baf335f378084f4ed7c21b60ccbafe148419a0a9a403f106ad19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e56402f62821baf335f378084f4ed7c21b60ccbafe148419a0a9a403f106ad19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6022ff28e83dafcd846a7f03b090262aa3229a0da89f1984e4b6a054a04ac24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6022ff28e83dafcd846a7f03b090262aa3229a0da89f1984e4b6a054a04ac24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89de56e32b796da11bf89dd598d04fbf2884745ddf0f5d87cccfcaf1c001d8d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89de56e32b796da11bf89dd598d04fbf2884745ddf0f5d87cccfcaf1c001d8d6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2dff3cacaba7f9238f482e67b670ea775e0f96621434b99f4e76c2dfdde0321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2dff3cacaba7f9238f482e67b670ea775e0f96621434b99f4e76c2dfdde0321\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a153389a4137ee580e5118f3248400460f9d014c992a956f508815755f583afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a153389a4137ee580e5118f3248400460f9d014c992a956f508815755f583afb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a8c051ad1c1f2373f79eb0781fa6db489aeea3d52ed6c0bb43eec11520ec2b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a8c051ad1c1f2373f79eb0781fa6db489aeea3d52ed6c0bb43eec11520ec2b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nl76g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:25Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.703846 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.703898 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.703910 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.703929 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.703941 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:25Z","lastTransitionTime":"2025-11-21T14:09:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.706023 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:25Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.717651 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7670227-d280-4847-b882-754429f56b0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://961961696b8750c248598f4cda032b18395e5c802275dcfe6d3e9f7ef3510f1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5854028821e6a360a259427b68e7bb4c2a7220e0c68f4cc91639b2ece0540a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-krv5b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:25Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.738126 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"222f65f2-0b82-4760-bcf7-779244720f01\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92f1d932a83d65ba6a9a906cc489ff51e76547d07d42656630c59421f833c4db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccd51aa0faf3b1327f207b355cddf86a7353c0051f7e5ce0396f3c4bc67e3b16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8971c38daad45bda3d61545f52120dad40992dbf2e1d6e1e776500ed58d73f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfee19f82e35869a641318a0c45bc7a2ae71d80039f5790e60e277be15c9d3ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec279f2f558a8ac653006321b18cbaf13185876fe303c5cc1c65988aef09edbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de095f3c760b13d5197f624eac3806a1f4186577f2b95eb522532d453eb266e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://814079abfa7d1b029825da6c6c5f1e57d439ff21ccb626c5d4e52e41d059ad3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://089c9d4e6d0fb1a4e9a65a34a033e15c12508d2dbe3b9086b464914a1dac41e5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-21T14:09:20Z\\\",\\\"message\\\":\\\"60\\\\nI1121 14:09:20.759274 6099 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1121 14:09:20.759278 6099 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1121 14:09:20.759284 6099 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1121 14:09:20.759289 6099 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1121 14:09:20.759296 6099 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1121 14:09:20.759303 6099 handler.go:208] Removed *v1.Node event handler 2\\\\nI1121 14:09:20.759302 6099 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1121 14:09:20.759313 6099 handler.go:208] Removed *v1.Node event handler 7\\\\nI1121 14:09:20.759343 6099 factory.go:656] Stopping watch factory\\\\nI1121 14:09:20.759354 6099 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1121 14:09:20.759360 6099 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1121 14:09:20.759390 6099 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1121 14:09:20.759533 6099 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1121 14:09:20.759593 6099 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:16Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://814079abfa7d1b029825da6c6c5f1e57d439ff21ccb626c5d4e52e41d059ad3a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-21T14:09:23Z\\\",\\\"message\\\":\\\"luster\\\\\\\", UUID:\\\\\\\"63b1440a-0908-4cab-8799-012fa1cf0b07\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-operator-lifecycle-manager/olm-operator-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-operator-lifecycle-manager/olm-operator-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-operator-lifecycle-manager/olm-operator-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.168\\\\\\\", Port:8443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1121 14:09:22.444620 6334 ovnkube.go:599] Stopped ovnkube\\\\nI1121 14:09:22.444715 6334 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1121 14:09:22.444771 6334 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b486a86e0e97ab409e590f30a1bf5cb9f584a7c67f96ff1236e42ca4f8675b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7877db6f5c2bf0eb31218c64d67bca48f921ff7ac62b4e495f29024c086bf738\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7877db6f5c2bf0eb31218c64d67bca48f921ff7ac62b4e495f29024c086bf738\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tpfb6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:25Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.750420 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7qfmq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"223b7002-3e67-4848-b1fb-db2921deb64e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381a46c6290da2b36e572f46c16f1a50aa3e25359bcfbc02c7a5a18e1484860e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2m27\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7qfmq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:25Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.769467 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6cd330fd-e8a6-49d5-bfc9-5f6122ccd49e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1bd64b9b938bc80284bcee54141e90cad613d339c4f27ef2934b29bc2a318916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://936f59996d26e42099dcd055f87fff9e3ac4283b57e1956fe97463f2334152d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca74b28a069dc2e609b0779164c3aa470a15ed9848af0dd5fc3479aaf750157b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://270c53b9a342f26be077a01cd4e87e9ccd0eb5d2c84208ae38cbf720463245e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66c521c26d541bbddb064dd9710f1a7b41090d2f47343977922272dfc6dab7af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6eb2ac410b49098e5b3603de5e6bf5ed4484fab0b300021d0bed2e5ed8a42527\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6eb2ac410b49098e5b3603de5e6bf5ed4484fab0b300021d0bed2e5ed8a42527\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03385e0c78e4806f60c160227ef4905a11003ed3d70f0e5c0dd0a477979d1d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03385e0c78e4806f60c160227ef4905a11003ed3d70f0e5c0dd0a477979d1d4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://50aea2fa8b3b141b4330ddce2c8e0715230c6f06dab64b3abfd98a3067224bec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50aea2fa8b3b141b4330ddce2c8e0715230c6f06dab64b3abfd98a3067224bec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:25Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.781405 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a6c0a99-565b-41f9-a8b5-c8990a0c8eb3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://737553dd1d16e2019ea89a376e3a868f9792ed5239aa7d6cc40a05d3d4f286ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6682a8886d2b9bce069cf9261d4701ef6907f19574fd035bd89bd05626023e5c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de079956e3d8de0101f3f87dcc120ef23876c4972382639e419b3f34a38cdfca\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://469b5be7467d4a58d3776ec0a967ed911be31b770b49965da4f5d2fb264d8a3d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1907bfbad7eb34da642e30d6aa81820aab457f92347333a364ac12396b0e9dae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"le observer\\\\nW1121 14:09:07.994401 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1121 14:09:07.994601 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1121 14:09:07.995290 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1461162674/tls.crt::/tmp/serving-cert-1461162674/tls.key\\\\\\\"\\\\nI1121 14:09:08.162382 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1121 14:09:08.165301 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1121 14:09:08.165318 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1121 14:09:08.165343 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1121 14:09:08.165348 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1121 14:09:08.172015 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1121 14:09:08.172083 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1121 14:09:08.172089 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1121 14:09:08.172094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1121 14:09:08.172098 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1121 14:09:08.172102 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1121 14:09:08.172106 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1121 14:09:08.172023 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1121 14:09:08.174214 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d78c87c440c7d8e30075f6143111b3f504fb99686484b8f19a9c8d61a55eccc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://101fb592fef069af009c613a1e9c9825024ed205498ab72334211ba90d6f7397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://101fb592fef069af009c613a1e9c9825024ed205498ab72334211ba90d6f7397\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:25Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.792635 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b78c993a-9966-4e84-85a1-b4f73aaf17f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce3a2602338a1c1e9380ee55f78360f3d9cf6ccc62678886dc5ae574a72e3d9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c256823f2fb49ecc121f6ef564cd86196f9b4a473e9bdea670173596ef14a823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c65f71f8746044f5d41687c670a0308cd09dc41a6531f9b5720e852b3d240756\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1dbe88e7e6ad63899e107ebb884a16edb9a42b21ae7c86d7fbfaf89785b0d36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:25Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.802545 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09877d9780caddd5f7f73e95fdf818a427ed645ed5997984113151f4b4759c13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:25Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.808159 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.808225 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.808242 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.808271 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.808284 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:25Z","lastTransitionTime":"2025-11-21T14:09:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.817248 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:25Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.830471 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6ljgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f3b70b5-9294-4f69-9528-500d28f34c89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://527cf28abdeef1b42d9c19fb88e3171dc05c4516c3a542a461b5b1daa85c6d25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78jhb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6ljgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:25Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.845423 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x5pws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc9e717e-af03-48bf-b005-62b3d084c6fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f968eee1d8a1888229e23abda59fd58fa5e86f02ca96ff80fbe093904b27e8df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9c4f5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0f99b9213902c003e6ca754134d2482404ace39c184fb3b431d2b600481d5b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9c4f5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-x5pws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:25Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.858277 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:25Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.872144 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c976d804a7ebeba2c879e920cbc389f96f7f5def4b4439e8c8980d5460ecdb65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:25Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.885719 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nl76g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a7d6101-d0d2-4828-b3ce-d01040c04640\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac58b6ef1bd08e5f950260c18120c72e50b90a0de182c6ab0a5dbbbe45cf7328\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e56402f62821baf335f378084f4ed7c21b60ccbafe148419a0a9a403f106ad19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e56402f62821baf335f378084f4ed7c21b60ccbafe148419a0a9a403f106ad19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6022ff28e83dafcd846a7f03b090262aa3229a0da89f1984e4b6a054a04ac24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6022ff28e83dafcd846a7f03b090262aa3229a0da89f1984e4b6a054a04ac24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89de56e32b796da11bf89dd598d04fbf2884745ddf0f5d87cccfcaf1c001d8d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89de56e32b796da11bf89dd598d04fbf2884745ddf0f5d87cccfcaf1c001d8d6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2dff3cacaba7f9238f482e67b670ea775e0f96621434b99f4e76c2dfdde0321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2dff3cacaba7f9238f482e67b670ea775e0f96621434b99f4e76c2dfdde0321\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a153389a4137ee580e5118f3248400460f9d014c992a956f508815755f583afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a153389a4137ee580e5118f3248400460f9d014c992a956f508815755f583afb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a8c051ad1c1f2373f79eb0781fa6db489aeea3d52ed6c0bb43eec11520ec2b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a8c051ad1c1f2373f79eb0781fa6db489aeea3d52ed6c0bb43eec11520ec2b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nl76g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:25Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.897903 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48c6b1ea7497b33c00b956e400426c1d4f3c6c66e0225133434460800eef9314\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ea6068a6b6f8123b402de354e5374a0f02a7191a922d1718ab38f8898f47c52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:25Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.908062 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4qwqj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f46e2ba8-9ee0-406d-a668-2ba424618b69\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f43aa89d4d9defcb50d0dc6f90293b5f426b9cfc398b780c2e10f6a5c9f1e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbsgb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4qwqj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:25Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.911060 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.911106 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.911116 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.911137 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.911150 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:25Z","lastTransitionTime":"2025-11-21T14:09:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:25 crc kubenswrapper[4897]: I1121 14:09:25.921928 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-rs4c8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7xkdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7xkdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:23Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-rs4c8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:25Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:26 crc kubenswrapper[4897]: I1121 14:09:26.014600 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:26 crc kubenswrapper[4897]: I1121 14:09:26.014663 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:26 crc kubenswrapper[4897]: I1121 14:09:26.014673 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:26 crc kubenswrapper[4897]: I1121 14:09:26.014692 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:26 crc kubenswrapper[4897]: I1121 14:09:26.014708 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:26Z","lastTransitionTime":"2025-11-21T14:09:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:26 crc kubenswrapper[4897]: I1121 14:09:26.117675 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:26 crc kubenswrapper[4897]: I1121 14:09:26.117739 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:26 crc kubenswrapper[4897]: I1121 14:09:26.117749 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:26 crc kubenswrapper[4897]: I1121 14:09:26.117768 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:26 crc kubenswrapper[4897]: I1121 14:09:26.117783 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:26Z","lastTransitionTime":"2025-11-21T14:09:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:26 crc kubenswrapper[4897]: I1121 14:09:26.220083 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:26 crc kubenswrapper[4897]: I1121 14:09:26.220137 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:26 crc kubenswrapper[4897]: I1121 14:09:26.220148 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:26 crc kubenswrapper[4897]: I1121 14:09:26.220165 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:26 crc kubenswrapper[4897]: I1121 14:09:26.220177 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:26Z","lastTransitionTime":"2025-11-21T14:09:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:26 crc kubenswrapper[4897]: I1121 14:09:26.323373 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:26 crc kubenswrapper[4897]: I1121 14:09:26.323443 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:26 crc kubenswrapper[4897]: I1121 14:09:26.323456 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:26 crc kubenswrapper[4897]: I1121 14:09:26.323473 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:26 crc kubenswrapper[4897]: I1121 14:09:26.323486 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:26Z","lastTransitionTime":"2025-11-21T14:09:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:26 crc kubenswrapper[4897]: I1121 14:09:26.425991 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:26 crc kubenswrapper[4897]: I1121 14:09:26.426045 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:26 crc kubenswrapper[4897]: I1121 14:09:26.426058 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:26 crc kubenswrapper[4897]: I1121 14:09:26.426080 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:26 crc kubenswrapper[4897]: I1121 14:09:26.426096 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:26Z","lastTransitionTime":"2025-11-21T14:09:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:26 crc kubenswrapper[4897]: I1121 14:09:26.528463 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:26 crc kubenswrapper[4897]: I1121 14:09:26.528568 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:26 crc kubenswrapper[4897]: I1121 14:09:26.528580 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:26 crc kubenswrapper[4897]: I1121 14:09:26.528602 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:26 crc kubenswrapper[4897]: I1121 14:09:26.528615 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:26Z","lastTransitionTime":"2025-11-21T14:09:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:26 crc kubenswrapper[4897]: I1121 14:09:26.630820 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:26 crc kubenswrapper[4897]: I1121 14:09:26.630864 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:26 crc kubenswrapper[4897]: I1121 14:09:26.630873 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:26 crc kubenswrapper[4897]: I1121 14:09:26.630896 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:26 crc kubenswrapper[4897]: I1121 14:09:26.630907 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:26Z","lastTransitionTime":"2025-11-21T14:09:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:26 crc kubenswrapper[4897]: I1121 14:09:26.734670 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:26 crc kubenswrapper[4897]: I1121 14:09:26.734724 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:26 crc kubenswrapper[4897]: I1121 14:09:26.734742 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:26 crc kubenswrapper[4897]: I1121 14:09:26.734763 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:26 crc kubenswrapper[4897]: I1121 14:09:26.734777 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:26Z","lastTransitionTime":"2025-11-21T14:09:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:26 crc kubenswrapper[4897]: I1121 14:09:26.837105 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:26 crc kubenswrapper[4897]: I1121 14:09:26.837174 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:26 crc kubenswrapper[4897]: I1121 14:09:26.837193 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:26 crc kubenswrapper[4897]: I1121 14:09:26.837221 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:26 crc kubenswrapper[4897]: I1121 14:09:26.837244 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:26Z","lastTransitionTime":"2025-11-21T14:09:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:26 crc kubenswrapper[4897]: I1121 14:09:26.940159 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:26 crc kubenswrapper[4897]: I1121 14:09:26.940222 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:26 crc kubenswrapper[4897]: I1121 14:09:26.940248 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:26 crc kubenswrapper[4897]: I1121 14:09:26.940276 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:26 crc kubenswrapper[4897]: I1121 14:09:26.940295 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:26Z","lastTransitionTime":"2025-11-21T14:09:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:27 crc kubenswrapper[4897]: I1121 14:09:27.043477 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:27 crc kubenswrapper[4897]: I1121 14:09:27.043548 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:27 crc kubenswrapper[4897]: I1121 14:09:27.043560 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:27 crc kubenswrapper[4897]: I1121 14:09:27.043579 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:27 crc kubenswrapper[4897]: I1121 14:09:27.043593 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:27Z","lastTransitionTime":"2025-11-21T14:09:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:27 crc kubenswrapper[4897]: I1121 14:09:27.088416 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 21 14:09:27 crc kubenswrapper[4897]: I1121 14:09:27.088536 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 21 14:09:27 crc kubenswrapper[4897]: I1121 14:09:27.088568 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs4c8" Nov 21 14:09:27 crc kubenswrapper[4897]: I1121 14:09:27.088567 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 21 14:09:27 crc kubenswrapper[4897]: E1121 14:09:27.088612 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 21 14:09:27 crc kubenswrapper[4897]: E1121 14:09:27.088675 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 21 14:09:27 crc kubenswrapper[4897]: E1121 14:09:27.088841 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs4c8" podUID="01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd" Nov 21 14:09:27 crc kubenswrapper[4897]: E1121 14:09:27.088987 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 21 14:09:27 crc kubenswrapper[4897]: I1121 14:09:27.146330 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:27 crc kubenswrapper[4897]: I1121 14:09:27.146376 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:27 crc kubenswrapper[4897]: I1121 14:09:27.146387 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:27 crc kubenswrapper[4897]: I1121 14:09:27.146404 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:27 crc kubenswrapper[4897]: I1121 14:09:27.146416 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:27Z","lastTransitionTime":"2025-11-21T14:09:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:27 crc kubenswrapper[4897]: I1121 14:09:27.248933 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:27 crc kubenswrapper[4897]: I1121 14:09:27.248993 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:27 crc kubenswrapper[4897]: I1121 14:09:27.249011 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:27 crc kubenswrapper[4897]: I1121 14:09:27.249032 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:27 crc kubenswrapper[4897]: I1121 14:09:27.249046 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:27Z","lastTransitionTime":"2025-11-21T14:09:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:27 crc kubenswrapper[4897]: I1121 14:09:27.351590 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:27 crc kubenswrapper[4897]: I1121 14:09:27.351635 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:27 crc kubenswrapper[4897]: I1121 14:09:27.351645 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:27 crc kubenswrapper[4897]: I1121 14:09:27.351666 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:27 crc kubenswrapper[4897]: I1121 14:09:27.351686 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:27Z","lastTransitionTime":"2025-11-21T14:09:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:27 crc kubenswrapper[4897]: I1121 14:09:27.453589 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:27 crc kubenswrapper[4897]: I1121 14:09:27.453643 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:27 crc kubenswrapper[4897]: I1121 14:09:27.453653 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:27 crc kubenswrapper[4897]: I1121 14:09:27.453672 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:27 crc kubenswrapper[4897]: I1121 14:09:27.453683 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:27Z","lastTransitionTime":"2025-11-21T14:09:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:27 crc kubenswrapper[4897]: I1121 14:09:27.457308 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd-metrics-certs\") pod \"network-metrics-daemon-rs4c8\" (UID: \"01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd\") " pod="openshift-multus/network-metrics-daemon-rs4c8" Nov 21 14:09:27 crc kubenswrapper[4897]: E1121 14:09:27.457497 4897 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 21 14:09:27 crc kubenswrapper[4897]: E1121 14:09:27.457590 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd-metrics-certs podName:01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd nodeName:}" failed. No retries permitted until 2025-11-21 14:09:31.457568448 +0000 UTC m=+48.742161923 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd-metrics-certs") pod "network-metrics-daemon-rs4c8" (UID: "01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 21 14:09:27 crc kubenswrapper[4897]: I1121 14:09:27.556690 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:27 crc kubenswrapper[4897]: I1121 14:09:27.556769 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:27 crc kubenswrapper[4897]: I1121 14:09:27.556789 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:27 crc kubenswrapper[4897]: I1121 14:09:27.556822 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:27 crc kubenswrapper[4897]: I1121 14:09:27.556840 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:27Z","lastTransitionTime":"2025-11-21T14:09:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:27 crc kubenswrapper[4897]: I1121 14:09:27.660407 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:27 crc kubenswrapper[4897]: I1121 14:09:27.660464 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:27 crc kubenswrapper[4897]: I1121 14:09:27.660476 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:27 crc kubenswrapper[4897]: I1121 14:09:27.660495 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:27 crc kubenswrapper[4897]: I1121 14:09:27.660528 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:27Z","lastTransitionTime":"2025-11-21T14:09:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:27 crc kubenswrapper[4897]: I1121 14:09:27.762434 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:27 crc kubenswrapper[4897]: I1121 14:09:27.762476 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:27 crc kubenswrapper[4897]: I1121 14:09:27.762486 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:27 crc kubenswrapper[4897]: I1121 14:09:27.762503 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:27 crc kubenswrapper[4897]: I1121 14:09:27.762526 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:27Z","lastTransitionTime":"2025-11-21T14:09:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:27 crc kubenswrapper[4897]: I1121 14:09:27.865940 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:27 crc kubenswrapper[4897]: I1121 14:09:27.865984 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:27 crc kubenswrapper[4897]: I1121 14:09:27.865993 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:27 crc kubenswrapper[4897]: I1121 14:09:27.866013 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:27 crc kubenswrapper[4897]: I1121 14:09:27.866025 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:27Z","lastTransitionTime":"2025-11-21T14:09:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:27 crc kubenswrapper[4897]: I1121 14:09:27.969214 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:27 crc kubenswrapper[4897]: I1121 14:09:27.969254 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:27 crc kubenswrapper[4897]: I1121 14:09:27.969265 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:27 crc kubenswrapper[4897]: I1121 14:09:27.969282 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:27 crc kubenswrapper[4897]: I1121 14:09:27.969293 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:27Z","lastTransitionTime":"2025-11-21T14:09:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:28 crc kubenswrapper[4897]: I1121 14:09:28.072256 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:28 crc kubenswrapper[4897]: I1121 14:09:28.072292 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:28 crc kubenswrapper[4897]: I1121 14:09:28.072302 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:28 crc kubenswrapper[4897]: I1121 14:09:28.072318 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:28 crc kubenswrapper[4897]: I1121 14:09:28.072331 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:28Z","lastTransitionTime":"2025-11-21T14:09:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:28 crc kubenswrapper[4897]: I1121 14:09:28.174803 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:28 crc kubenswrapper[4897]: I1121 14:09:28.174854 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:28 crc kubenswrapper[4897]: I1121 14:09:28.174867 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:28 crc kubenswrapper[4897]: I1121 14:09:28.174885 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:28 crc kubenswrapper[4897]: I1121 14:09:28.174898 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:28Z","lastTransitionTime":"2025-11-21T14:09:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:28 crc kubenswrapper[4897]: I1121 14:09:28.277149 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:28 crc kubenswrapper[4897]: I1121 14:09:28.277195 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:28 crc kubenswrapper[4897]: I1121 14:09:28.277206 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:28 crc kubenswrapper[4897]: I1121 14:09:28.277224 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:28 crc kubenswrapper[4897]: I1121 14:09:28.277237 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:28Z","lastTransitionTime":"2025-11-21T14:09:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:28 crc kubenswrapper[4897]: I1121 14:09:28.379347 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:28 crc kubenswrapper[4897]: I1121 14:09:28.379389 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:28 crc kubenswrapper[4897]: I1121 14:09:28.379402 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:28 crc kubenswrapper[4897]: I1121 14:09:28.379419 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:28 crc kubenswrapper[4897]: I1121 14:09:28.379435 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:28Z","lastTransitionTime":"2025-11-21T14:09:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:28 crc kubenswrapper[4897]: I1121 14:09:28.482904 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:28 crc kubenswrapper[4897]: I1121 14:09:28.483011 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:28 crc kubenswrapper[4897]: I1121 14:09:28.483038 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:28 crc kubenswrapper[4897]: I1121 14:09:28.483072 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:28 crc kubenswrapper[4897]: I1121 14:09:28.483097 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:28Z","lastTransitionTime":"2025-11-21T14:09:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:28 crc kubenswrapper[4897]: I1121 14:09:28.586312 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:28 crc kubenswrapper[4897]: I1121 14:09:28.586383 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:28 crc kubenswrapper[4897]: I1121 14:09:28.586394 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:28 crc kubenswrapper[4897]: I1121 14:09:28.586409 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:28 crc kubenswrapper[4897]: I1121 14:09:28.586420 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:28Z","lastTransitionTime":"2025-11-21T14:09:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:28 crc kubenswrapper[4897]: I1121 14:09:28.689265 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:28 crc kubenswrapper[4897]: I1121 14:09:28.689327 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:28 crc kubenswrapper[4897]: I1121 14:09:28.689338 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:28 crc kubenswrapper[4897]: I1121 14:09:28.689359 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:28 crc kubenswrapper[4897]: I1121 14:09:28.689374 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:28Z","lastTransitionTime":"2025-11-21T14:09:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:28 crc kubenswrapper[4897]: I1121 14:09:28.791650 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:28 crc kubenswrapper[4897]: I1121 14:09:28.791707 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:28 crc kubenswrapper[4897]: I1121 14:09:28.791721 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:28 crc kubenswrapper[4897]: I1121 14:09:28.791741 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:28 crc kubenswrapper[4897]: I1121 14:09:28.791753 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:28Z","lastTransitionTime":"2025-11-21T14:09:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:28 crc kubenswrapper[4897]: I1121 14:09:28.894248 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:28 crc kubenswrapper[4897]: I1121 14:09:28.894335 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:28 crc kubenswrapper[4897]: I1121 14:09:28.894348 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:28 crc kubenswrapper[4897]: I1121 14:09:28.894372 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:28 crc kubenswrapper[4897]: I1121 14:09:28.894389 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:28Z","lastTransitionTime":"2025-11-21T14:09:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:28 crc kubenswrapper[4897]: I1121 14:09:28.997643 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:28 crc kubenswrapper[4897]: I1121 14:09:28.997699 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:28 crc kubenswrapper[4897]: I1121 14:09:28.997711 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:28 crc kubenswrapper[4897]: I1121 14:09:28.997732 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:28 crc kubenswrapper[4897]: I1121 14:09:28.997748 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:28Z","lastTransitionTime":"2025-11-21T14:09:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:29 crc kubenswrapper[4897]: I1121 14:09:29.088848 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs4c8" Nov 21 14:09:29 crc kubenswrapper[4897]: I1121 14:09:29.088953 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 21 14:09:29 crc kubenswrapper[4897]: I1121 14:09:29.089008 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 21 14:09:29 crc kubenswrapper[4897]: I1121 14:09:29.088889 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 21 14:09:29 crc kubenswrapper[4897]: E1121 14:09:29.089170 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs4c8" podUID="01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd" Nov 21 14:09:29 crc kubenswrapper[4897]: E1121 14:09:29.089277 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 21 14:09:29 crc kubenswrapper[4897]: E1121 14:09:29.089417 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 21 14:09:29 crc kubenswrapper[4897]: E1121 14:09:29.089684 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 21 14:09:29 crc kubenswrapper[4897]: I1121 14:09:29.100024 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:29 crc kubenswrapper[4897]: I1121 14:09:29.100078 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:29 crc kubenswrapper[4897]: I1121 14:09:29.100090 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:29 crc kubenswrapper[4897]: I1121 14:09:29.100111 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:29 crc kubenswrapper[4897]: I1121 14:09:29.100125 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:29Z","lastTransitionTime":"2025-11-21T14:09:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:29 crc kubenswrapper[4897]: I1121 14:09:29.202906 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:29 crc kubenswrapper[4897]: I1121 14:09:29.202963 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:29 crc kubenswrapper[4897]: I1121 14:09:29.202975 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:29 crc kubenswrapper[4897]: I1121 14:09:29.202995 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:29 crc kubenswrapper[4897]: I1121 14:09:29.203008 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:29Z","lastTransitionTime":"2025-11-21T14:09:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:29 crc kubenswrapper[4897]: I1121 14:09:29.305200 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:29 crc kubenswrapper[4897]: I1121 14:09:29.305247 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:29 crc kubenswrapper[4897]: I1121 14:09:29.305260 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:29 crc kubenswrapper[4897]: I1121 14:09:29.305280 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:29 crc kubenswrapper[4897]: I1121 14:09:29.305294 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:29Z","lastTransitionTime":"2025-11-21T14:09:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:29 crc kubenswrapper[4897]: I1121 14:09:29.408611 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:29 crc kubenswrapper[4897]: I1121 14:09:29.408665 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:29 crc kubenswrapper[4897]: I1121 14:09:29.408678 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:29 crc kubenswrapper[4897]: I1121 14:09:29.408695 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:29 crc kubenswrapper[4897]: I1121 14:09:29.408707 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:29Z","lastTransitionTime":"2025-11-21T14:09:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:29 crc kubenswrapper[4897]: I1121 14:09:29.512746 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:29 crc kubenswrapper[4897]: I1121 14:09:29.512810 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:29 crc kubenswrapper[4897]: I1121 14:09:29.512824 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:29 crc kubenswrapper[4897]: I1121 14:09:29.512849 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:29 crc kubenswrapper[4897]: I1121 14:09:29.512865 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:29Z","lastTransitionTime":"2025-11-21T14:09:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:29 crc kubenswrapper[4897]: I1121 14:09:29.616593 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:29 crc kubenswrapper[4897]: I1121 14:09:29.616659 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:29 crc kubenswrapper[4897]: I1121 14:09:29.616676 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:29 crc kubenswrapper[4897]: I1121 14:09:29.616697 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:29 crc kubenswrapper[4897]: I1121 14:09:29.616714 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:29Z","lastTransitionTime":"2025-11-21T14:09:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:29 crc kubenswrapper[4897]: I1121 14:09:29.718842 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:29 crc kubenswrapper[4897]: I1121 14:09:29.718902 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:29 crc kubenswrapper[4897]: I1121 14:09:29.718910 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:29 crc kubenswrapper[4897]: I1121 14:09:29.718928 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:29 crc kubenswrapper[4897]: I1121 14:09:29.718948 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:29Z","lastTransitionTime":"2025-11-21T14:09:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:29 crc kubenswrapper[4897]: I1121 14:09:29.822594 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:29 crc kubenswrapper[4897]: I1121 14:09:29.822671 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:29 crc kubenswrapper[4897]: I1121 14:09:29.822684 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:29 crc kubenswrapper[4897]: I1121 14:09:29.822711 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:29 crc kubenswrapper[4897]: I1121 14:09:29.822724 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:29Z","lastTransitionTime":"2025-11-21T14:09:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:29 crc kubenswrapper[4897]: I1121 14:09:29.925915 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:29 crc kubenswrapper[4897]: I1121 14:09:29.925991 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:29 crc kubenswrapper[4897]: I1121 14:09:29.926009 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:29 crc kubenswrapper[4897]: I1121 14:09:29.926033 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:29 crc kubenswrapper[4897]: I1121 14:09:29.926045 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:29Z","lastTransitionTime":"2025-11-21T14:09:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:30 crc kubenswrapper[4897]: I1121 14:09:30.028141 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:30 crc kubenswrapper[4897]: I1121 14:09:30.028202 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:30 crc kubenswrapper[4897]: I1121 14:09:30.028214 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:30 crc kubenswrapper[4897]: I1121 14:09:30.028233 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:30 crc kubenswrapper[4897]: I1121 14:09:30.028246 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:30Z","lastTransitionTime":"2025-11-21T14:09:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:30 crc kubenswrapper[4897]: I1121 14:09:30.131172 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:30 crc kubenswrapper[4897]: I1121 14:09:30.131223 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:30 crc kubenswrapper[4897]: I1121 14:09:30.131235 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:30 crc kubenswrapper[4897]: I1121 14:09:30.131254 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:30 crc kubenswrapper[4897]: I1121 14:09:30.131265 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:30Z","lastTransitionTime":"2025-11-21T14:09:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:30 crc kubenswrapper[4897]: I1121 14:09:30.233834 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:30 crc kubenswrapper[4897]: I1121 14:09:30.233911 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:30 crc kubenswrapper[4897]: I1121 14:09:30.233923 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:30 crc kubenswrapper[4897]: I1121 14:09:30.233945 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:30 crc kubenswrapper[4897]: I1121 14:09:30.233956 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:30Z","lastTransitionTime":"2025-11-21T14:09:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:30 crc kubenswrapper[4897]: I1121 14:09:30.336706 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:30 crc kubenswrapper[4897]: I1121 14:09:30.336760 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:30 crc kubenswrapper[4897]: I1121 14:09:30.336780 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:30 crc kubenswrapper[4897]: I1121 14:09:30.336806 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:30 crc kubenswrapper[4897]: I1121 14:09:30.336820 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:30Z","lastTransitionTime":"2025-11-21T14:09:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:30 crc kubenswrapper[4897]: I1121 14:09:30.439527 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:30 crc kubenswrapper[4897]: I1121 14:09:30.439559 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:30 crc kubenswrapper[4897]: I1121 14:09:30.439569 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:30 crc kubenswrapper[4897]: I1121 14:09:30.439583 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:30 crc kubenswrapper[4897]: I1121 14:09:30.439592 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:30Z","lastTransitionTime":"2025-11-21T14:09:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:30 crc kubenswrapper[4897]: I1121 14:09:30.543167 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:30 crc kubenswrapper[4897]: I1121 14:09:30.543221 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:30 crc kubenswrapper[4897]: I1121 14:09:30.543230 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:30 crc kubenswrapper[4897]: I1121 14:09:30.543246 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:30 crc kubenswrapper[4897]: I1121 14:09:30.543257 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:30Z","lastTransitionTime":"2025-11-21T14:09:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:30 crc kubenswrapper[4897]: I1121 14:09:30.646573 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:30 crc kubenswrapper[4897]: I1121 14:09:30.646615 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:30 crc kubenswrapper[4897]: I1121 14:09:30.646629 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:30 crc kubenswrapper[4897]: I1121 14:09:30.646647 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:30 crc kubenswrapper[4897]: I1121 14:09:30.646659 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:30Z","lastTransitionTime":"2025-11-21T14:09:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:30 crc kubenswrapper[4897]: I1121 14:09:30.749103 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:30 crc kubenswrapper[4897]: I1121 14:09:30.749150 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:30 crc kubenswrapper[4897]: I1121 14:09:30.749160 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:30 crc kubenswrapper[4897]: I1121 14:09:30.749175 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:30 crc kubenswrapper[4897]: I1121 14:09:30.749188 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:30Z","lastTransitionTime":"2025-11-21T14:09:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:30 crc kubenswrapper[4897]: I1121 14:09:30.852135 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:30 crc kubenswrapper[4897]: I1121 14:09:30.852190 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:30 crc kubenswrapper[4897]: I1121 14:09:30.852202 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:30 crc kubenswrapper[4897]: I1121 14:09:30.852221 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:30 crc kubenswrapper[4897]: I1121 14:09:30.852232 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:30Z","lastTransitionTime":"2025-11-21T14:09:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:30 crc kubenswrapper[4897]: I1121 14:09:30.954905 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:30 crc kubenswrapper[4897]: I1121 14:09:30.954948 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:30 crc kubenswrapper[4897]: I1121 14:09:30.954958 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:30 crc kubenswrapper[4897]: I1121 14:09:30.954972 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:30 crc kubenswrapper[4897]: I1121 14:09:30.954984 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:30Z","lastTransitionTime":"2025-11-21T14:09:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:31 crc kubenswrapper[4897]: I1121 14:09:31.058170 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:31 crc kubenswrapper[4897]: I1121 14:09:31.058232 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:31 crc kubenswrapper[4897]: I1121 14:09:31.058245 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:31 crc kubenswrapper[4897]: I1121 14:09:31.058265 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:31 crc kubenswrapper[4897]: I1121 14:09:31.058280 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:31Z","lastTransitionTime":"2025-11-21T14:09:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:31 crc kubenswrapper[4897]: I1121 14:09:31.088714 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 21 14:09:31 crc kubenswrapper[4897]: I1121 14:09:31.088813 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 21 14:09:31 crc kubenswrapper[4897]: I1121 14:09:31.088851 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs4c8" Nov 21 14:09:31 crc kubenswrapper[4897]: E1121 14:09:31.088857 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 21 14:09:31 crc kubenswrapper[4897]: I1121 14:09:31.088815 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 21 14:09:31 crc kubenswrapper[4897]: E1121 14:09:31.088959 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 21 14:09:31 crc kubenswrapper[4897]: E1121 14:09:31.089080 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 21 14:09:31 crc kubenswrapper[4897]: E1121 14:09:31.089184 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs4c8" podUID="01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd" Nov 21 14:09:31 crc kubenswrapper[4897]: I1121 14:09:31.160220 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:31 crc kubenswrapper[4897]: I1121 14:09:31.160255 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:31 crc kubenswrapper[4897]: I1121 14:09:31.160265 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:31 crc kubenswrapper[4897]: I1121 14:09:31.160278 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:31 crc kubenswrapper[4897]: I1121 14:09:31.160290 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:31Z","lastTransitionTime":"2025-11-21T14:09:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:31 crc kubenswrapper[4897]: I1121 14:09:31.262693 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:31 crc kubenswrapper[4897]: I1121 14:09:31.262751 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:31 crc kubenswrapper[4897]: I1121 14:09:31.262765 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:31 crc kubenswrapper[4897]: I1121 14:09:31.262783 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:31 crc kubenswrapper[4897]: I1121 14:09:31.262798 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:31Z","lastTransitionTime":"2025-11-21T14:09:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:31 crc kubenswrapper[4897]: I1121 14:09:31.365884 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:31 crc kubenswrapper[4897]: I1121 14:09:31.365934 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:31 crc kubenswrapper[4897]: I1121 14:09:31.365942 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:31 crc kubenswrapper[4897]: I1121 14:09:31.365958 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:31 crc kubenswrapper[4897]: I1121 14:09:31.365968 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:31Z","lastTransitionTime":"2025-11-21T14:09:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:31 crc kubenswrapper[4897]: I1121 14:09:31.468658 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:31 crc kubenswrapper[4897]: I1121 14:09:31.468718 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:31 crc kubenswrapper[4897]: I1121 14:09:31.468731 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:31 crc kubenswrapper[4897]: I1121 14:09:31.468753 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:31 crc kubenswrapper[4897]: I1121 14:09:31.468768 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:31Z","lastTransitionTime":"2025-11-21T14:09:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:31 crc kubenswrapper[4897]: I1121 14:09:31.503472 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd-metrics-certs\") pod \"network-metrics-daemon-rs4c8\" (UID: \"01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd\") " pod="openshift-multus/network-metrics-daemon-rs4c8" Nov 21 14:09:31 crc kubenswrapper[4897]: E1121 14:09:31.503726 4897 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 21 14:09:31 crc kubenswrapper[4897]: E1121 14:09:31.503843 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd-metrics-certs podName:01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd nodeName:}" failed. No retries permitted until 2025-11-21 14:09:39.503814398 +0000 UTC m=+56.788407883 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd-metrics-certs") pod "network-metrics-daemon-rs4c8" (UID: "01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 21 14:09:31 crc kubenswrapper[4897]: I1121 14:09:31.571369 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:31 crc kubenswrapper[4897]: I1121 14:09:31.571427 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:31 crc kubenswrapper[4897]: I1121 14:09:31.571445 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:31 crc kubenswrapper[4897]: I1121 14:09:31.571490 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:31 crc kubenswrapper[4897]: I1121 14:09:31.571534 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:31Z","lastTransitionTime":"2025-11-21T14:09:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:31 crc kubenswrapper[4897]: I1121 14:09:31.673944 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:31 crc kubenswrapper[4897]: I1121 14:09:31.674027 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:31 crc kubenswrapper[4897]: I1121 14:09:31.674049 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:31 crc kubenswrapper[4897]: I1121 14:09:31.674072 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:31 crc kubenswrapper[4897]: I1121 14:09:31.674090 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:31Z","lastTransitionTime":"2025-11-21T14:09:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:31 crc kubenswrapper[4897]: I1121 14:09:31.777550 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:31 crc kubenswrapper[4897]: I1121 14:09:31.777638 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:31 crc kubenswrapper[4897]: I1121 14:09:31.777649 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:31 crc kubenswrapper[4897]: I1121 14:09:31.777670 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:31 crc kubenswrapper[4897]: I1121 14:09:31.777682 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:31Z","lastTransitionTime":"2025-11-21T14:09:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:31 crc kubenswrapper[4897]: I1121 14:09:31.880673 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:31 crc kubenswrapper[4897]: I1121 14:09:31.880709 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:31 crc kubenswrapper[4897]: I1121 14:09:31.880735 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:31 crc kubenswrapper[4897]: I1121 14:09:31.880755 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:31 crc kubenswrapper[4897]: I1121 14:09:31.880767 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:31Z","lastTransitionTime":"2025-11-21T14:09:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:31 crc kubenswrapper[4897]: I1121 14:09:31.983919 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:31 crc kubenswrapper[4897]: I1121 14:09:31.983969 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:31 crc kubenswrapper[4897]: I1121 14:09:31.983989 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:31 crc kubenswrapper[4897]: I1121 14:09:31.984009 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:31 crc kubenswrapper[4897]: I1121 14:09:31.984020 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:31Z","lastTransitionTime":"2025-11-21T14:09:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:32 crc kubenswrapper[4897]: I1121 14:09:32.086819 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:32 crc kubenswrapper[4897]: I1121 14:09:32.086861 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:32 crc kubenswrapper[4897]: I1121 14:09:32.086872 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:32 crc kubenswrapper[4897]: I1121 14:09:32.086891 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:32 crc kubenswrapper[4897]: I1121 14:09:32.086903 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:32Z","lastTransitionTime":"2025-11-21T14:09:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:32 crc kubenswrapper[4897]: I1121 14:09:32.188931 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:32 crc kubenswrapper[4897]: I1121 14:09:32.188972 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:32 crc kubenswrapper[4897]: I1121 14:09:32.188982 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:32 crc kubenswrapper[4897]: I1121 14:09:32.188996 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:32 crc kubenswrapper[4897]: I1121 14:09:32.189005 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:32Z","lastTransitionTime":"2025-11-21T14:09:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:32 crc kubenswrapper[4897]: I1121 14:09:32.291947 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:32 crc kubenswrapper[4897]: I1121 14:09:32.291998 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:32 crc kubenswrapper[4897]: I1121 14:09:32.292007 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:32 crc kubenswrapper[4897]: I1121 14:09:32.292024 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:32 crc kubenswrapper[4897]: I1121 14:09:32.292037 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:32Z","lastTransitionTime":"2025-11-21T14:09:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:32 crc kubenswrapper[4897]: I1121 14:09:32.395855 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:32 crc kubenswrapper[4897]: I1121 14:09:32.395898 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:32 crc kubenswrapper[4897]: I1121 14:09:32.395907 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:32 crc kubenswrapper[4897]: I1121 14:09:32.395926 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:32 crc kubenswrapper[4897]: I1121 14:09:32.395938 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:32Z","lastTransitionTime":"2025-11-21T14:09:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:32 crc kubenswrapper[4897]: I1121 14:09:32.498756 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:32 crc kubenswrapper[4897]: I1121 14:09:32.498868 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:32 crc kubenswrapper[4897]: I1121 14:09:32.498884 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:32 crc kubenswrapper[4897]: I1121 14:09:32.498910 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:32 crc kubenswrapper[4897]: I1121 14:09:32.498924 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:32Z","lastTransitionTime":"2025-11-21T14:09:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:32 crc kubenswrapper[4897]: I1121 14:09:32.601828 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:32 crc kubenswrapper[4897]: I1121 14:09:32.601878 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:32 crc kubenswrapper[4897]: I1121 14:09:32.601886 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:32 crc kubenswrapper[4897]: I1121 14:09:32.601904 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:32 crc kubenswrapper[4897]: I1121 14:09:32.601918 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:32Z","lastTransitionTime":"2025-11-21T14:09:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:32 crc kubenswrapper[4897]: I1121 14:09:32.704054 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:32 crc kubenswrapper[4897]: I1121 14:09:32.704110 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:32 crc kubenswrapper[4897]: I1121 14:09:32.704121 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:32 crc kubenswrapper[4897]: I1121 14:09:32.704142 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:32 crc kubenswrapper[4897]: I1121 14:09:32.704156 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:32Z","lastTransitionTime":"2025-11-21T14:09:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:32 crc kubenswrapper[4897]: I1121 14:09:32.806768 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:32 crc kubenswrapper[4897]: I1121 14:09:32.806801 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:32 crc kubenswrapper[4897]: I1121 14:09:32.806810 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:32 crc kubenswrapper[4897]: I1121 14:09:32.806824 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:32 crc kubenswrapper[4897]: I1121 14:09:32.806832 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:32Z","lastTransitionTime":"2025-11-21T14:09:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:32 crc kubenswrapper[4897]: I1121 14:09:32.909196 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:32 crc kubenswrapper[4897]: I1121 14:09:32.909235 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:32 crc kubenswrapper[4897]: I1121 14:09:32.909245 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:32 crc kubenswrapper[4897]: I1121 14:09:32.909261 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:32 crc kubenswrapper[4897]: I1121 14:09:32.909271 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:32Z","lastTransitionTime":"2025-11-21T14:09:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:33 crc kubenswrapper[4897]: I1121 14:09:33.011721 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:33 crc kubenswrapper[4897]: I1121 14:09:33.011765 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:33 crc kubenswrapper[4897]: I1121 14:09:33.011776 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:33 crc kubenswrapper[4897]: I1121 14:09:33.011792 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:33 crc kubenswrapper[4897]: I1121 14:09:33.011804 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:33Z","lastTransitionTime":"2025-11-21T14:09:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:33 crc kubenswrapper[4897]: I1121 14:09:33.088688 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 21 14:09:33 crc kubenswrapper[4897]: I1121 14:09:33.088760 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 21 14:09:33 crc kubenswrapper[4897]: I1121 14:09:33.088688 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 21 14:09:33 crc kubenswrapper[4897]: E1121 14:09:33.088825 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 21 14:09:33 crc kubenswrapper[4897]: I1121 14:09:33.088691 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs4c8" Nov 21 14:09:33 crc kubenswrapper[4897]: E1121 14:09:33.088892 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 21 14:09:33 crc kubenswrapper[4897]: E1121 14:09:33.088956 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs4c8" podUID="01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd" Nov 21 14:09:33 crc kubenswrapper[4897]: E1121 14:09:33.089024 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 21 14:09:33 crc kubenswrapper[4897]: I1121 14:09:33.113942 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:33 crc kubenswrapper[4897]: I1121 14:09:33.113977 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:33 crc kubenswrapper[4897]: I1121 14:09:33.114008 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:33 crc kubenswrapper[4897]: I1121 14:09:33.114030 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:33 crc kubenswrapper[4897]: I1121 14:09:33.114042 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:33Z","lastTransitionTime":"2025-11-21T14:09:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:33 crc kubenswrapper[4897]: I1121 14:09:33.217242 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:33 crc kubenswrapper[4897]: I1121 14:09:33.217295 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:33 crc kubenswrapper[4897]: I1121 14:09:33.217304 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:33 crc kubenswrapper[4897]: I1121 14:09:33.217320 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:33 crc kubenswrapper[4897]: I1121 14:09:33.217332 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:33Z","lastTransitionTime":"2025-11-21T14:09:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:33 crc kubenswrapper[4897]: I1121 14:09:33.319674 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:33 crc kubenswrapper[4897]: I1121 14:09:33.319750 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:33 crc kubenswrapper[4897]: I1121 14:09:33.319767 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:33 crc kubenswrapper[4897]: I1121 14:09:33.319797 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:33 crc kubenswrapper[4897]: I1121 14:09:33.319820 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:33Z","lastTransitionTime":"2025-11-21T14:09:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:33 crc kubenswrapper[4897]: I1121 14:09:33.422564 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:33 crc kubenswrapper[4897]: I1121 14:09:33.422607 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:33 crc kubenswrapper[4897]: I1121 14:09:33.422618 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:33 crc kubenswrapper[4897]: I1121 14:09:33.422634 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:33 crc kubenswrapper[4897]: I1121 14:09:33.422646 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:33Z","lastTransitionTime":"2025-11-21T14:09:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:33 crc kubenswrapper[4897]: I1121 14:09:33.525740 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:33 crc kubenswrapper[4897]: I1121 14:09:33.525794 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:33 crc kubenswrapper[4897]: I1121 14:09:33.525812 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:33 crc kubenswrapper[4897]: I1121 14:09:33.525831 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:33 crc kubenswrapper[4897]: I1121 14:09:33.525847 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:33Z","lastTransitionTime":"2025-11-21T14:09:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:33 crc kubenswrapper[4897]: I1121 14:09:33.628360 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:33 crc kubenswrapper[4897]: I1121 14:09:33.628416 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:33 crc kubenswrapper[4897]: I1121 14:09:33.628432 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:33 crc kubenswrapper[4897]: I1121 14:09:33.628480 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:33 crc kubenswrapper[4897]: I1121 14:09:33.628496 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:33Z","lastTransitionTime":"2025-11-21T14:09:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:33 crc kubenswrapper[4897]: I1121 14:09:33.730701 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:33 crc kubenswrapper[4897]: I1121 14:09:33.730735 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:33 crc kubenswrapper[4897]: I1121 14:09:33.730745 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:33 crc kubenswrapper[4897]: I1121 14:09:33.730760 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:33 crc kubenswrapper[4897]: I1121 14:09:33.730769 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:33Z","lastTransitionTime":"2025-11-21T14:09:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:33 crc kubenswrapper[4897]: I1121 14:09:33.833564 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:33 crc kubenswrapper[4897]: I1121 14:09:33.833629 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:33 crc kubenswrapper[4897]: I1121 14:09:33.833646 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:33 crc kubenswrapper[4897]: I1121 14:09:33.833667 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:33 crc kubenswrapper[4897]: I1121 14:09:33.833682 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:33Z","lastTransitionTime":"2025-11-21T14:09:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:33 crc kubenswrapper[4897]: I1121 14:09:33.936214 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:33 crc kubenswrapper[4897]: I1121 14:09:33.936265 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:33 crc kubenswrapper[4897]: I1121 14:09:33.936275 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:33 crc kubenswrapper[4897]: I1121 14:09:33.936293 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:33 crc kubenswrapper[4897]: I1121 14:09:33.936304 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:33Z","lastTransitionTime":"2025-11-21T14:09:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:34 crc kubenswrapper[4897]: I1121 14:09:34.039199 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:34 crc kubenswrapper[4897]: I1121 14:09:34.039248 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:34 crc kubenswrapper[4897]: I1121 14:09:34.039260 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:34 crc kubenswrapper[4897]: I1121 14:09:34.039281 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:34 crc kubenswrapper[4897]: I1121 14:09:34.039295 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:34Z","lastTransitionTime":"2025-11-21T14:09:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:34 crc kubenswrapper[4897]: I1121 14:09:34.102098 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:34Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:34 crc kubenswrapper[4897]: I1121 14:09:34.117249 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c976d804a7ebeba2c879e920cbc389f96f7f5def4b4439e8c8980d5460ecdb65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:34Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:34 crc kubenswrapper[4897]: I1121 14:09:34.132583 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nl76g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a7d6101-d0d2-4828-b3ce-d01040c04640\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac58b6ef1bd08e5f950260c18120c72e50b90a0de182c6ab0a5dbbbe45cf7328\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e56402f62821baf335f378084f4ed7c21b60ccbafe148419a0a9a403f106ad19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e56402f62821baf335f378084f4ed7c21b60ccbafe148419a0a9a403f106ad19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6022ff28e83dafcd846a7f03b090262aa3229a0da89f1984e4b6a054a04ac24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6022ff28e83dafcd846a7f03b090262aa3229a0da89f1984e4b6a054a04ac24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89de56e32b796da11bf89dd598d04fbf2884745ddf0f5d87cccfcaf1c001d8d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89de56e32b796da11bf89dd598d04fbf2884745ddf0f5d87cccfcaf1c001d8d6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2dff3cacaba7f9238f482e67b670ea775e0f96621434b99f4e76c2dfdde0321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2dff3cacaba7f9238f482e67b670ea775e0f96621434b99f4e76c2dfdde0321\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a153389a4137ee580e5118f3248400460f9d014c992a956f508815755f583afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a153389a4137ee580e5118f3248400460f9d014c992a956f508815755f583afb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a8c051ad1c1f2373f79eb0781fa6db489aeea3d52ed6c0bb43eec11520ec2b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a8c051ad1c1f2373f79eb0781fa6db489aeea3d52ed6c0bb43eec11520ec2b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nl76g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:34Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:34 crc kubenswrapper[4897]: I1121 14:09:34.141846 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:34 crc kubenswrapper[4897]: I1121 14:09:34.141893 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:34 crc kubenswrapper[4897]: I1121 14:09:34.141905 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:34 crc kubenswrapper[4897]: I1121 14:09:34.141924 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:34 crc kubenswrapper[4897]: I1121 14:09:34.141943 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:34Z","lastTransitionTime":"2025-11-21T14:09:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:34 crc kubenswrapper[4897]: I1121 14:09:34.146727 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48c6b1ea7497b33c00b956e400426c1d4f3c6c66e0225133434460800eef9314\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ea6068a6b6f8123b402de354e5374a0f02a7191a922d1718ab38f8898f47c52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:34Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:34 crc kubenswrapper[4897]: I1121 14:09:34.161123 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4qwqj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f46e2ba8-9ee0-406d-a668-2ba424618b69\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f43aa89d4d9defcb50d0dc6f90293b5f426b9cfc398b780c2e10f6a5c9f1e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbsgb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4qwqj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:34Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:34 crc kubenswrapper[4897]: I1121 14:09:34.172235 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-rs4c8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7xkdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7xkdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:23Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-rs4c8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:34Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:34 crc kubenswrapper[4897]: I1121 14:09:34.189219 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:34Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:34 crc kubenswrapper[4897]: I1121 14:09:34.201967 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7670227-d280-4847-b882-754429f56b0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://961961696b8750c248598f4cda032b18395e5c802275dcfe6d3e9f7ef3510f1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5854028821e6a360a259427b68e7bb4c2a7220e0c68f4cc91639b2ece0540a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-krv5b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:34Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:34 crc kubenswrapper[4897]: I1121 14:09:34.222569 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"222f65f2-0b82-4760-bcf7-779244720f01\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92f1d932a83d65ba6a9a906cc489ff51e76547d07d42656630c59421f833c4db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccd51aa0faf3b1327f207b355cddf86a7353c0051f7e5ce0396f3c4bc67e3b16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8971c38daad45bda3d61545f52120dad40992dbf2e1d6e1e776500ed58d73f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfee19f82e35869a641318a0c45bc7a2ae71d80039f5790e60e277be15c9d3ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec279f2f558a8ac653006321b18cbaf13185876fe303c5cc1c65988aef09edbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de095f3c760b13d5197f624eac3806a1f4186577f2b95eb522532d453eb266e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://814079abfa7d1b029825da6c6c5f1e57d439ff21ccb626c5d4e52e41d059ad3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://089c9d4e6d0fb1a4e9a65a34a033e15c12508d2dbe3b9086b464914a1dac41e5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-21T14:09:20Z\\\",\\\"message\\\":\\\"60\\\\nI1121 14:09:20.759274 6099 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1121 14:09:20.759278 6099 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1121 14:09:20.759284 6099 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1121 14:09:20.759289 6099 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1121 14:09:20.759296 6099 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1121 14:09:20.759303 6099 handler.go:208] Removed *v1.Node event handler 2\\\\nI1121 14:09:20.759302 6099 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1121 14:09:20.759313 6099 handler.go:208] Removed *v1.Node event handler 7\\\\nI1121 14:09:20.759343 6099 factory.go:656] Stopping watch factory\\\\nI1121 14:09:20.759354 6099 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1121 14:09:20.759360 6099 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1121 14:09:20.759390 6099 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1121 14:09:20.759533 6099 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1121 14:09:20.759593 6099 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:16Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://814079abfa7d1b029825da6c6c5f1e57d439ff21ccb626c5d4e52e41d059ad3a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-21T14:09:23Z\\\",\\\"message\\\":\\\"luster\\\\\\\", UUID:\\\\\\\"63b1440a-0908-4cab-8799-012fa1cf0b07\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-operator-lifecycle-manager/olm-operator-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-operator-lifecycle-manager/olm-operator-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-operator-lifecycle-manager/olm-operator-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.168\\\\\\\", Port:8443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1121 14:09:22.444620 6334 ovnkube.go:599] Stopped ovnkube\\\\nI1121 14:09:22.444715 6334 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1121 14:09:22.444771 6334 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b486a86e0e97ab409e590f30a1bf5cb9f584a7c67f96ff1236e42ca4f8675b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7877db6f5c2bf0eb31218c64d67bca48f921ff7ac62b4e495f29024c086bf738\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7877db6f5c2bf0eb31218c64d67bca48f921ff7ac62b4e495f29024c086bf738\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tpfb6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:34Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:34 crc kubenswrapper[4897]: I1121 14:09:34.238992 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7qfmq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"223b7002-3e67-4848-b1fb-db2921deb64e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381a46c6290da2b36e572f46c16f1a50aa3e25359bcfbc02c7a5a18e1484860e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2m27\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7qfmq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:34Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:34 crc kubenswrapper[4897]: I1121 14:09:34.244614 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:34 crc kubenswrapper[4897]: I1121 14:09:34.244647 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:34 crc kubenswrapper[4897]: I1121 14:09:34.244655 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:34 crc kubenswrapper[4897]: I1121 14:09:34.244669 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:34 crc kubenswrapper[4897]: I1121 14:09:34.244680 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:34Z","lastTransitionTime":"2025-11-21T14:09:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:34 crc kubenswrapper[4897]: I1121 14:09:34.252110 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6ljgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f3b70b5-9294-4f69-9528-500d28f34c89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://527cf28abdeef1b42d9c19fb88e3171dc05c4516c3a542a461b5b1daa85c6d25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78jhb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6ljgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:34Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:34 crc kubenswrapper[4897]: I1121 14:09:34.264085 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x5pws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc9e717e-af03-48bf-b005-62b3d084c6fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f968eee1d8a1888229e23abda59fd58fa5e86f02ca96ff80fbe093904b27e8df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9c4f5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0f99b9213902c003e6ca754134d2482404ace39c184fb3b431d2b600481d5b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9c4f5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-x5pws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:34Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:34 crc kubenswrapper[4897]: I1121 14:09:34.289104 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6cd330fd-e8a6-49d5-bfc9-5f6122ccd49e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1bd64b9b938bc80284bcee54141e90cad613d339c4f27ef2934b29bc2a318916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://936f59996d26e42099dcd055f87fff9e3ac4283b57e1956fe97463f2334152d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca74b28a069dc2e609b0779164c3aa470a15ed9848af0dd5fc3479aaf750157b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://270c53b9a342f26be077a01cd4e87e9ccd0eb5d2c84208ae38cbf720463245e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66c521c26d541bbddb064dd9710f1a7b41090d2f47343977922272dfc6dab7af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6eb2ac410b49098e5b3603de5e6bf5ed4484fab0b300021d0bed2e5ed8a42527\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6eb2ac410b49098e5b3603de5e6bf5ed4484fab0b300021d0bed2e5ed8a42527\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03385e0c78e4806f60c160227ef4905a11003ed3d70f0e5c0dd0a477979d1d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03385e0c78e4806f60c160227ef4905a11003ed3d70f0e5c0dd0a477979d1d4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://50aea2fa8b3b141b4330ddce2c8e0715230c6f06dab64b3abfd98a3067224bec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50aea2fa8b3b141b4330ddce2c8e0715230c6f06dab64b3abfd98a3067224bec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:34Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:34 crc kubenswrapper[4897]: I1121 14:09:34.304593 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a6c0a99-565b-41f9-a8b5-c8990a0c8eb3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://737553dd1d16e2019ea89a376e3a868f9792ed5239aa7d6cc40a05d3d4f286ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6682a8886d2b9bce069cf9261d4701ef6907f19574fd035bd89bd05626023e5c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de079956e3d8de0101f3f87dcc120ef23876c4972382639e419b3f34a38cdfca\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://469b5be7467d4a58d3776ec0a967ed911be31b770b49965da4f5d2fb264d8a3d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1907bfbad7eb34da642e30d6aa81820aab457f92347333a364ac12396b0e9dae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"le observer\\\\nW1121 14:09:07.994401 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1121 14:09:07.994601 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1121 14:09:07.995290 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1461162674/tls.crt::/tmp/serving-cert-1461162674/tls.key\\\\\\\"\\\\nI1121 14:09:08.162382 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1121 14:09:08.165301 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1121 14:09:08.165318 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1121 14:09:08.165343 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1121 14:09:08.165348 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1121 14:09:08.172015 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1121 14:09:08.172083 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1121 14:09:08.172089 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1121 14:09:08.172094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1121 14:09:08.172098 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1121 14:09:08.172102 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1121 14:09:08.172106 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1121 14:09:08.172023 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1121 14:09:08.174214 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d78c87c440c7d8e30075f6143111b3f504fb99686484b8f19a9c8d61a55eccc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://101fb592fef069af009c613a1e9c9825024ed205498ab72334211ba90d6f7397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://101fb592fef069af009c613a1e9c9825024ed205498ab72334211ba90d6f7397\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:34Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:34 crc kubenswrapper[4897]: I1121 14:09:34.317686 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b78c993a-9966-4e84-85a1-b4f73aaf17f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce3a2602338a1c1e9380ee55f78360f3d9cf6ccc62678886dc5ae574a72e3d9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c256823f2fb49ecc121f6ef564cd86196f9b4a473e9bdea670173596ef14a823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c65f71f8746044f5d41687c670a0308cd09dc41a6531f9b5720e852b3d240756\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1dbe88e7e6ad63899e107ebb884a16edb9a42b21ae7c86d7fbfaf89785b0d36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:34Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:34 crc kubenswrapper[4897]: I1121 14:09:34.336160 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09877d9780caddd5f7f73e95fdf818a427ed645ed5997984113151f4b4759c13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:34Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:34 crc kubenswrapper[4897]: I1121 14:09:34.346609 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:34 crc kubenswrapper[4897]: I1121 14:09:34.346647 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:34 crc kubenswrapper[4897]: I1121 14:09:34.346656 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:34 crc kubenswrapper[4897]: I1121 14:09:34.346672 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:34 crc kubenswrapper[4897]: I1121 14:09:34.346682 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:34Z","lastTransitionTime":"2025-11-21T14:09:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:34 crc kubenswrapper[4897]: I1121 14:09:34.350746 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:34Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:34 crc kubenswrapper[4897]: I1121 14:09:34.449599 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:34 crc kubenswrapper[4897]: I1121 14:09:34.449653 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:34 crc kubenswrapper[4897]: I1121 14:09:34.449668 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:34 crc kubenswrapper[4897]: I1121 14:09:34.449688 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:34 crc kubenswrapper[4897]: I1121 14:09:34.449701 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:34Z","lastTransitionTime":"2025-11-21T14:09:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:34 crc kubenswrapper[4897]: I1121 14:09:34.551801 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:34 crc kubenswrapper[4897]: I1121 14:09:34.551853 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:34 crc kubenswrapper[4897]: I1121 14:09:34.551865 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:34 crc kubenswrapper[4897]: I1121 14:09:34.551883 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:34 crc kubenswrapper[4897]: I1121 14:09:34.551893 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:34Z","lastTransitionTime":"2025-11-21T14:09:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:34 crc kubenswrapper[4897]: I1121 14:09:34.653860 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:34 crc kubenswrapper[4897]: I1121 14:09:34.653907 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:34 crc kubenswrapper[4897]: I1121 14:09:34.653917 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:34 crc kubenswrapper[4897]: I1121 14:09:34.653934 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:34 crc kubenswrapper[4897]: I1121 14:09:34.653947 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:34Z","lastTransitionTime":"2025-11-21T14:09:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:34 crc kubenswrapper[4897]: I1121 14:09:34.756569 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:34 crc kubenswrapper[4897]: I1121 14:09:34.756619 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:34 crc kubenswrapper[4897]: I1121 14:09:34.756629 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:34 crc kubenswrapper[4897]: I1121 14:09:34.756646 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:34 crc kubenswrapper[4897]: I1121 14:09:34.756657 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:34Z","lastTransitionTime":"2025-11-21T14:09:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:34 crc kubenswrapper[4897]: I1121 14:09:34.859615 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:34 crc kubenswrapper[4897]: I1121 14:09:34.859678 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:34 crc kubenswrapper[4897]: I1121 14:09:34.859692 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:34 crc kubenswrapper[4897]: I1121 14:09:34.859715 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:34 crc kubenswrapper[4897]: I1121 14:09:34.859731 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:34Z","lastTransitionTime":"2025-11-21T14:09:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:34 crc kubenswrapper[4897]: I1121 14:09:34.962253 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:34 crc kubenswrapper[4897]: I1121 14:09:34.962302 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:34 crc kubenswrapper[4897]: I1121 14:09:34.962313 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:34 crc kubenswrapper[4897]: I1121 14:09:34.962329 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:34 crc kubenswrapper[4897]: I1121 14:09:34.962340 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:34Z","lastTransitionTime":"2025-11-21T14:09:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:35 crc kubenswrapper[4897]: I1121 14:09:35.064794 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:35 crc kubenswrapper[4897]: I1121 14:09:35.064838 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:35 crc kubenswrapper[4897]: I1121 14:09:35.064848 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:35 crc kubenswrapper[4897]: I1121 14:09:35.064864 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:35 crc kubenswrapper[4897]: I1121 14:09:35.064875 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:35Z","lastTransitionTime":"2025-11-21T14:09:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:35 crc kubenswrapper[4897]: I1121 14:09:35.088086 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 21 14:09:35 crc kubenswrapper[4897]: E1121 14:09:35.088216 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 21 14:09:35 crc kubenswrapper[4897]: I1121 14:09:35.088095 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 21 14:09:35 crc kubenswrapper[4897]: I1121 14:09:35.088090 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 21 14:09:35 crc kubenswrapper[4897]: E1121 14:09:35.088295 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 21 14:09:35 crc kubenswrapper[4897]: I1121 14:09:35.088110 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs4c8" Nov 21 14:09:35 crc kubenswrapper[4897]: E1121 14:09:35.088707 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 21 14:09:35 crc kubenswrapper[4897]: E1121 14:09:35.088781 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs4c8" podUID="01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd" Nov 21 14:09:35 crc kubenswrapper[4897]: I1121 14:09:35.167925 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:35 crc kubenswrapper[4897]: I1121 14:09:35.167990 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:35 crc kubenswrapper[4897]: I1121 14:09:35.168028 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:35 crc kubenswrapper[4897]: I1121 14:09:35.168066 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:35 crc kubenswrapper[4897]: I1121 14:09:35.168089 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:35Z","lastTransitionTime":"2025-11-21T14:09:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:35 crc kubenswrapper[4897]: I1121 14:09:35.270879 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:35 crc kubenswrapper[4897]: I1121 14:09:35.270913 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:35 crc kubenswrapper[4897]: I1121 14:09:35.270922 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:35 crc kubenswrapper[4897]: I1121 14:09:35.270936 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:35 crc kubenswrapper[4897]: I1121 14:09:35.270946 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:35Z","lastTransitionTime":"2025-11-21T14:09:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:35 crc kubenswrapper[4897]: I1121 14:09:35.373886 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:35 crc kubenswrapper[4897]: I1121 14:09:35.373968 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:35 crc kubenswrapper[4897]: I1121 14:09:35.373989 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:35 crc kubenswrapper[4897]: I1121 14:09:35.374095 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:35 crc kubenswrapper[4897]: I1121 14:09:35.374127 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:35Z","lastTransitionTime":"2025-11-21T14:09:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:35 crc kubenswrapper[4897]: I1121 14:09:35.476588 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:35 crc kubenswrapper[4897]: I1121 14:09:35.476628 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:35 crc kubenswrapper[4897]: I1121 14:09:35.476637 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:35 crc kubenswrapper[4897]: I1121 14:09:35.476651 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:35 crc kubenswrapper[4897]: I1121 14:09:35.476667 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:35Z","lastTransitionTime":"2025-11-21T14:09:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:35 crc kubenswrapper[4897]: I1121 14:09:35.579923 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:35 crc kubenswrapper[4897]: I1121 14:09:35.579991 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:35 crc kubenswrapper[4897]: I1121 14:09:35.580009 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:35 crc kubenswrapper[4897]: I1121 14:09:35.580035 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:35 crc kubenswrapper[4897]: I1121 14:09:35.580053 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:35Z","lastTransitionTime":"2025-11-21T14:09:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:35 crc kubenswrapper[4897]: I1121 14:09:35.682620 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:35 crc kubenswrapper[4897]: I1121 14:09:35.682688 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:35 crc kubenswrapper[4897]: I1121 14:09:35.682710 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:35 crc kubenswrapper[4897]: I1121 14:09:35.682743 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:35 crc kubenswrapper[4897]: I1121 14:09:35.682771 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:35Z","lastTransitionTime":"2025-11-21T14:09:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:35 crc kubenswrapper[4897]: I1121 14:09:35.742497 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:35 crc kubenswrapper[4897]: I1121 14:09:35.742565 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:35 crc kubenswrapper[4897]: I1121 14:09:35.742575 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:35 crc kubenswrapper[4897]: I1121 14:09:35.742593 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:35 crc kubenswrapper[4897]: I1121 14:09:35.742607 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:35Z","lastTransitionTime":"2025-11-21T14:09:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:35 crc kubenswrapper[4897]: E1121 14:09:35.755567 4897 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8b8bdf9e-0fef-41ac-98ad-710dc2137183\\\",\\\"systemUUID\\\":\\\"8a210af6-a71f-4ea9-90a3-aac32fb868b7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:35Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:35 crc kubenswrapper[4897]: I1121 14:09:35.759581 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:35 crc kubenswrapper[4897]: I1121 14:09:35.759638 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:35 crc kubenswrapper[4897]: I1121 14:09:35.759647 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:35 crc kubenswrapper[4897]: I1121 14:09:35.759666 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:35 crc kubenswrapper[4897]: I1121 14:09:35.759678 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:35Z","lastTransitionTime":"2025-11-21T14:09:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:35 crc kubenswrapper[4897]: E1121 14:09:35.772605 4897 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8b8bdf9e-0fef-41ac-98ad-710dc2137183\\\",\\\"systemUUID\\\":\\\"8a210af6-a71f-4ea9-90a3-aac32fb868b7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:35Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:35 crc kubenswrapper[4897]: I1121 14:09:35.777607 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:35 crc kubenswrapper[4897]: I1121 14:09:35.777661 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:35 crc kubenswrapper[4897]: I1121 14:09:35.777678 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:35 crc kubenswrapper[4897]: I1121 14:09:35.777704 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:35 crc kubenswrapper[4897]: I1121 14:09:35.777726 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:35Z","lastTransitionTime":"2025-11-21T14:09:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:35 crc kubenswrapper[4897]: E1121 14:09:35.794485 4897 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8b8bdf9e-0fef-41ac-98ad-710dc2137183\\\",\\\"systemUUID\\\":\\\"8a210af6-a71f-4ea9-90a3-aac32fb868b7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:35Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:35 crc kubenswrapper[4897]: I1121 14:09:35.799200 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:35 crc kubenswrapper[4897]: I1121 14:09:35.799261 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:35 crc kubenswrapper[4897]: I1121 14:09:35.799286 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:35 crc kubenswrapper[4897]: I1121 14:09:35.799309 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:35 crc kubenswrapper[4897]: I1121 14:09:35.799320 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:35Z","lastTransitionTime":"2025-11-21T14:09:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:35 crc kubenswrapper[4897]: E1121 14:09:35.814326 4897 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8b8bdf9e-0fef-41ac-98ad-710dc2137183\\\",\\\"systemUUID\\\":\\\"8a210af6-a71f-4ea9-90a3-aac32fb868b7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:35Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:35 crc kubenswrapper[4897]: I1121 14:09:35.818973 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:35 crc kubenswrapper[4897]: I1121 14:09:35.819008 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:35 crc kubenswrapper[4897]: I1121 14:09:35.819017 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:35 crc kubenswrapper[4897]: I1121 14:09:35.819030 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:35 crc kubenswrapper[4897]: I1121 14:09:35.819040 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:35Z","lastTransitionTime":"2025-11-21T14:09:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:35 crc kubenswrapper[4897]: E1121 14:09:35.830882 4897 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8b8bdf9e-0fef-41ac-98ad-710dc2137183\\\",\\\"systemUUID\\\":\\\"8a210af6-a71f-4ea9-90a3-aac32fb868b7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:35Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:35 crc kubenswrapper[4897]: E1121 14:09:35.831035 4897 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 21 14:09:35 crc kubenswrapper[4897]: I1121 14:09:35.832812 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:35 crc kubenswrapper[4897]: I1121 14:09:35.832852 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:35 crc kubenswrapper[4897]: I1121 14:09:35.832864 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:35 crc kubenswrapper[4897]: I1121 14:09:35.832880 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:35 crc kubenswrapper[4897]: I1121 14:09:35.832890 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:35Z","lastTransitionTime":"2025-11-21T14:09:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:35 crc kubenswrapper[4897]: I1121 14:09:35.935730 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:35 crc kubenswrapper[4897]: I1121 14:09:35.935787 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:35 crc kubenswrapper[4897]: I1121 14:09:35.935802 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:35 crc kubenswrapper[4897]: I1121 14:09:35.935821 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:35 crc kubenswrapper[4897]: I1121 14:09:35.935834 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:35Z","lastTransitionTime":"2025-11-21T14:09:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:36 crc kubenswrapper[4897]: I1121 14:09:36.037769 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:36 crc kubenswrapper[4897]: I1121 14:09:36.037815 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:36 crc kubenswrapper[4897]: I1121 14:09:36.037828 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:36 crc kubenswrapper[4897]: I1121 14:09:36.037846 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:36 crc kubenswrapper[4897]: I1121 14:09:36.037888 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:36Z","lastTransitionTime":"2025-11-21T14:09:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:36 crc kubenswrapper[4897]: I1121 14:09:36.140231 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:36 crc kubenswrapper[4897]: I1121 14:09:36.140274 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:36 crc kubenswrapper[4897]: I1121 14:09:36.140289 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:36 crc kubenswrapper[4897]: I1121 14:09:36.140305 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:36 crc kubenswrapper[4897]: I1121 14:09:36.140316 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:36Z","lastTransitionTime":"2025-11-21T14:09:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:36 crc kubenswrapper[4897]: I1121 14:09:36.242636 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:36 crc kubenswrapper[4897]: I1121 14:09:36.242733 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:36 crc kubenswrapper[4897]: I1121 14:09:36.242749 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:36 crc kubenswrapper[4897]: I1121 14:09:36.242772 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:36 crc kubenswrapper[4897]: I1121 14:09:36.242785 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:36Z","lastTransitionTime":"2025-11-21T14:09:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:36 crc kubenswrapper[4897]: I1121 14:09:36.344896 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:36 crc kubenswrapper[4897]: I1121 14:09:36.344956 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:36 crc kubenswrapper[4897]: I1121 14:09:36.344979 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:36 crc kubenswrapper[4897]: I1121 14:09:36.345001 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:36 crc kubenswrapper[4897]: I1121 14:09:36.345017 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:36Z","lastTransitionTime":"2025-11-21T14:09:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:36 crc kubenswrapper[4897]: I1121 14:09:36.447213 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:36 crc kubenswrapper[4897]: I1121 14:09:36.447276 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:36 crc kubenswrapper[4897]: I1121 14:09:36.447292 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:36 crc kubenswrapper[4897]: I1121 14:09:36.447314 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:36 crc kubenswrapper[4897]: I1121 14:09:36.447330 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:36Z","lastTransitionTime":"2025-11-21T14:09:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:36 crc kubenswrapper[4897]: I1121 14:09:36.550175 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:36 crc kubenswrapper[4897]: I1121 14:09:36.550225 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:36 crc kubenswrapper[4897]: I1121 14:09:36.550237 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:36 crc kubenswrapper[4897]: I1121 14:09:36.550257 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:36 crc kubenswrapper[4897]: I1121 14:09:36.550270 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:36Z","lastTransitionTime":"2025-11-21T14:09:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:36 crc kubenswrapper[4897]: I1121 14:09:36.653454 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:36 crc kubenswrapper[4897]: I1121 14:09:36.653493 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:36 crc kubenswrapper[4897]: I1121 14:09:36.653528 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:36 crc kubenswrapper[4897]: I1121 14:09:36.653544 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:36 crc kubenswrapper[4897]: I1121 14:09:36.653554 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:36Z","lastTransitionTime":"2025-11-21T14:09:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:36 crc kubenswrapper[4897]: I1121 14:09:36.756483 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:36 crc kubenswrapper[4897]: I1121 14:09:36.756557 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:36 crc kubenswrapper[4897]: I1121 14:09:36.756574 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:36 crc kubenswrapper[4897]: I1121 14:09:36.756600 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:36 crc kubenswrapper[4897]: I1121 14:09:36.756619 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:36Z","lastTransitionTime":"2025-11-21T14:09:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:36 crc kubenswrapper[4897]: I1121 14:09:36.859568 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:36 crc kubenswrapper[4897]: I1121 14:09:36.859637 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:36 crc kubenswrapper[4897]: I1121 14:09:36.859654 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:36 crc kubenswrapper[4897]: I1121 14:09:36.859681 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:36 crc kubenswrapper[4897]: I1121 14:09:36.859704 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:36Z","lastTransitionTime":"2025-11-21T14:09:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:36 crc kubenswrapper[4897]: I1121 14:09:36.962827 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:36 crc kubenswrapper[4897]: I1121 14:09:36.962892 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:36 crc kubenswrapper[4897]: I1121 14:09:36.962908 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:36 crc kubenswrapper[4897]: I1121 14:09:36.962935 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:36 crc kubenswrapper[4897]: I1121 14:09:36.962959 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:36Z","lastTransitionTime":"2025-11-21T14:09:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:37 crc kubenswrapper[4897]: I1121 14:09:37.064934 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:37 crc kubenswrapper[4897]: I1121 14:09:37.064970 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:37 crc kubenswrapper[4897]: I1121 14:09:37.064983 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:37 crc kubenswrapper[4897]: I1121 14:09:37.065003 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:37 crc kubenswrapper[4897]: I1121 14:09:37.065016 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:37Z","lastTransitionTime":"2025-11-21T14:09:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:37 crc kubenswrapper[4897]: I1121 14:09:37.088647 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 21 14:09:37 crc kubenswrapper[4897]: I1121 14:09:37.088698 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs4c8" Nov 21 14:09:37 crc kubenswrapper[4897]: I1121 14:09:37.088690 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 21 14:09:37 crc kubenswrapper[4897]: I1121 14:09:37.088727 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 21 14:09:37 crc kubenswrapper[4897]: E1121 14:09:37.088762 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 21 14:09:37 crc kubenswrapper[4897]: E1121 14:09:37.088939 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 21 14:09:37 crc kubenswrapper[4897]: E1121 14:09:37.088993 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs4c8" podUID="01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd" Nov 21 14:09:37 crc kubenswrapper[4897]: E1121 14:09:37.089568 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 21 14:09:37 crc kubenswrapper[4897]: I1121 14:09:37.090324 4897 scope.go:117] "RemoveContainer" containerID="814079abfa7d1b029825da6c6c5f1e57d439ff21ccb626c5d4e52e41d059ad3a" Nov 21 14:09:37 crc kubenswrapper[4897]: I1121 14:09:37.108377 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:37Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:37 crc kubenswrapper[4897]: I1121 14:09:37.121209 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6ljgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f3b70b5-9294-4f69-9528-500d28f34c89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://527cf28abdeef1b42d9c19fb88e3171dc05c4516c3a542a461b5b1daa85c6d25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78jhb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6ljgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:37Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:37 crc kubenswrapper[4897]: I1121 14:09:37.136278 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x5pws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc9e717e-af03-48bf-b005-62b3d084c6fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f968eee1d8a1888229e23abda59fd58fa5e86f02ca96ff80fbe093904b27e8df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9c4f5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0f99b9213902c003e6ca754134d2482404ace39c184fb3b431d2b600481d5b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9c4f5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-x5pws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:37Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:37 crc kubenswrapper[4897]: I1121 14:09:37.153945 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6cd330fd-e8a6-49d5-bfc9-5f6122ccd49e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1bd64b9b938bc80284bcee54141e90cad613d339c4f27ef2934b29bc2a318916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://936f59996d26e42099dcd055f87fff9e3ac4283b57e1956fe97463f2334152d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca74b28a069dc2e609b0779164c3aa470a15ed9848af0dd5fc3479aaf750157b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://270c53b9a342f26be077a01cd4e87e9ccd0eb5d2c84208ae38cbf720463245e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66c521c26d541bbddb064dd9710f1a7b41090d2f47343977922272dfc6dab7af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6eb2ac410b49098e5b3603de5e6bf5ed4484fab0b300021d0bed2e5ed8a42527\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6eb2ac410b49098e5b3603de5e6bf5ed4484fab0b300021d0bed2e5ed8a42527\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03385e0c78e4806f60c160227ef4905a11003ed3d70f0e5c0dd0a477979d1d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03385e0c78e4806f60c160227ef4905a11003ed3d70f0e5c0dd0a477979d1d4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://50aea2fa8b3b141b4330ddce2c8e0715230c6f06dab64b3abfd98a3067224bec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50aea2fa8b3b141b4330ddce2c8e0715230c6f06dab64b3abfd98a3067224bec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:37Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:37 crc kubenswrapper[4897]: I1121 14:09:37.166798 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:37 crc kubenswrapper[4897]: I1121 14:09:37.166843 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:37 crc kubenswrapper[4897]: I1121 14:09:37.166858 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:37 crc kubenswrapper[4897]: I1121 14:09:37.166880 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:37 crc kubenswrapper[4897]: I1121 14:09:37.166891 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:37Z","lastTransitionTime":"2025-11-21T14:09:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:37 crc kubenswrapper[4897]: I1121 14:09:37.170679 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a6c0a99-565b-41f9-a8b5-c8990a0c8eb3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://737553dd1d16e2019ea89a376e3a868f9792ed5239aa7d6cc40a05d3d4f286ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6682a8886d2b9bce069cf9261d4701ef6907f19574fd035bd89bd05626023e5c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de079956e3d8de0101f3f87dcc120ef23876c4972382639e419b3f34a38cdfca\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://469b5be7467d4a58d3776ec0a967ed911be31b770b49965da4f5d2fb264d8a3d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1907bfbad7eb34da642e30d6aa81820aab457f92347333a364ac12396b0e9dae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"le observer\\\\nW1121 14:09:07.994401 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1121 14:09:07.994601 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1121 14:09:07.995290 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1461162674/tls.crt::/tmp/serving-cert-1461162674/tls.key\\\\\\\"\\\\nI1121 14:09:08.162382 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1121 14:09:08.165301 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1121 14:09:08.165318 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1121 14:09:08.165343 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1121 14:09:08.165348 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1121 14:09:08.172015 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1121 14:09:08.172083 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1121 14:09:08.172089 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1121 14:09:08.172094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1121 14:09:08.172098 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1121 14:09:08.172102 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1121 14:09:08.172106 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1121 14:09:08.172023 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1121 14:09:08.174214 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d78c87c440c7d8e30075f6143111b3f504fb99686484b8f19a9c8d61a55eccc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://101fb592fef069af009c613a1e9c9825024ed205498ab72334211ba90d6f7397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://101fb592fef069af009c613a1e9c9825024ed205498ab72334211ba90d6f7397\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:37Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:37 crc kubenswrapper[4897]: I1121 14:09:37.182139 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b78c993a-9966-4e84-85a1-b4f73aaf17f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce3a2602338a1c1e9380ee55f78360f3d9cf6ccc62678886dc5ae574a72e3d9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c256823f2fb49ecc121f6ef564cd86196f9b4a473e9bdea670173596ef14a823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c65f71f8746044f5d41687c670a0308cd09dc41a6531f9b5720e852b3d240756\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1dbe88e7e6ad63899e107ebb884a16edb9a42b21ae7c86d7fbfaf89785b0d36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:37Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:37 crc kubenswrapper[4897]: I1121 14:09:37.193880 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09877d9780caddd5f7f73e95fdf818a427ed645ed5997984113151f4b4759c13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:37Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:37 crc kubenswrapper[4897]: I1121 14:09:37.207967 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:37Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:37 crc kubenswrapper[4897]: I1121 14:09:37.221955 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c976d804a7ebeba2c879e920cbc389f96f7f5def4b4439e8c8980d5460ecdb65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:37Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:37 crc kubenswrapper[4897]: I1121 14:09:37.235365 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nl76g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a7d6101-d0d2-4828-b3ce-d01040c04640\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac58b6ef1bd08e5f950260c18120c72e50b90a0de182c6ab0a5dbbbe45cf7328\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e56402f62821baf335f378084f4ed7c21b60ccbafe148419a0a9a403f106ad19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e56402f62821baf335f378084f4ed7c21b60ccbafe148419a0a9a403f106ad19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6022ff28e83dafcd846a7f03b090262aa3229a0da89f1984e4b6a054a04ac24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6022ff28e83dafcd846a7f03b090262aa3229a0da89f1984e4b6a054a04ac24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89de56e32b796da11bf89dd598d04fbf2884745ddf0f5d87cccfcaf1c001d8d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89de56e32b796da11bf89dd598d04fbf2884745ddf0f5d87cccfcaf1c001d8d6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2dff3cacaba7f9238f482e67b670ea775e0f96621434b99f4e76c2dfdde0321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2dff3cacaba7f9238f482e67b670ea775e0f96621434b99f4e76c2dfdde0321\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a153389a4137ee580e5118f3248400460f9d014c992a956f508815755f583afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a153389a4137ee580e5118f3248400460f9d014c992a956f508815755f583afb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a8c051ad1c1f2373f79eb0781fa6db489aeea3d52ed6c0bb43eec11520ec2b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a8c051ad1c1f2373f79eb0781fa6db489aeea3d52ed6c0bb43eec11520ec2b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nl76g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:37Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:37 crc kubenswrapper[4897]: I1121 14:09:37.247022 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48c6b1ea7497b33c00b956e400426c1d4f3c6c66e0225133434460800eef9314\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ea6068a6b6f8123b402de354e5374a0f02a7191a922d1718ab38f8898f47c52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:37Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:37 crc kubenswrapper[4897]: I1121 14:09:37.258596 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4qwqj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f46e2ba8-9ee0-406d-a668-2ba424618b69\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f43aa89d4d9defcb50d0dc6f90293b5f426b9cfc398b780c2e10f6a5c9f1e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbsgb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4qwqj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:37Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:37 crc kubenswrapper[4897]: I1121 14:09:37.269158 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:37 crc kubenswrapper[4897]: I1121 14:09:37.269217 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:37 crc kubenswrapper[4897]: I1121 14:09:37.269235 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:37 crc kubenswrapper[4897]: I1121 14:09:37.269271 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:37 crc kubenswrapper[4897]: I1121 14:09:37.269289 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:37Z","lastTransitionTime":"2025-11-21T14:09:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:37 crc kubenswrapper[4897]: I1121 14:09:37.271040 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-rs4c8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7xkdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7xkdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:23Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-rs4c8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:37Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:37 crc kubenswrapper[4897]: I1121 14:09:37.282817 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7qfmq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"223b7002-3e67-4848-b1fb-db2921deb64e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381a46c6290da2b36e572f46c16f1a50aa3e25359bcfbc02c7a5a18e1484860e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2m27\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7qfmq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:37Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:37 crc kubenswrapper[4897]: I1121 14:09:37.295003 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:37Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:37 crc kubenswrapper[4897]: I1121 14:09:37.306034 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7670227-d280-4847-b882-754429f56b0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://961961696b8750c248598f4cda032b18395e5c802275dcfe6d3e9f7ef3510f1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5854028821e6a360a259427b68e7bb4c2a7220e0c68f4cc91639b2ece0540a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-krv5b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:37Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:37 crc kubenswrapper[4897]: I1121 14:09:37.325348 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"222f65f2-0b82-4760-bcf7-779244720f01\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92f1d932a83d65ba6a9a906cc489ff51e76547d07d42656630c59421f833c4db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccd51aa0faf3b1327f207b355cddf86a7353c0051f7e5ce0396f3c4bc67e3b16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8971c38daad45bda3d61545f52120dad40992dbf2e1d6e1e776500ed58d73f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfee19f82e35869a641318a0c45bc7a2ae71d80039f5790e60e277be15c9d3ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec279f2f558a8ac653006321b18cbaf13185876fe303c5cc1c65988aef09edbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de095f3c760b13d5197f624eac3806a1f4186577f2b95eb522532d453eb266e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://814079abfa7d1b029825da6c6c5f1e57d439ff21ccb626c5d4e52e41d059ad3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://814079abfa7d1b029825da6c6c5f1e57d439ff21ccb626c5d4e52e41d059ad3a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-21T14:09:23Z\\\",\\\"message\\\":\\\"luster\\\\\\\", UUID:\\\\\\\"63b1440a-0908-4cab-8799-012fa1cf0b07\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-operator-lifecycle-manager/olm-operator-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-operator-lifecycle-manager/olm-operator-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-operator-lifecycle-manager/olm-operator-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.168\\\\\\\", Port:8443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1121 14:09:22.444620 6334 ovnkube.go:599] Stopped ovnkube\\\\nI1121 14:09:22.444715 6334 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1121 14:09:22.444771 6334 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:21Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-tpfb6_openshift-ovn-kubernetes(222f65f2-0b82-4760-bcf7-779244720f01)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b486a86e0e97ab409e590f30a1bf5cb9f584a7c67f96ff1236e42ca4f8675b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7877db6f5c2bf0eb31218c64d67bca48f921ff7ac62b4e495f29024c086bf738\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7877db6f5c2bf0eb31218c64d67bca48f921ff7ac62b4e495f29024c086bf738\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tpfb6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:37Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:37 crc kubenswrapper[4897]: I1121 14:09:37.372240 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:37 crc kubenswrapper[4897]: I1121 14:09:37.372278 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:37 crc kubenswrapper[4897]: I1121 14:09:37.372287 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:37 crc kubenswrapper[4897]: I1121 14:09:37.372304 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:37 crc kubenswrapper[4897]: I1121 14:09:37.372313 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:37Z","lastTransitionTime":"2025-11-21T14:09:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:37 crc kubenswrapper[4897]: I1121 14:09:37.474141 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:37 crc kubenswrapper[4897]: I1121 14:09:37.474188 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:37 crc kubenswrapper[4897]: I1121 14:09:37.474198 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:37 crc kubenswrapper[4897]: I1121 14:09:37.474216 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:37 crc kubenswrapper[4897]: I1121 14:09:37.474228 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:37Z","lastTransitionTime":"2025-11-21T14:09:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:37 crc kubenswrapper[4897]: I1121 14:09:37.483748 4897 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tpfb6_222f65f2-0b82-4760-bcf7-779244720f01/ovnkube-controller/1.log" Nov 21 14:09:37 crc kubenswrapper[4897]: I1121 14:09:37.485973 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" event={"ID":"222f65f2-0b82-4760-bcf7-779244720f01","Type":"ContainerStarted","Data":"9e285a47c1193b01d8cd3cb16ca750bea2f24d3ad941dd438f96d5203f9203de"} Nov 21 14:09:37 crc kubenswrapper[4897]: I1121 14:09:37.486437 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" Nov 21 14:09:37 crc kubenswrapper[4897]: I1121 14:09:37.502638 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:37Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:37 crc kubenswrapper[4897]: I1121 14:09:37.514037 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7670227-d280-4847-b882-754429f56b0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://961961696b8750c248598f4cda032b18395e5c802275dcfe6d3e9f7ef3510f1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5854028821e6a360a259427b68e7bb4c2a7220e0c68f4cc91639b2ece0540a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-krv5b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:37Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:37 crc kubenswrapper[4897]: I1121 14:09:37.535908 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"222f65f2-0b82-4760-bcf7-779244720f01\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92f1d932a83d65ba6a9a906cc489ff51e76547d07d42656630c59421f833c4db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccd51aa0faf3b1327f207b355cddf86a7353c0051f7e5ce0396f3c4bc67e3b16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8971c38daad45bda3d61545f52120dad40992dbf2e1d6e1e776500ed58d73f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfee19f82e35869a641318a0c45bc7a2ae71d80039f5790e60e277be15c9d3ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec279f2f558a8ac653006321b18cbaf13185876fe303c5cc1c65988aef09edbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de095f3c760b13d5197f624eac3806a1f4186577f2b95eb522532d453eb266e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e285a47c1193b01d8cd3cb16ca750bea2f24d3ad941dd438f96d5203f9203de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://814079abfa7d1b029825da6c6c5f1e57d439ff21ccb626c5d4e52e41d059ad3a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-21T14:09:23Z\\\",\\\"message\\\":\\\"luster\\\\\\\", UUID:\\\\\\\"63b1440a-0908-4cab-8799-012fa1cf0b07\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-operator-lifecycle-manager/olm-operator-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-operator-lifecycle-manager/olm-operator-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-operator-lifecycle-manager/olm-operator-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.168\\\\\\\", Port:8443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1121 14:09:22.444620 6334 ovnkube.go:599] Stopped ovnkube\\\\nI1121 14:09:22.444715 6334 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1121 14:09:22.444771 6334 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:21Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b486a86e0e97ab409e590f30a1bf5cb9f584a7c67f96ff1236e42ca4f8675b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7877db6f5c2bf0eb31218c64d67bca48f921ff7ac62b4e495f29024c086bf738\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7877db6f5c2bf0eb31218c64d67bca48f921ff7ac62b4e495f29024c086bf738\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tpfb6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:37Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:37 crc kubenswrapper[4897]: I1121 14:09:37.549271 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7qfmq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"223b7002-3e67-4848-b1fb-db2921deb64e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381a46c6290da2b36e572f46c16f1a50aa3e25359bcfbc02c7a5a18e1484860e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2m27\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7qfmq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:37Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:37 crc kubenswrapper[4897]: I1121 14:09:37.564319 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6ljgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f3b70b5-9294-4f69-9528-500d28f34c89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://527cf28abdeef1b42d9c19fb88e3171dc05c4516c3a542a461b5b1daa85c6d25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78jhb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6ljgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:37Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:37 crc kubenswrapper[4897]: I1121 14:09:37.576335 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:37 crc kubenswrapper[4897]: I1121 14:09:37.576381 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:37 crc kubenswrapper[4897]: I1121 14:09:37.576392 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:37 crc kubenswrapper[4897]: I1121 14:09:37.576409 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:37 crc kubenswrapper[4897]: I1121 14:09:37.576421 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:37Z","lastTransitionTime":"2025-11-21T14:09:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:37 crc kubenswrapper[4897]: I1121 14:09:37.578360 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x5pws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc9e717e-af03-48bf-b005-62b3d084c6fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f968eee1d8a1888229e23abda59fd58fa5e86f02ca96ff80fbe093904b27e8df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9c4f5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0f99b9213902c003e6ca754134d2482404ace39c184fb3b431d2b600481d5b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9c4f5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-x5pws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:37Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:37 crc kubenswrapper[4897]: I1121 14:09:37.600874 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6cd330fd-e8a6-49d5-bfc9-5f6122ccd49e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1bd64b9b938bc80284bcee54141e90cad613d339c4f27ef2934b29bc2a318916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://936f59996d26e42099dcd055f87fff9e3ac4283b57e1956fe97463f2334152d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca74b28a069dc2e609b0779164c3aa470a15ed9848af0dd5fc3479aaf750157b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://270c53b9a342f26be077a01cd4e87e9ccd0eb5d2c84208ae38cbf720463245e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66c521c26d541bbddb064dd9710f1a7b41090d2f47343977922272dfc6dab7af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6eb2ac410b49098e5b3603de5e6bf5ed4484fab0b300021d0bed2e5ed8a42527\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6eb2ac410b49098e5b3603de5e6bf5ed4484fab0b300021d0bed2e5ed8a42527\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03385e0c78e4806f60c160227ef4905a11003ed3d70f0e5c0dd0a477979d1d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03385e0c78e4806f60c160227ef4905a11003ed3d70f0e5c0dd0a477979d1d4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://50aea2fa8b3b141b4330ddce2c8e0715230c6f06dab64b3abfd98a3067224bec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50aea2fa8b3b141b4330ddce2c8e0715230c6f06dab64b3abfd98a3067224bec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:37Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:37 crc kubenswrapper[4897]: I1121 14:09:37.616499 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a6c0a99-565b-41f9-a8b5-c8990a0c8eb3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://737553dd1d16e2019ea89a376e3a868f9792ed5239aa7d6cc40a05d3d4f286ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6682a8886d2b9bce069cf9261d4701ef6907f19574fd035bd89bd05626023e5c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de079956e3d8de0101f3f87dcc120ef23876c4972382639e419b3f34a38cdfca\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://469b5be7467d4a58d3776ec0a967ed911be31b770b49965da4f5d2fb264d8a3d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1907bfbad7eb34da642e30d6aa81820aab457f92347333a364ac12396b0e9dae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"le observer\\\\nW1121 14:09:07.994401 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1121 14:09:07.994601 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1121 14:09:07.995290 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1461162674/tls.crt::/tmp/serving-cert-1461162674/tls.key\\\\\\\"\\\\nI1121 14:09:08.162382 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1121 14:09:08.165301 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1121 14:09:08.165318 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1121 14:09:08.165343 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1121 14:09:08.165348 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1121 14:09:08.172015 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1121 14:09:08.172083 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1121 14:09:08.172089 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1121 14:09:08.172094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1121 14:09:08.172098 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1121 14:09:08.172102 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1121 14:09:08.172106 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1121 14:09:08.172023 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1121 14:09:08.174214 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d78c87c440c7d8e30075f6143111b3f504fb99686484b8f19a9c8d61a55eccc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://101fb592fef069af009c613a1e9c9825024ed205498ab72334211ba90d6f7397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://101fb592fef069af009c613a1e9c9825024ed205498ab72334211ba90d6f7397\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:37Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:37 crc kubenswrapper[4897]: I1121 14:09:37.631491 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b78c993a-9966-4e84-85a1-b4f73aaf17f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce3a2602338a1c1e9380ee55f78360f3d9cf6ccc62678886dc5ae574a72e3d9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c256823f2fb49ecc121f6ef564cd86196f9b4a473e9bdea670173596ef14a823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c65f71f8746044f5d41687c670a0308cd09dc41a6531f9b5720e852b3d240756\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1dbe88e7e6ad63899e107ebb884a16edb9a42b21ae7c86d7fbfaf89785b0d36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:37Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:37 crc kubenswrapper[4897]: I1121 14:09:37.644268 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09877d9780caddd5f7f73e95fdf818a427ed645ed5997984113151f4b4759c13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:37Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:37 crc kubenswrapper[4897]: I1121 14:09:37.663929 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:37Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:37 crc kubenswrapper[4897]: I1121 14:09:37.676794 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:37Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:37 crc kubenswrapper[4897]: I1121 14:09:37.678458 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:37 crc kubenswrapper[4897]: I1121 14:09:37.678535 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:37 crc kubenswrapper[4897]: I1121 14:09:37.678548 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:37 crc kubenswrapper[4897]: I1121 14:09:37.678568 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:37 crc kubenswrapper[4897]: I1121 14:09:37.678598 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:37Z","lastTransitionTime":"2025-11-21T14:09:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:37 crc kubenswrapper[4897]: I1121 14:09:37.689641 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c976d804a7ebeba2c879e920cbc389f96f7f5def4b4439e8c8980d5460ecdb65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:37Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:37 crc kubenswrapper[4897]: I1121 14:09:37.704622 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nl76g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a7d6101-d0d2-4828-b3ce-d01040c04640\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac58b6ef1bd08e5f950260c18120c72e50b90a0de182c6ab0a5dbbbe45cf7328\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e56402f62821baf335f378084f4ed7c21b60ccbafe148419a0a9a403f106ad19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e56402f62821baf335f378084f4ed7c21b60ccbafe148419a0a9a403f106ad19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6022ff28e83dafcd846a7f03b090262aa3229a0da89f1984e4b6a054a04ac24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6022ff28e83dafcd846a7f03b090262aa3229a0da89f1984e4b6a054a04ac24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89de56e32b796da11bf89dd598d04fbf2884745ddf0f5d87cccfcaf1c001d8d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89de56e32b796da11bf89dd598d04fbf2884745ddf0f5d87cccfcaf1c001d8d6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2dff3cacaba7f9238f482e67b670ea775e0f96621434b99f4e76c2dfdde0321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2dff3cacaba7f9238f482e67b670ea775e0f96621434b99f4e76c2dfdde0321\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a153389a4137ee580e5118f3248400460f9d014c992a956f508815755f583afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a153389a4137ee580e5118f3248400460f9d014c992a956f508815755f583afb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a8c051ad1c1f2373f79eb0781fa6db489aeea3d52ed6c0bb43eec11520ec2b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a8c051ad1c1f2373f79eb0781fa6db489aeea3d52ed6c0bb43eec11520ec2b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nl76g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:37Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:37 crc kubenswrapper[4897]: I1121 14:09:37.715827 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48c6b1ea7497b33c00b956e400426c1d4f3c6c66e0225133434460800eef9314\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ea6068a6b6f8123b402de354e5374a0f02a7191a922d1718ab38f8898f47c52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:37Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:37 crc kubenswrapper[4897]: I1121 14:09:37.724625 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4qwqj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f46e2ba8-9ee0-406d-a668-2ba424618b69\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f43aa89d4d9defcb50d0dc6f90293b5f426b9cfc398b780c2e10f6a5c9f1e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbsgb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4qwqj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:37Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:37 crc kubenswrapper[4897]: I1121 14:09:37.733463 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-rs4c8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7xkdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7xkdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:23Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-rs4c8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:37Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:37 crc kubenswrapper[4897]: I1121 14:09:37.780450 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:37 crc kubenswrapper[4897]: I1121 14:09:37.780496 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:37 crc kubenswrapper[4897]: I1121 14:09:37.780526 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:37 crc kubenswrapper[4897]: I1121 14:09:37.780544 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:37 crc kubenswrapper[4897]: I1121 14:09:37.780556 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:37Z","lastTransitionTime":"2025-11-21T14:09:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:37 crc kubenswrapper[4897]: I1121 14:09:37.882347 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:37 crc kubenswrapper[4897]: I1121 14:09:37.882384 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:37 crc kubenswrapper[4897]: I1121 14:09:37.882392 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:37 crc kubenswrapper[4897]: I1121 14:09:37.882406 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:37 crc kubenswrapper[4897]: I1121 14:09:37.882415 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:37Z","lastTransitionTime":"2025-11-21T14:09:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:37 crc kubenswrapper[4897]: I1121 14:09:37.985062 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:37 crc kubenswrapper[4897]: I1121 14:09:37.985118 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:37 crc kubenswrapper[4897]: I1121 14:09:37.985131 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:37 crc kubenswrapper[4897]: I1121 14:09:37.985150 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:37 crc kubenswrapper[4897]: I1121 14:09:37.985160 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:37Z","lastTransitionTime":"2025-11-21T14:09:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:38 crc kubenswrapper[4897]: I1121 14:09:38.087732 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:38 crc kubenswrapper[4897]: I1121 14:09:38.087782 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:38 crc kubenswrapper[4897]: I1121 14:09:38.087798 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:38 crc kubenswrapper[4897]: I1121 14:09:38.087821 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:38 crc kubenswrapper[4897]: I1121 14:09:38.087838 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:38Z","lastTransitionTime":"2025-11-21T14:09:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:38 crc kubenswrapper[4897]: I1121 14:09:38.189779 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:38 crc kubenswrapper[4897]: I1121 14:09:38.189818 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:38 crc kubenswrapper[4897]: I1121 14:09:38.189827 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:38 crc kubenswrapper[4897]: I1121 14:09:38.189843 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:38 crc kubenswrapper[4897]: I1121 14:09:38.189853 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:38Z","lastTransitionTime":"2025-11-21T14:09:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:38 crc kubenswrapper[4897]: I1121 14:09:38.292759 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:38 crc kubenswrapper[4897]: I1121 14:09:38.292821 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:38 crc kubenswrapper[4897]: I1121 14:09:38.292832 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:38 crc kubenswrapper[4897]: I1121 14:09:38.292855 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:38 crc kubenswrapper[4897]: I1121 14:09:38.292864 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:38Z","lastTransitionTime":"2025-11-21T14:09:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:38 crc kubenswrapper[4897]: I1121 14:09:38.395984 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:38 crc kubenswrapper[4897]: I1121 14:09:38.396030 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:38 crc kubenswrapper[4897]: I1121 14:09:38.396042 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:38 crc kubenswrapper[4897]: I1121 14:09:38.396058 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:38 crc kubenswrapper[4897]: I1121 14:09:38.396069 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:38Z","lastTransitionTime":"2025-11-21T14:09:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:38 crc kubenswrapper[4897]: I1121 14:09:38.490962 4897 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tpfb6_222f65f2-0b82-4760-bcf7-779244720f01/ovnkube-controller/2.log" Nov 21 14:09:38 crc kubenswrapper[4897]: I1121 14:09:38.491806 4897 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tpfb6_222f65f2-0b82-4760-bcf7-779244720f01/ovnkube-controller/1.log" Nov 21 14:09:38 crc kubenswrapper[4897]: I1121 14:09:38.494794 4897 generic.go:334] "Generic (PLEG): container finished" podID="222f65f2-0b82-4760-bcf7-779244720f01" containerID="9e285a47c1193b01d8cd3cb16ca750bea2f24d3ad941dd438f96d5203f9203de" exitCode=1 Nov 21 14:09:38 crc kubenswrapper[4897]: I1121 14:09:38.494848 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" event={"ID":"222f65f2-0b82-4760-bcf7-779244720f01","Type":"ContainerDied","Data":"9e285a47c1193b01d8cd3cb16ca750bea2f24d3ad941dd438f96d5203f9203de"} Nov 21 14:09:38 crc kubenswrapper[4897]: I1121 14:09:38.494911 4897 scope.go:117] "RemoveContainer" containerID="814079abfa7d1b029825da6c6c5f1e57d439ff21ccb626c5d4e52e41d059ad3a" Nov 21 14:09:38 crc kubenswrapper[4897]: I1121 14:09:38.495555 4897 scope.go:117] "RemoveContainer" containerID="9e285a47c1193b01d8cd3cb16ca750bea2f24d3ad941dd438f96d5203f9203de" Nov 21 14:09:38 crc kubenswrapper[4897]: E1121 14:09:38.495740 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-tpfb6_openshift-ovn-kubernetes(222f65f2-0b82-4760-bcf7-779244720f01)\"" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" podUID="222f65f2-0b82-4760-bcf7-779244720f01" Nov 21 14:09:38 crc kubenswrapper[4897]: I1121 14:09:38.500730 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:38 crc kubenswrapper[4897]: I1121 14:09:38.500767 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:38 crc kubenswrapper[4897]: I1121 14:09:38.500775 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:38 crc kubenswrapper[4897]: I1121 14:09:38.500792 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:38 crc kubenswrapper[4897]: I1121 14:09:38.500804 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:38Z","lastTransitionTime":"2025-11-21T14:09:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:38 crc kubenswrapper[4897]: I1121 14:09:38.512759 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:38Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:38 crc kubenswrapper[4897]: I1121 14:09:38.529189 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c976d804a7ebeba2c879e920cbc389f96f7f5def4b4439e8c8980d5460ecdb65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:38Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:38 crc kubenswrapper[4897]: I1121 14:09:38.547394 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nl76g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a7d6101-d0d2-4828-b3ce-d01040c04640\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac58b6ef1bd08e5f950260c18120c72e50b90a0de182c6ab0a5dbbbe45cf7328\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e56402f62821baf335f378084f4ed7c21b60ccbafe148419a0a9a403f106ad19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e56402f62821baf335f378084f4ed7c21b60ccbafe148419a0a9a403f106ad19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6022ff28e83dafcd846a7f03b090262aa3229a0da89f1984e4b6a054a04ac24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6022ff28e83dafcd846a7f03b090262aa3229a0da89f1984e4b6a054a04ac24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89de56e32b796da11bf89dd598d04fbf2884745ddf0f5d87cccfcaf1c001d8d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89de56e32b796da11bf89dd598d04fbf2884745ddf0f5d87cccfcaf1c001d8d6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2dff3cacaba7f9238f482e67b670ea775e0f96621434b99f4e76c2dfdde0321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2dff3cacaba7f9238f482e67b670ea775e0f96621434b99f4e76c2dfdde0321\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a153389a4137ee580e5118f3248400460f9d014c992a956f508815755f583afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a153389a4137ee580e5118f3248400460f9d014c992a956f508815755f583afb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a8c051ad1c1f2373f79eb0781fa6db489aeea3d52ed6c0bb43eec11520ec2b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a8c051ad1c1f2373f79eb0781fa6db489aeea3d52ed6c0bb43eec11520ec2b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nl76g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:38Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:38 crc kubenswrapper[4897]: I1121 14:09:38.562316 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48c6b1ea7497b33c00b956e400426c1d4f3c6c66e0225133434460800eef9314\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ea6068a6b6f8123b402de354e5374a0f02a7191a922d1718ab38f8898f47c52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:38Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:38 crc kubenswrapper[4897]: I1121 14:09:38.571955 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4qwqj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f46e2ba8-9ee0-406d-a668-2ba424618b69\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f43aa89d4d9defcb50d0dc6f90293b5f426b9cfc398b780c2e10f6a5c9f1e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbsgb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4qwqj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:38Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:38 crc kubenswrapper[4897]: I1121 14:09:38.582639 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-rs4c8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7xkdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7xkdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:23Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-rs4c8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:38Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:38 crc kubenswrapper[4897]: I1121 14:09:38.592462 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7qfmq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"223b7002-3e67-4848-b1fb-db2921deb64e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381a46c6290da2b36e572f46c16f1a50aa3e25359bcfbc02c7a5a18e1484860e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2m27\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7qfmq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:38Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:38 crc kubenswrapper[4897]: I1121 14:09:38.603054 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:38 crc kubenswrapper[4897]: I1121 14:09:38.603096 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:38 crc kubenswrapper[4897]: I1121 14:09:38.603108 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:38 crc kubenswrapper[4897]: I1121 14:09:38.603126 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:38 crc kubenswrapper[4897]: I1121 14:09:38.603139 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:38Z","lastTransitionTime":"2025-11-21T14:09:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:38 crc kubenswrapper[4897]: I1121 14:09:38.604447 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:38Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:38 crc kubenswrapper[4897]: I1121 14:09:38.616250 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7670227-d280-4847-b882-754429f56b0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://961961696b8750c248598f4cda032b18395e5c802275dcfe6d3e9f7ef3510f1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5854028821e6a360a259427b68e7bb4c2a7220e0c68f4cc91639b2ece0540a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-krv5b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:38Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:38 crc kubenswrapper[4897]: I1121 14:09:38.636025 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"222f65f2-0b82-4760-bcf7-779244720f01\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92f1d932a83d65ba6a9a906cc489ff51e76547d07d42656630c59421f833c4db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccd51aa0faf3b1327f207b355cddf86a7353c0051f7e5ce0396f3c4bc67e3b16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8971c38daad45bda3d61545f52120dad40992dbf2e1d6e1e776500ed58d73f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfee19f82e35869a641318a0c45bc7a2ae71d80039f5790e60e277be15c9d3ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec279f2f558a8ac653006321b18cbaf13185876fe303c5cc1c65988aef09edbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de095f3c760b13d5197f624eac3806a1f4186577f2b95eb522532d453eb266e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e285a47c1193b01d8cd3cb16ca750bea2f24d3ad941dd438f96d5203f9203de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://814079abfa7d1b029825da6c6c5f1e57d439ff21ccb626c5d4e52e41d059ad3a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-21T14:09:23Z\\\",\\\"message\\\":\\\"luster\\\\\\\", UUID:\\\\\\\"63b1440a-0908-4cab-8799-012fa1cf0b07\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-operator-lifecycle-manager/olm-operator-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-operator-lifecycle-manager/olm-operator-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-operator-lifecycle-manager/olm-operator-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.168\\\\\\\", Port:8443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1121 14:09:22.444620 6334 ovnkube.go:599] Stopped ovnkube\\\\nI1121 14:09:22.444715 6334 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1121 14:09:22.444771 6334 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:21Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e285a47c1193b01d8cd3cb16ca750bea2f24d3ad941dd438f96d5203f9203de\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-21T14:09:38Z\\\",\\\"message\\\":\\\":776] Recording success event on pod openshift-etcd/etcd-crc\\\\nI1121 14:09:38.005810 6569 default_network_controller.go:776] Recording success event on pod openshift-image-registry/node-ca-7qfmq\\\\nI1121 14:09:38.005827 6569 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-additional-cni-plugins-nl76g\\\\nI1121 14:09:38.005832 6569 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-additional-cni-plugins-nl76g\\\\nI1121 14:09:38.005837 6569 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-additional-cni-plugins-nl76g in node crc\\\\nI1121 14:09:38.005842 6569 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-additional-cni-plugins-nl76g after 0 failed attempt(s)\\\\nI1121 14:09:38.005846 6569 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-additional-cni-plugins-nl76g\\\\nI1121 14:09:38.005834 6569 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-operator-lifecycle-manager/package-server-manager-metrics]} name:Service_openshift-operator-lifecycle-manager/package-server-manager-metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b486a86e0e97ab409e590f30a1bf5cb9f584a7c67f96ff1236e42ca4f8675b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7877db6f5c2bf0eb31218c64d67bca48f921ff7ac62b4e495f29024c086bf738\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7877db6f5c2bf0eb31218c64d67bca48f921ff7ac62b4e495f29024c086bf738\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tpfb6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:38Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:38 crc kubenswrapper[4897]: I1121 14:09:38.649191 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:38Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:38 crc kubenswrapper[4897]: I1121 14:09:38.660460 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6ljgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f3b70b5-9294-4f69-9528-500d28f34c89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://527cf28abdeef1b42d9c19fb88e3171dc05c4516c3a542a461b5b1daa85c6d25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78jhb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6ljgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:38Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:38 crc kubenswrapper[4897]: I1121 14:09:38.673416 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x5pws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc9e717e-af03-48bf-b005-62b3d084c6fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f968eee1d8a1888229e23abda59fd58fa5e86f02ca96ff80fbe093904b27e8df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9c4f5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0f99b9213902c003e6ca754134d2482404ace39c184fb3b431d2b600481d5b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9c4f5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-x5pws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:38Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:38 crc kubenswrapper[4897]: I1121 14:09:38.693752 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6cd330fd-e8a6-49d5-bfc9-5f6122ccd49e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1bd64b9b938bc80284bcee54141e90cad613d339c4f27ef2934b29bc2a318916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://936f59996d26e42099dcd055f87fff9e3ac4283b57e1956fe97463f2334152d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca74b28a069dc2e609b0779164c3aa470a15ed9848af0dd5fc3479aaf750157b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://270c53b9a342f26be077a01cd4e87e9ccd0eb5d2c84208ae38cbf720463245e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66c521c26d541bbddb064dd9710f1a7b41090d2f47343977922272dfc6dab7af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6eb2ac410b49098e5b3603de5e6bf5ed4484fab0b300021d0bed2e5ed8a42527\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6eb2ac410b49098e5b3603de5e6bf5ed4484fab0b300021d0bed2e5ed8a42527\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03385e0c78e4806f60c160227ef4905a11003ed3d70f0e5c0dd0a477979d1d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03385e0c78e4806f60c160227ef4905a11003ed3d70f0e5c0dd0a477979d1d4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://50aea2fa8b3b141b4330ddce2c8e0715230c6f06dab64b3abfd98a3067224bec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50aea2fa8b3b141b4330ddce2c8e0715230c6f06dab64b3abfd98a3067224bec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:38Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:38 crc kubenswrapper[4897]: I1121 14:09:38.706145 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:38 crc kubenswrapper[4897]: I1121 14:09:38.706192 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:38 crc kubenswrapper[4897]: I1121 14:09:38.706201 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:38 crc kubenswrapper[4897]: I1121 14:09:38.706217 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:38 crc kubenswrapper[4897]: I1121 14:09:38.706232 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:38Z","lastTransitionTime":"2025-11-21T14:09:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:38 crc kubenswrapper[4897]: I1121 14:09:38.710733 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a6c0a99-565b-41f9-a8b5-c8990a0c8eb3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://737553dd1d16e2019ea89a376e3a868f9792ed5239aa7d6cc40a05d3d4f286ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6682a8886d2b9bce069cf9261d4701ef6907f19574fd035bd89bd05626023e5c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de079956e3d8de0101f3f87dcc120ef23876c4972382639e419b3f34a38cdfca\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://469b5be7467d4a58d3776ec0a967ed911be31b770b49965da4f5d2fb264d8a3d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1907bfbad7eb34da642e30d6aa81820aab457f92347333a364ac12396b0e9dae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"le observer\\\\nW1121 14:09:07.994401 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1121 14:09:07.994601 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1121 14:09:07.995290 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1461162674/tls.crt::/tmp/serving-cert-1461162674/tls.key\\\\\\\"\\\\nI1121 14:09:08.162382 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1121 14:09:08.165301 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1121 14:09:08.165318 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1121 14:09:08.165343 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1121 14:09:08.165348 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1121 14:09:08.172015 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1121 14:09:08.172083 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1121 14:09:08.172089 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1121 14:09:08.172094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1121 14:09:08.172098 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1121 14:09:08.172102 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1121 14:09:08.172106 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1121 14:09:08.172023 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1121 14:09:08.174214 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d78c87c440c7d8e30075f6143111b3f504fb99686484b8f19a9c8d61a55eccc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://101fb592fef069af009c613a1e9c9825024ed205498ab72334211ba90d6f7397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://101fb592fef069af009c613a1e9c9825024ed205498ab72334211ba90d6f7397\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:38Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:38 crc kubenswrapper[4897]: I1121 14:09:38.724964 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b78c993a-9966-4e84-85a1-b4f73aaf17f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce3a2602338a1c1e9380ee55f78360f3d9cf6ccc62678886dc5ae574a72e3d9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c256823f2fb49ecc121f6ef564cd86196f9b4a473e9bdea670173596ef14a823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c65f71f8746044f5d41687c670a0308cd09dc41a6531f9b5720e852b3d240756\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1dbe88e7e6ad63899e107ebb884a16edb9a42b21ae7c86d7fbfaf89785b0d36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:38Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:38 crc kubenswrapper[4897]: I1121 14:09:38.737186 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09877d9780caddd5f7f73e95fdf818a427ed645ed5997984113151f4b4759c13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:38Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:38 crc kubenswrapper[4897]: I1121 14:09:38.808565 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:38 crc kubenswrapper[4897]: I1121 14:09:38.808609 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:38 crc kubenswrapper[4897]: I1121 14:09:38.808618 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:38 crc kubenswrapper[4897]: I1121 14:09:38.808637 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:38 crc kubenswrapper[4897]: I1121 14:09:38.808647 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:38Z","lastTransitionTime":"2025-11-21T14:09:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:38 crc kubenswrapper[4897]: I1121 14:09:38.910553 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:38 crc kubenswrapper[4897]: I1121 14:09:38.910596 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:38 crc kubenswrapper[4897]: I1121 14:09:38.910607 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:38 crc kubenswrapper[4897]: I1121 14:09:38.910623 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:38 crc kubenswrapper[4897]: I1121 14:09:38.910632 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:38Z","lastTransitionTime":"2025-11-21T14:09:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:39 crc kubenswrapper[4897]: I1121 14:09:39.013859 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:39 crc kubenswrapper[4897]: I1121 14:09:39.013909 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:39 crc kubenswrapper[4897]: I1121 14:09:39.013918 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:39 crc kubenswrapper[4897]: I1121 14:09:39.013937 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:39 crc kubenswrapper[4897]: I1121 14:09:39.013949 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:39Z","lastTransitionTime":"2025-11-21T14:09:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:39 crc kubenswrapper[4897]: I1121 14:09:39.088184 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs4c8" Nov 21 14:09:39 crc kubenswrapper[4897]: I1121 14:09:39.088245 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 21 14:09:39 crc kubenswrapper[4897]: E1121 14:09:39.088353 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs4c8" podUID="01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd" Nov 21 14:09:39 crc kubenswrapper[4897]: I1121 14:09:39.088258 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 21 14:09:39 crc kubenswrapper[4897]: E1121 14:09:39.088473 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 21 14:09:39 crc kubenswrapper[4897]: E1121 14:09:39.088681 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 21 14:09:39 crc kubenswrapper[4897]: I1121 14:09:39.088857 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 21 14:09:39 crc kubenswrapper[4897]: E1121 14:09:39.088935 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 21 14:09:39 crc kubenswrapper[4897]: I1121 14:09:39.116950 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:39 crc kubenswrapper[4897]: I1121 14:09:39.117000 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:39 crc kubenswrapper[4897]: I1121 14:09:39.117014 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:39 crc kubenswrapper[4897]: I1121 14:09:39.117080 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:39 crc kubenswrapper[4897]: I1121 14:09:39.117103 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:39Z","lastTransitionTime":"2025-11-21T14:09:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:39 crc kubenswrapper[4897]: I1121 14:09:39.219414 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:39 crc kubenswrapper[4897]: I1121 14:09:39.219461 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:39 crc kubenswrapper[4897]: I1121 14:09:39.219473 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:39 crc kubenswrapper[4897]: I1121 14:09:39.219489 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:39 crc kubenswrapper[4897]: I1121 14:09:39.219500 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:39Z","lastTransitionTime":"2025-11-21T14:09:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:39 crc kubenswrapper[4897]: I1121 14:09:39.322725 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:39 crc kubenswrapper[4897]: I1121 14:09:39.322762 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:39 crc kubenswrapper[4897]: I1121 14:09:39.322772 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:39 crc kubenswrapper[4897]: I1121 14:09:39.322789 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:39 crc kubenswrapper[4897]: I1121 14:09:39.322798 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:39Z","lastTransitionTime":"2025-11-21T14:09:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:39 crc kubenswrapper[4897]: I1121 14:09:39.330079 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 21 14:09:39 crc kubenswrapper[4897]: I1121 14:09:39.345026 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Nov 21 14:09:39 crc kubenswrapper[4897]: I1121 14:09:39.357381 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6cd330fd-e8a6-49d5-bfc9-5f6122ccd49e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1bd64b9b938bc80284bcee54141e90cad613d339c4f27ef2934b29bc2a318916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://936f59996d26e42099dcd055f87fff9e3ac4283b57e1956fe97463f2334152d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca74b28a069dc2e609b0779164c3aa470a15ed9848af0dd5fc3479aaf750157b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://270c53b9a342f26be077a01cd4e87e9ccd0eb5d2c84208ae38cbf720463245e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66c521c26d541bbddb064dd9710f1a7b41090d2f47343977922272dfc6dab7af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6eb2ac410b49098e5b3603de5e6bf5ed4484fab0b300021d0bed2e5ed8a42527\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6eb2ac410b49098e5b3603de5e6bf5ed4484fab0b300021d0bed2e5ed8a42527\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03385e0c78e4806f60c160227ef4905a11003ed3d70f0e5c0dd0a477979d1d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03385e0c78e4806f60c160227ef4905a11003ed3d70f0e5c0dd0a477979d1d4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://50aea2fa8b3b141b4330ddce2c8e0715230c6f06dab64b3abfd98a3067224bec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50aea2fa8b3b141b4330ddce2c8e0715230c6f06dab64b3abfd98a3067224bec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:39Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:39 crc kubenswrapper[4897]: I1121 14:09:39.375784 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a6c0a99-565b-41f9-a8b5-c8990a0c8eb3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://737553dd1d16e2019ea89a376e3a868f9792ed5239aa7d6cc40a05d3d4f286ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6682a8886d2b9bce069cf9261d4701ef6907f19574fd035bd89bd05626023e5c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de079956e3d8de0101f3f87dcc120ef23876c4972382639e419b3f34a38cdfca\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://469b5be7467d4a58d3776ec0a967ed911be31b770b49965da4f5d2fb264d8a3d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1907bfbad7eb34da642e30d6aa81820aab457f92347333a364ac12396b0e9dae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"le observer\\\\nW1121 14:09:07.994401 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1121 14:09:07.994601 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1121 14:09:07.995290 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1461162674/tls.crt::/tmp/serving-cert-1461162674/tls.key\\\\\\\"\\\\nI1121 14:09:08.162382 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1121 14:09:08.165301 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1121 14:09:08.165318 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1121 14:09:08.165343 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1121 14:09:08.165348 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1121 14:09:08.172015 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1121 14:09:08.172083 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1121 14:09:08.172089 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1121 14:09:08.172094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1121 14:09:08.172098 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1121 14:09:08.172102 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1121 14:09:08.172106 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1121 14:09:08.172023 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1121 14:09:08.174214 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d78c87c440c7d8e30075f6143111b3f504fb99686484b8f19a9c8d61a55eccc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://101fb592fef069af009c613a1e9c9825024ed205498ab72334211ba90d6f7397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://101fb592fef069af009c613a1e9c9825024ed205498ab72334211ba90d6f7397\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:39Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:39 crc kubenswrapper[4897]: I1121 14:09:39.391640 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b78c993a-9966-4e84-85a1-b4f73aaf17f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce3a2602338a1c1e9380ee55f78360f3d9cf6ccc62678886dc5ae574a72e3d9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c256823f2fb49ecc121f6ef564cd86196f9b4a473e9bdea670173596ef14a823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c65f71f8746044f5d41687c670a0308cd09dc41a6531f9b5720e852b3d240756\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1dbe88e7e6ad63899e107ebb884a16edb9a42b21ae7c86d7fbfaf89785b0d36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:39Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:39 crc kubenswrapper[4897]: I1121 14:09:39.408439 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09877d9780caddd5f7f73e95fdf818a427ed645ed5997984113151f4b4759c13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:39Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:39 crc kubenswrapper[4897]: I1121 14:09:39.425789 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:39Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:39 crc kubenswrapper[4897]: I1121 14:09:39.425892 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:39 crc kubenswrapper[4897]: I1121 14:09:39.425934 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:39 crc kubenswrapper[4897]: I1121 14:09:39.425946 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:39 crc kubenswrapper[4897]: I1121 14:09:39.425966 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:39 crc kubenswrapper[4897]: I1121 14:09:39.425979 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:39Z","lastTransitionTime":"2025-11-21T14:09:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:39 crc kubenswrapper[4897]: I1121 14:09:39.442764 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6ljgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f3b70b5-9294-4f69-9528-500d28f34c89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://527cf28abdeef1b42d9c19fb88e3171dc05c4516c3a542a461b5b1daa85c6d25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78jhb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6ljgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:39Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:39 crc kubenswrapper[4897]: I1121 14:09:39.454006 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x5pws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc9e717e-af03-48bf-b005-62b3d084c6fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f968eee1d8a1888229e23abda59fd58fa5e86f02ca96ff80fbe093904b27e8df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9c4f5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0f99b9213902c003e6ca754134d2482404ace39c184fb3b431d2b600481d5b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9c4f5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-x5pws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:39Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:39 crc kubenswrapper[4897]: I1121 14:09:39.469100 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:39Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:39 crc kubenswrapper[4897]: I1121 14:09:39.471045 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 21 14:09:39 crc kubenswrapper[4897]: I1121 14:09:39.482326 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c976d804a7ebeba2c879e920cbc389f96f7f5def4b4439e8c8980d5460ecdb65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:39Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:39 crc kubenswrapper[4897]: I1121 14:09:39.495858 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nl76g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a7d6101-d0d2-4828-b3ce-d01040c04640\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac58b6ef1bd08e5f950260c18120c72e50b90a0de182c6ab0a5dbbbe45cf7328\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e56402f62821baf335f378084f4ed7c21b60ccbafe148419a0a9a403f106ad19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e56402f62821baf335f378084f4ed7c21b60ccbafe148419a0a9a403f106ad19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6022ff28e83dafcd846a7f03b090262aa3229a0da89f1984e4b6a054a04ac24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6022ff28e83dafcd846a7f03b090262aa3229a0da89f1984e4b6a054a04ac24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89de56e32b796da11bf89dd598d04fbf2884745ddf0f5d87cccfcaf1c001d8d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89de56e32b796da11bf89dd598d04fbf2884745ddf0f5d87cccfcaf1c001d8d6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2dff3cacaba7f9238f482e67b670ea775e0f96621434b99f4e76c2dfdde0321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2dff3cacaba7f9238f482e67b670ea775e0f96621434b99f4e76c2dfdde0321\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a153389a4137ee580e5118f3248400460f9d014c992a956f508815755f583afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a153389a4137ee580e5118f3248400460f9d014c992a956f508815755f583afb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a8c051ad1c1f2373f79eb0781fa6db489aeea3d52ed6c0bb43eec11520ec2b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a8c051ad1c1f2373f79eb0781fa6db489aeea3d52ed6c0bb43eec11520ec2b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nl76g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:39Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:39 crc kubenswrapper[4897]: I1121 14:09:39.500022 4897 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tpfb6_222f65f2-0b82-4760-bcf7-779244720f01/ovnkube-controller/2.log" Nov 21 14:09:39 crc kubenswrapper[4897]: I1121 14:09:39.503159 4897 scope.go:117] "RemoveContainer" containerID="9e285a47c1193b01d8cd3cb16ca750bea2f24d3ad941dd438f96d5203f9203de" Nov 21 14:09:39 crc kubenswrapper[4897]: E1121 14:09:39.503316 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-tpfb6_openshift-ovn-kubernetes(222f65f2-0b82-4760-bcf7-779244720f01)\"" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" podUID="222f65f2-0b82-4760-bcf7-779244720f01" Nov 21 14:09:39 crc kubenswrapper[4897]: I1121 14:09:39.512171 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48c6b1ea7497b33c00b956e400426c1d4f3c6c66e0225133434460800eef9314\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ea6068a6b6f8123b402de354e5374a0f02a7191a922d1718ab38f8898f47c52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:39Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:39 crc kubenswrapper[4897]: I1121 14:09:39.551642 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4qwqj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f46e2ba8-9ee0-406d-a668-2ba424618b69\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f43aa89d4d9defcb50d0dc6f90293b5f426b9cfc398b780c2e10f6a5c9f1e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbsgb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4qwqj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:39Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:39 crc kubenswrapper[4897]: I1121 14:09:39.552691 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:39 crc kubenswrapper[4897]: I1121 14:09:39.552731 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:39 crc kubenswrapper[4897]: I1121 14:09:39.552742 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:39 crc kubenswrapper[4897]: I1121 14:09:39.552759 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:39 crc kubenswrapper[4897]: I1121 14:09:39.552771 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:39Z","lastTransitionTime":"2025-11-21T14:09:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:39 crc kubenswrapper[4897]: I1121 14:09:39.564054 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-rs4c8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7xkdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7xkdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:23Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-rs4c8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:39Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:39 crc kubenswrapper[4897]: I1121 14:09:39.576112 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:39Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:39 crc kubenswrapper[4897]: I1121 14:09:39.586252 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7670227-d280-4847-b882-754429f56b0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://961961696b8750c248598f4cda032b18395e5c802275dcfe6d3e9f7ef3510f1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5854028821e6a360a259427b68e7bb4c2a7220e0c68f4cc91639b2ece0540a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-krv5b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:39Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:39 crc kubenswrapper[4897]: I1121 14:09:39.587812 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd-metrics-certs\") pod \"network-metrics-daemon-rs4c8\" (UID: \"01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd\") " pod="openshift-multus/network-metrics-daemon-rs4c8" Nov 21 14:09:39 crc kubenswrapper[4897]: E1121 14:09:39.588085 4897 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 21 14:09:39 crc kubenswrapper[4897]: E1121 14:09:39.588214 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd-metrics-certs podName:01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd nodeName:}" failed. No retries permitted until 2025-11-21 14:09:55.588177465 +0000 UTC m=+72.872771040 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd-metrics-certs") pod "network-metrics-daemon-rs4c8" (UID: "01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 21 14:09:39 crc kubenswrapper[4897]: I1121 14:09:39.608076 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"222f65f2-0b82-4760-bcf7-779244720f01\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92f1d932a83d65ba6a9a906cc489ff51e76547d07d42656630c59421f833c4db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccd51aa0faf3b1327f207b355cddf86a7353c0051f7e5ce0396f3c4bc67e3b16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8971c38daad45bda3d61545f52120dad40992dbf2e1d6e1e776500ed58d73f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfee19f82e35869a641318a0c45bc7a2ae71d80039f5790e60e277be15c9d3ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec279f2f558a8ac653006321b18cbaf13185876fe303c5cc1c65988aef09edbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de095f3c760b13d5197f624eac3806a1f4186577f2b95eb522532d453eb266e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e285a47c1193b01d8cd3cb16ca750bea2f24d3ad941dd438f96d5203f9203de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://814079abfa7d1b029825da6c6c5f1e57d439ff21ccb626c5d4e52e41d059ad3a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-21T14:09:23Z\\\",\\\"message\\\":\\\"luster\\\\\\\", UUID:\\\\\\\"63b1440a-0908-4cab-8799-012fa1cf0b07\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-operator-lifecycle-manager/olm-operator-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-operator-lifecycle-manager/olm-operator-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-operator-lifecycle-manager/olm-operator-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.168\\\\\\\", Port:8443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1121 14:09:22.444620 6334 ovnkube.go:599] Stopped ovnkube\\\\nI1121 14:09:22.444715 6334 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1121 14:09:22.444771 6334 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:21Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e285a47c1193b01d8cd3cb16ca750bea2f24d3ad941dd438f96d5203f9203de\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-21T14:09:38Z\\\",\\\"message\\\":\\\":776] Recording success event on pod openshift-etcd/etcd-crc\\\\nI1121 14:09:38.005810 6569 default_network_controller.go:776] Recording success event on pod openshift-image-registry/node-ca-7qfmq\\\\nI1121 14:09:38.005827 6569 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-additional-cni-plugins-nl76g\\\\nI1121 14:09:38.005832 6569 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-additional-cni-plugins-nl76g\\\\nI1121 14:09:38.005837 6569 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-additional-cni-plugins-nl76g in node crc\\\\nI1121 14:09:38.005842 6569 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-additional-cni-plugins-nl76g after 0 failed attempt(s)\\\\nI1121 14:09:38.005846 6569 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-additional-cni-plugins-nl76g\\\\nI1121 14:09:38.005834 6569 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-operator-lifecycle-manager/package-server-manager-metrics]} name:Service_openshift-operator-lifecycle-manager/package-server-manager-metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b486a86e0e97ab409e590f30a1bf5cb9f584a7c67f96ff1236e42ca4f8675b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7877db6f5c2bf0eb31218c64d67bca48f921ff7ac62b4e495f29024c086bf738\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7877db6f5c2bf0eb31218c64d67bca48f921ff7ac62b4e495f29024c086bf738\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tpfb6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:39Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:39 crc kubenswrapper[4897]: I1121 14:09:39.617966 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7qfmq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"223b7002-3e67-4848-b1fb-db2921deb64e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381a46c6290da2b36e572f46c16f1a50aa3e25359bcfbc02c7a5a18e1484860e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2m27\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7qfmq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:39Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:39 crc kubenswrapper[4897]: I1121 14:09:39.630410 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b78c993a-9966-4e84-85a1-b4f73aaf17f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce3a2602338a1c1e9380ee55f78360f3d9cf6ccc62678886dc5ae574a72e3d9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c256823f2fb49ecc121f6ef564cd86196f9b4a473e9bdea670173596ef14a823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c65f71f8746044f5d41687c670a0308cd09dc41a6531f9b5720e852b3d240756\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1dbe88e7e6ad63899e107ebb884a16edb9a42b21ae7c86d7fbfaf89785b0d36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:39Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:39 crc kubenswrapper[4897]: I1121 14:09:39.640736 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09877d9780caddd5f7f73e95fdf818a427ed645ed5997984113151f4b4759c13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:39Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:39 crc kubenswrapper[4897]: I1121 14:09:39.651560 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:39Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:39 crc kubenswrapper[4897]: I1121 14:09:39.655767 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:39 crc kubenswrapper[4897]: I1121 14:09:39.655810 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:39 crc kubenswrapper[4897]: I1121 14:09:39.655826 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:39 crc kubenswrapper[4897]: I1121 14:09:39.655845 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:39 crc kubenswrapper[4897]: I1121 14:09:39.655863 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:39Z","lastTransitionTime":"2025-11-21T14:09:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:39 crc kubenswrapper[4897]: I1121 14:09:39.668771 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6ljgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f3b70b5-9294-4f69-9528-500d28f34c89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://527cf28abdeef1b42d9c19fb88e3171dc05c4516c3a542a461b5b1daa85c6d25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78jhb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6ljgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:39Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:39 crc kubenswrapper[4897]: I1121 14:09:39.679494 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x5pws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc9e717e-af03-48bf-b005-62b3d084c6fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f968eee1d8a1888229e23abda59fd58fa5e86f02ca96ff80fbe093904b27e8df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9c4f5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0f99b9213902c003e6ca754134d2482404ace39c184fb3b431d2b600481d5b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9c4f5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-x5pws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:39Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:39 crc kubenswrapper[4897]: I1121 14:09:39.701299 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6cd330fd-e8a6-49d5-bfc9-5f6122ccd49e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1bd64b9b938bc80284bcee54141e90cad613d339c4f27ef2934b29bc2a318916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://936f59996d26e42099dcd055f87fff9e3ac4283b57e1956fe97463f2334152d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca74b28a069dc2e609b0779164c3aa470a15ed9848af0dd5fc3479aaf750157b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://270c53b9a342f26be077a01cd4e87e9ccd0eb5d2c84208ae38cbf720463245e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66c521c26d541bbddb064dd9710f1a7b41090d2f47343977922272dfc6dab7af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6eb2ac410b49098e5b3603de5e6bf5ed4484fab0b300021d0bed2e5ed8a42527\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6eb2ac410b49098e5b3603de5e6bf5ed4484fab0b300021d0bed2e5ed8a42527\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03385e0c78e4806f60c160227ef4905a11003ed3d70f0e5c0dd0a477979d1d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03385e0c78e4806f60c160227ef4905a11003ed3d70f0e5c0dd0a477979d1d4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://50aea2fa8b3b141b4330ddce2c8e0715230c6f06dab64b3abfd98a3067224bec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50aea2fa8b3b141b4330ddce2c8e0715230c6f06dab64b3abfd98a3067224bec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:39Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:39 crc kubenswrapper[4897]: I1121 14:09:39.718792 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a6c0a99-565b-41f9-a8b5-c8990a0c8eb3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://737553dd1d16e2019ea89a376e3a868f9792ed5239aa7d6cc40a05d3d4f286ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6682a8886d2b9bce069cf9261d4701ef6907f19574fd035bd89bd05626023e5c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de079956e3d8de0101f3f87dcc120ef23876c4972382639e419b3f34a38cdfca\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://469b5be7467d4a58d3776ec0a967ed911be31b770b49965da4f5d2fb264d8a3d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1907bfbad7eb34da642e30d6aa81820aab457f92347333a364ac12396b0e9dae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"le observer\\\\nW1121 14:09:07.994401 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1121 14:09:07.994601 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1121 14:09:07.995290 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1461162674/tls.crt::/tmp/serving-cert-1461162674/tls.key\\\\\\\"\\\\nI1121 14:09:08.162382 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1121 14:09:08.165301 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1121 14:09:08.165318 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1121 14:09:08.165343 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1121 14:09:08.165348 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1121 14:09:08.172015 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1121 14:09:08.172083 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1121 14:09:08.172089 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1121 14:09:08.172094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1121 14:09:08.172098 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1121 14:09:08.172102 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1121 14:09:08.172106 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1121 14:09:08.172023 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1121 14:09:08.174214 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d78c87c440c7d8e30075f6143111b3f504fb99686484b8f19a9c8d61a55eccc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://101fb592fef069af009c613a1e9c9825024ed205498ab72334211ba90d6f7397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://101fb592fef069af009c613a1e9c9825024ed205498ab72334211ba90d6f7397\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:39Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:39 crc kubenswrapper[4897]: I1121 14:09:39.734391 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nl76g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a7d6101-d0d2-4828-b3ce-d01040c04640\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac58b6ef1bd08e5f950260c18120c72e50b90a0de182c6ab0a5dbbbe45cf7328\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e56402f62821baf335f378084f4ed7c21b60ccbafe148419a0a9a403f106ad19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e56402f62821baf335f378084f4ed7c21b60ccbafe148419a0a9a403f106ad19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6022ff28e83dafcd846a7f03b090262aa3229a0da89f1984e4b6a054a04ac24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6022ff28e83dafcd846a7f03b090262aa3229a0da89f1984e4b6a054a04ac24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89de56e32b796da11bf89dd598d04fbf2884745ddf0f5d87cccfcaf1c001d8d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89de56e32b796da11bf89dd598d04fbf2884745ddf0f5d87cccfcaf1c001d8d6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2dff3cacaba7f9238f482e67b670ea775e0f96621434b99f4e76c2dfdde0321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2dff3cacaba7f9238f482e67b670ea775e0f96621434b99f4e76c2dfdde0321\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a153389a4137ee580e5118f3248400460f9d014c992a956f508815755f583afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a153389a4137ee580e5118f3248400460f9d014c992a956f508815755f583afb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a8c051ad1c1f2373f79eb0781fa6db489aeea3d52ed6c0bb43eec11520ec2b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a8c051ad1c1f2373f79eb0781fa6db489aeea3d52ed6c0bb43eec11520ec2b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nl76g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:39Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:39 crc kubenswrapper[4897]: I1121 14:09:39.745734 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:39Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:39 crc kubenswrapper[4897]: I1121 14:09:39.758832 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:39 crc kubenswrapper[4897]: I1121 14:09:39.758871 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:39 crc kubenswrapper[4897]: I1121 14:09:39.758882 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:39 crc kubenswrapper[4897]: I1121 14:09:39.758900 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:39 crc kubenswrapper[4897]: I1121 14:09:39.758913 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:39Z","lastTransitionTime":"2025-11-21T14:09:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:39 crc kubenswrapper[4897]: I1121 14:09:39.760945 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c976d804a7ebeba2c879e920cbc389f96f7f5def4b4439e8c8980d5460ecdb65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:39Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:39 crc kubenswrapper[4897]: I1121 14:09:39.792017 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4qwqj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f46e2ba8-9ee0-406d-a668-2ba424618b69\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f43aa89d4d9defcb50d0dc6f90293b5f426b9cfc398b780c2e10f6a5c9f1e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbsgb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4qwqj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:39Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:39 crc kubenswrapper[4897]: I1121 14:09:39.808706 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-rs4c8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7xkdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7xkdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:23Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-rs4c8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:39Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:39 crc kubenswrapper[4897]: I1121 14:09:39.833002 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48c6b1ea7497b33c00b956e400426c1d4f3c6c66e0225133434460800eef9314\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ea6068a6b6f8123b402de354e5374a0f02a7191a922d1718ab38f8898f47c52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:39Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:39 crc kubenswrapper[4897]: I1121 14:09:39.847274 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7670227-d280-4847-b882-754429f56b0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://961961696b8750c248598f4cda032b18395e5c802275dcfe6d3e9f7ef3510f1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5854028821e6a360a259427b68e7bb4c2a7220e0c68f4cc91639b2ece0540a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-krv5b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:39Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:39 crc kubenswrapper[4897]: I1121 14:09:39.861452 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:39 crc kubenswrapper[4897]: I1121 14:09:39.861481 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:39 crc kubenswrapper[4897]: I1121 14:09:39.861490 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:39 crc kubenswrapper[4897]: I1121 14:09:39.861517 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:39 crc kubenswrapper[4897]: I1121 14:09:39.861526 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:39Z","lastTransitionTime":"2025-11-21T14:09:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:39 crc kubenswrapper[4897]: I1121 14:09:39.868123 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"222f65f2-0b82-4760-bcf7-779244720f01\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92f1d932a83d65ba6a9a906cc489ff51e76547d07d42656630c59421f833c4db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccd51aa0faf3b1327f207b355cddf86a7353c0051f7e5ce0396f3c4bc67e3b16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8971c38daad45bda3d61545f52120dad40992dbf2e1d6e1e776500ed58d73f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfee19f82e35869a641318a0c45bc7a2ae71d80039f5790e60e277be15c9d3ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec279f2f558a8ac653006321b18cbaf13185876fe303c5cc1c65988aef09edbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de095f3c760b13d5197f624eac3806a1f4186577f2b95eb522532d453eb266e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e285a47c1193b01d8cd3cb16ca750bea2f24d3ad941dd438f96d5203f9203de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e285a47c1193b01d8cd3cb16ca750bea2f24d3ad941dd438f96d5203f9203de\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-21T14:09:38Z\\\",\\\"message\\\":\\\":776] Recording success event on pod openshift-etcd/etcd-crc\\\\nI1121 14:09:38.005810 6569 default_network_controller.go:776] Recording success event on pod openshift-image-registry/node-ca-7qfmq\\\\nI1121 14:09:38.005827 6569 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-additional-cni-plugins-nl76g\\\\nI1121 14:09:38.005832 6569 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-additional-cni-plugins-nl76g\\\\nI1121 14:09:38.005837 6569 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-additional-cni-plugins-nl76g in node crc\\\\nI1121 14:09:38.005842 6569 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-additional-cni-plugins-nl76g after 0 failed attempt(s)\\\\nI1121 14:09:38.005846 6569 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-additional-cni-plugins-nl76g\\\\nI1121 14:09:38.005834 6569 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-operator-lifecycle-manager/package-server-manager-metrics]} name:Service_openshift-operator-lifecycle-manager/package-server-manager-metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:37Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-tpfb6_openshift-ovn-kubernetes(222f65f2-0b82-4760-bcf7-779244720f01)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b486a86e0e97ab409e590f30a1bf5cb9f584a7c67f96ff1236e42ca4f8675b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7877db6f5c2bf0eb31218c64d67bca48f921ff7ac62b4e495f29024c086bf738\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7877db6f5c2bf0eb31218c64d67bca48f921ff7ac62b4e495f29024c086bf738\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tpfb6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:39Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:39 crc kubenswrapper[4897]: I1121 14:09:39.879767 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7qfmq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"223b7002-3e67-4848-b1fb-db2921deb64e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381a46c6290da2b36e572f46c16f1a50aa3e25359bcfbc02c7a5a18e1484860e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2m27\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7qfmq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:39Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:39 crc kubenswrapper[4897]: I1121 14:09:39.892055 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"402271c9-35e5-4bed-8410-c84c2e132878\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08651f8381818947818e0bb640c089be4a07022efc650befbb65ad58466bf8c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://419ab9403696491ef53a3d812d4c19d676f8fb1a1a4dfcdfcc0c6255377d2918\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://acc9629bb6b0dfbe5b4efe2f9f980e4e3865fdc6742782ebc49a5a7e0f775b9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0fc71bebc463919826d01260f38b63bfdb01199fd8dd6ae90d9ac9cd75db703\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0fc71bebc463919826d01260f38b63bfdb01199fd8dd6ae90d9ac9cd75db703\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:39Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:39 crc kubenswrapper[4897]: I1121 14:09:39.905149 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:39Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:39 crc kubenswrapper[4897]: I1121 14:09:39.964196 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:39 crc kubenswrapper[4897]: I1121 14:09:39.964260 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:39 crc kubenswrapper[4897]: I1121 14:09:39.964274 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:39 crc kubenswrapper[4897]: I1121 14:09:39.964315 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:39 crc kubenswrapper[4897]: I1121 14:09:39.964327 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:39Z","lastTransitionTime":"2025-11-21T14:09:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:40 crc kubenswrapper[4897]: I1121 14:09:40.067421 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:40 crc kubenswrapper[4897]: I1121 14:09:40.067492 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:40 crc kubenswrapper[4897]: I1121 14:09:40.067588 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:40 crc kubenswrapper[4897]: I1121 14:09:40.067624 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:40 crc kubenswrapper[4897]: I1121 14:09:40.067649 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:40Z","lastTransitionTime":"2025-11-21T14:09:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:40 crc kubenswrapper[4897]: I1121 14:09:40.171141 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:40 crc kubenswrapper[4897]: I1121 14:09:40.171204 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:40 crc kubenswrapper[4897]: I1121 14:09:40.171224 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:40 crc kubenswrapper[4897]: I1121 14:09:40.171250 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:40 crc kubenswrapper[4897]: I1121 14:09:40.171269 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:40Z","lastTransitionTime":"2025-11-21T14:09:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:40 crc kubenswrapper[4897]: I1121 14:09:40.274164 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:40 crc kubenswrapper[4897]: I1121 14:09:40.274314 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:40 crc kubenswrapper[4897]: I1121 14:09:40.274341 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:40 crc kubenswrapper[4897]: I1121 14:09:40.274399 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:40 crc kubenswrapper[4897]: I1121 14:09:40.274418 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:40Z","lastTransitionTime":"2025-11-21T14:09:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:40 crc kubenswrapper[4897]: I1121 14:09:40.377606 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:40 crc kubenswrapper[4897]: I1121 14:09:40.377646 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:40 crc kubenswrapper[4897]: I1121 14:09:40.377655 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:40 crc kubenswrapper[4897]: I1121 14:09:40.377670 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:40 crc kubenswrapper[4897]: I1121 14:09:40.377683 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:40Z","lastTransitionTime":"2025-11-21T14:09:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:40 crc kubenswrapper[4897]: I1121 14:09:40.481478 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:40 crc kubenswrapper[4897]: I1121 14:09:40.481566 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:40 crc kubenswrapper[4897]: I1121 14:09:40.481582 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:40 crc kubenswrapper[4897]: I1121 14:09:40.481602 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:40 crc kubenswrapper[4897]: I1121 14:09:40.481617 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:40Z","lastTransitionTime":"2025-11-21T14:09:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:40 crc kubenswrapper[4897]: I1121 14:09:40.585890 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:40 crc kubenswrapper[4897]: I1121 14:09:40.585951 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:40 crc kubenswrapper[4897]: I1121 14:09:40.586002 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:40 crc kubenswrapper[4897]: I1121 14:09:40.586026 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:40 crc kubenswrapper[4897]: I1121 14:09:40.586052 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:40Z","lastTransitionTime":"2025-11-21T14:09:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:40 crc kubenswrapper[4897]: I1121 14:09:40.688555 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:40 crc kubenswrapper[4897]: I1121 14:09:40.688604 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:40 crc kubenswrapper[4897]: I1121 14:09:40.688617 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:40 crc kubenswrapper[4897]: I1121 14:09:40.688639 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:40 crc kubenswrapper[4897]: I1121 14:09:40.688654 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:40Z","lastTransitionTime":"2025-11-21T14:09:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:40 crc kubenswrapper[4897]: I1121 14:09:40.791057 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:40 crc kubenswrapper[4897]: I1121 14:09:40.791098 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:40 crc kubenswrapper[4897]: I1121 14:09:40.791107 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:40 crc kubenswrapper[4897]: I1121 14:09:40.791121 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:40 crc kubenswrapper[4897]: I1121 14:09:40.791130 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:40Z","lastTransitionTime":"2025-11-21T14:09:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:40 crc kubenswrapper[4897]: I1121 14:09:40.894270 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:40 crc kubenswrapper[4897]: I1121 14:09:40.894327 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:40 crc kubenswrapper[4897]: I1121 14:09:40.894341 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:40 crc kubenswrapper[4897]: I1121 14:09:40.894363 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:40 crc kubenswrapper[4897]: I1121 14:09:40.894375 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:40Z","lastTransitionTime":"2025-11-21T14:09:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:40 crc kubenswrapper[4897]: I1121 14:09:40.901175 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 21 14:09:40 crc kubenswrapper[4897]: I1121 14:09:40.901313 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 21 14:09:40 crc kubenswrapper[4897]: E1121 14:09:40.901378 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-21 14:10:12.901336285 +0000 UTC m=+90.185929760 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:09:40 crc kubenswrapper[4897]: E1121 14:09:40.901442 4897 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 21 14:09:40 crc kubenswrapper[4897]: E1121 14:09:40.901536 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-21 14:10:12.901496469 +0000 UTC m=+90.186089944 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 21 14:09:40 crc kubenswrapper[4897]: E1121 14:09:40.901565 4897 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 21 14:09:40 crc kubenswrapper[4897]: I1121 14:09:40.901456 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 21 14:09:40 crc kubenswrapper[4897]: E1121 14:09:40.901608 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-21 14:10:12.901599621 +0000 UTC m=+90.186193096 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 21 14:09:40 crc kubenswrapper[4897]: I1121 14:09:40.996911 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:40 crc kubenswrapper[4897]: I1121 14:09:40.996970 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:40 crc kubenswrapper[4897]: I1121 14:09:40.996983 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:40 crc kubenswrapper[4897]: I1121 14:09:40.997000 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:40 crc kubenswrapper[4897]: I1121 14:09:40.997010 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:40Z","lastTransitionTime":"2025-11-21T14:09:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:41 crc kubenswrapper[4897]: I1121 14:09:41.002372 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 21 14:09:41 crc kubenswrapper[4897]: I1121 14:09:41.002489 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 21 14:09:41 crc kubenswrapper[4897]: E1121 14:09:41.002638 4897 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 21 14:09:41 crc kubenswrapper[4897]: E1121 14:09:41.002663 4897 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 21 14:09:41 crc kubenswrapper[4897]: E1121 14:09:41.002617 4897 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 21 14:09:41 crc kubenswrapper[4897]: E1121 14:09:41.002728 4897 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 21 14:09:41 crc kubenswrapper[4897]: E1121 14:09:41.002746 4897 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 21 14:09:41 crc kubenswrapper[4897]: E1121 14:09:41.002675 4897 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 21 14:09:41 crc kubenswrapper[4897]: E1121 14:09:41.002862 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-21 14:10:13.002828858 +0000 UTC m=+90.287422373 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 21 14:09:41 crc kubenswrapper[4897]: E1121 14:09:41.002893 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-21 14:10:13.00288116 +0000 UTC m=+90.287474855 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 21 14:09:41 crc kubenswrapper[4897]: I1121 14:09:41.088964 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 21 14:09:41 crc kubenswrapper[4897]: I1121 14:09:41.089020 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs4c8" Nov 21 14:09:41 crc kubenswrapper[4897]: I1121 14:09:41.089121 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 21 14:09:41 crc kubenswrapper[4897]: I1121 14:09:41.089177 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 21 14:09:41 crc kubenswrapper[4897]: E1121 14:09:41.089212 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 21 14:09:41 crc kubenswrapper[4897]: E1121 14:09:41.089317 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 21 14:09:41 crc kubenswrapper[4897]: E1121 14:09:41.089440 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs4c8" podUID="01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd" Nov 21 14:09:41 crc kubenswrapper[4897]: E1121 14:09:41.089616 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 21 14:09:41 crc kubenswrapper[4897]: I1121 14:09:41.101072 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:41 crc kubenswrapper[4897]: I1121 14:09:41.101137 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:41 crc kubenswrapper[4897]: I1121 14:09:41.101161 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:41 crc kubenswrapper[4897]: I1121 14:09:41.101184 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:41 crc kubenswrapper[4897]: I1121 14:09:41.101197 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:41Z","lastTransitionTime":"2025-11-21T14:09:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:41 crc kubenswrapper[4897]: I1121 14:09:41.204680 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:41 crc kubenswrapper[4897]: I1121 14:09:41.204747 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:41 crc kubenswrapper[4897]: I1121 14:09:41.204757 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:41 crc kubenswrapper[4897]: I1121 14:09:41.204778 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:41 crc kubenswrapper[4897]: I1121 14:09:41.204791 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:41Z","lastTransitionTime":"2025-11-21T14:09:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:41 crc kubenswrapper[4897]: I1121 14:09:41.307397 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:41 crc kubenswrapper[4897]: I1121 14:09:41.307438 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:41 crc kubenswrapper[4897]: I1121 14:09:41.307447 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:41 crc kubenswrapper[4897]: I1121 14:09:41.307462 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:41 crc kubenswrapper[4897]: I1121 14:09:41.307474 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:41Z","lastTransitionTime":"2025-11-21T14:09:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:41 crc kubenswrapper[4897]: I1121 14:09:41.410191 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:41 crc kubenswrapper[4897]: I1121 14:09:41.410257 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:41 crc kubenswrapper[4897]: I1121 14:09:41.410275 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:41 crc kubenswrapper[4897]: I1121 14:09:41.410299 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:41 crc kubenswrapper[4897]: I1121 14:09:41.410319 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:41Z","lastTransitionTime":"2025-11-21T14:09:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:41 crc kubenswrapper[4897]: I1121 14:09:41.513226 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:41 crc kubenswrapper[4897]: I1121 14:09:41.513271 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:41 crc kubenswrapper[4897]: I1121 14:09:41.513279 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:41 crc kubenswrapper[4897]: I1121 14:09:41.513293 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:41 crc kubenswrapper[4897]: I1121 14:09:41.513304 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:41Z","lastTransitionTime":"2025-11-21T14:09:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:41 crc kubenswrapper[4897]: I1121 14:09:41.616211 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:41 crc kubenswrapper[4897]: I1121 14:09:41.616261 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:41 crc kubenswrapper[4897]: I1121 14:09:41.616270 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:41 crc kubenswrapper[4897]: I1121 14:09:41.616288 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:41 crc kubenswrapper[4897]: I1121 14:09:41.616301 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:41Z","lastTransitionTime":"2025-11-21T14:09:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:41 crc kubenswrapper[4897]: I1121 14:09:41.719900 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:41 crc kubenswrapper[4897]: I1121 14:09:41.719955 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:41 crc kubenswrapper[4897]: I1121 14:09:41.719970 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:41 crc kubenswrapper[4897]: I1121 14:09:41.719989 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:41 crc kubenswrapper[4897]: I1121 14:09:41.720001 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:41Z","lastTransitionTime":"2025-11-21T14:09:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:41 crc kubenswrapper[4897]: I1121 14:09:41.824063 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:41 crc kubenswrapper[4897]: I1121 14:09:41.824112 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:41 crc kubenswrapper[4897]: I1121 14:09:41.824124 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:41 crc kubenswrapper[4897]: I1121 14:09:41.824142 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:41 crc kubenswrapper[4897]: I1121 14:09:41.824156 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:41Z","lastTransitionTime":"2025-11-21T14:09:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:41 crc kubenswrapper[4897]: I1121 14:09:41.927605 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:41 crc kubenswrapper[4897]: I1121 14:09:41.927643 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:41 crc kubenswrapper[4897]: I1121 14:09:41.927652 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:41 crc kubenswrapper[4897]: I1121 14:09:41.927668 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:41 crc kubenswrapper[4897]: I1121 14:09:41.927678 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:41Z","lastTransitionTime":"2025-11-21T14:09:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:42 crc kubenswrapper[4897]: I1121 14:09:42.031270 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:42 crc kubenswrapper[4897]: I1121 14:09:42.031335 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:42 crc kubenswrapper[4897]: I1121 14:09:42.031353 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:42 crc kubenswrapper[4897]: I1121 14:09:42.031375 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:42 crc kubenswrapper[4897]: I1121 14:09:42.031390 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:42Z","lastTransitionTime":"2025-11-21T14:09:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:42 crc kubenswrapper[4897]: I1121 14:09:42.134289 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:42 crc kubenswrapper[4897]: I1121 14:09:42.134337 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:42 crc kubenswrapper[4897]: I1121 14:09:42.134350 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:42 crc kubenswrapper[4897]: I1121 14:09:42.134366 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:42 crc kubenswrapper[4897]: I1121 14:09:42.134377 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:42Z","lastTransitionTime":"2025-11-21T14:09:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:42 crc kubenswrapper[4897]: I1121 14:09:42.237164 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:42 crc kubenswrapper[4897]: I1121 14:09:42.237216 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:42 crc kubenswrapper[4897]: I1121 14:09:42.237227 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:42 crc kubenswrapper[4897]: I1121 14:09:42.237244 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:42 crc kubenswrapper[4897]: I1121 14:09:42.237255 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:42Z","lastTransitionTime":"2025-11-21T14:09:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:42 crc kubenswrapper[4897]: I1121 14:09:42.339795 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:42 crc kubenswrapper[4897]: I1121 14:09:42.339862 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:42 crc kubenswrapper[4897]: I1121 14:09:42.339875 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:42 crc kubenswrapper[4897]: I1121 14:09:42.339899 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:42 crc kubenswrapper[4897]: I1121 14:09:42.339913 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:42Z","lastTransitionTime":"2025-11-21T14:09:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:42 crc kubenswrapper[4897]: I1121 14:09:42.442387 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:42 crc kubenswrapper[4897]: I1121 14:09:42.442428 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:42 crc kubenswrapper[4897]: I1121 14:09:42.442440 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:42 crc kubenswrapper[4897]: I1121 14:09:42.442471 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:42 crc kubenswrapper[4897]: I1121 14:09:42.442482 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:42Z","lastTransitionTime":"2025-11-21T14:09:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:42 crc kubenswrapper[4897]: I1121 14:09:42.545921 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:42 crc kubenswrapper[4897]: I1121 14:09:42.545967 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:42 crc kubenswrapper[4897]: I1121 14:09:42.545975 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:42 crc kubenswrapper[4897]: I1121 14:09:42.545991 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:42 crc kubenswrapper[4897]: I1121 14:09:42.546002 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:42Z","lastTransitionTime":"2025-11-21T14:09:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:42 crc kubenswrapper[4897]: I1121 14:09:42.648753 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:42 crc kubenswrapper[4897]: I1121 14:09:42.648792 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:42 crc kubenswrapper[4897]: I1121 14:09:42.648801 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:42 crc kubenswrapper[4897]: I1121 14:09:42.648815 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:42 crc kubenswrapper[4897]: I1121 14:09:42.648825 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:42Z","lastTransitionTime":"2025-11-21T14:09:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:42 crc kubenswrapper[4897]: I1121 14:09:42.751968 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:42 crc kubenswrapper[4897]: I1121 14:09:42.752012 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:42 crc kubenswrapper[4897]: I1121 14:09:42.752025 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:42 crc kubenswrapper[4897]: I1121 14:09:42.752045 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:42 crc kubenswrapper[4897]: I1121 14:09:42.752062 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:42Z","lastTransitionTime":"2025-11-21T14:09:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:42 crc kubenswrapper[4897]: I1121 14:09:42.854669 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:42 crc kubenswrapper[4897]: I1121 14:09:42.854713 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:42 crc kubenswrapper[4897]: I1121 14:09:42.854723 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:42 crc kubenswrapper[4897]: I1121 14:09:42.854742 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:42 crc kubenswrapper[4897]: I1121 14:09:42.854753 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:42Z","lastTransitionTime":"2025-11-21T14:09:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:42 crc kubenswrapper[4897]: I1121 14:09:42.957491 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:42 crc kubenswrapper[4897]: I1121 14:09:42.957557 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:42 crc kubenswrapper[4897]: I1121 14:09:42.957568 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:42 crc kubenswrapper[4897]: I1121 14:09:42.957585 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:42 crc kubenswrapper[4897]: I1121 14:09:42.957599 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:42Z","lastTransitionTime":"2025-11-21T14:09:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:43 crc kubenswrapper[4897]: I1121 14:09:43.060634 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:43 crc kubenswrapper[4897]: I1121 14:09:43.060695 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:43 crc kubenswrapper[4897]: I1121 14:09:43.060708 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:43 crc kubenswrapper[4897]: I1121 14:09:43.060730 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:43 crc kubenswrapper[4897]: I1121 14:09:43.060747 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:43Z","lastTransitionTime":"2025-11-21T14:09:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:43 crc kubenswrapper[4897]: I1121 14:09:43.088806 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 21 14:09:43 crc kubenswrapper[4897]: I1121 14:09:43.088849 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 21 14:09:43 crc kubenswrapper[4897]: I1121 14:09:43.088872 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 21 14:09:43 crc kubenswrapper[4897]: I1121 14:09:43.088942 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs4c8" Nov 21 14:09:43 crc kubenswrapper[4897]: E1121 14:09:43.089048 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 21 14:09:43 crc kubenswrapper[4897]: E1121 14:09:43.089133 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 21 14:09:43 crc kubenswrapper[4897]: E1121 14:09:43.089267 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 21 14:09:43 crc kubenswrapper[4897]: E1121 14:09:43.089405 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs4c8" podUID="01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd" Nov 21 14:09:43 crc kubenswrapper[4897]: I1121 14:09:43.163368 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:43 crc kubenswrapper[4897]: I1121 14:09:43.163411 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:43 crc kubenswrapper[4897]: I1121 14:09:43.163419 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:43 crc kubenswrapper[4897]: I1121 14:09:43.163435 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:43 crc kubenswrapper[4897]: I1121 14:09:43.163446 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:43Z","lastTransitionTime":"2025-11-21T14:09:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:43 crc kubenswrapper[4897]: I1121 14:09:43.266401 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:43 crc kubenswrapper[4897]: I1121 14:09:43.266452 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:43 crc kubenswrapper[4897]: I1121 14:09:43.266463 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:43 crc kubenswrapper[4897]: I1121 14:09:43.266483 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:43 crc kubenswrapper[4897]: I1121 14:09:43.266498 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:43Z","lastTransitionTime":"2025-11-21T14:09:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:43 crc kubenswrapper[4897]: I1121 14:09:43.368850 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:43 crc kubenswrapper[4897]: I1121 14:09:43.368903 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:43 crc kubenswrapper[4897]: I1121 14:09:43.368912 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:43 crc kubenswrapper[4897]: I1121 14:09:43.368928 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:43 crc kubenswrapper[4897]: I1121 14:09:43.368941 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:43Z","lastTransitionTime":"2025-11-21T14:09:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:43 crc kubenswrapper[4897]: I1121 14:09:43.471489 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:43 crc kubenswrapper[4897]: I1121 14:09:43.471547 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:43 crc kubenswrapper[4897]: I1121 14:09:43.471555 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:43 crc kubenswrapper[4897]: I1121 14:09:43.471571 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:43 crc kubenswrapper[4897]: I1121 14:09:43.471581 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:43Z","lastTransitionTime":"2025-11-21T14:09:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:43 crc kubenswrapper[4897]: I1121 14:09:43.575044 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:43 crc kubenswrapper[4897]: I1121 14:09:43.575098 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:43 crc kubenswrapper[4897]: I1121 14:09:43.575113 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:43 crc kubenswrapper[4897]: I1121 14:09:43.575133 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:43 crc kubenswrapper[4897]: I1121 14:09:43.575146 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:43Z","lastTransitionTime":"2025-11-21T14:09:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:43 crc kubenswrapper[4897]: I1121 14:09:43.680225 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:43 crc kubenswrapper[4897]: I1121 14:09:43.680257 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:43 crc kubenswrapper[4897]: I1121 14:09:43.680267 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:43 crc kubenswrapper[4897]: I1121 14:09:43.680280 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:43 crc kubenswrapper[4897]: I1121 14:09:43.680290 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:43Z","lastTransitionTime":"2025-11-21T14:09:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:43 crc kubenswrapper[4897]: I1121 14:09:43.783621 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:43 crc kubenswrapper[4897]: I1121 14:09:43.783658 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:43 crc kubenswrapper[4897]: I1121 14:09:43.783669 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:43 crc kubenswrapper[4897]: I1121 14:09:43.783684 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:43 crc kubenswrapper[4897]: I1121 14:09:43.783696 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:43Z","lastTransitionTime":"2025-11-21T14:09:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:43 crc kubenswrapper[4897]: I1121 14:09:43.886479 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:43 crc kubenswrapper[4897]: I1121 14:09:43.886583 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:43 crc kubenswrapper[4897]: I1121 14:09:43.886592 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:43 crc kubenswrapper[4897]: I1121 14:09:43.886610 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:43 crc kubenswrapper[4897]: I1121 14:09:43.886621 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:43Z","lastTransitionTime":"2025-11-21T14:09:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:43 crc kubenswrapper[4897]: I1121 14:09:43.988646 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:43 crc kubenswrapper[4897]: I1121 14:09:43.988684 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:43 crc kubenswrapper[4897]: I1121 14:09:43.988694 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:43 crc kubenswrapper[4897]: I1121 14:09:43.988712 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:43 crc kubenswrapper[4897]: I1121 14:09:43.988722 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:43Z","lastTransitionTime":"2025-11-21T14:09:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:44 crc kubenswrapper[4897]: I1121 14:09:44.092676 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:44 crc kubenswrapper[4897]: I1121 14:09:44.092717 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:44 crc kubenswrapper[4897]: I1121 14:09:44.092725 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:44 crc kubenswrapper[4897]: I1121 14:09:44.092743 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:44 crc kubenswrapper[4897]: I1121 14:09:44.092755 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:44Z","lastTransitionTime":"2025-11-21T14:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:44 crc kubenswrapper[4897]: I1121 14:09:44.101274 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48c6b1ea7497b33c00b956e400426c1d4f3c6c66e0225133434460800eef9314\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ea6068a6b6f8123b402de354e5374a0f02a7191a922d1718ab38f8898f47c52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:44Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:44 crc kubenswrapper[4897]: I1121 14:09:44.110879 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4qwqj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f46e2ba8-9ee0-406d-a668-2ba424618b69\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f43aa89d4d9defcb50d0dc6f90293b5f426b9cfc398b780c2e10f6a5c9f1e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbsgb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4qwqj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:44Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:44 crc kubenswrapper[4897]: I1121 14:09:44.121352 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-rs4c8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7xkdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7xkdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:23Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-rs4c8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:44Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:44 crc kubenswrapper[4897]: I1121 14:09:44.132167 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:44Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:44 crc kubenswrapper[4897]: I1121 14:09:44.142381 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7670227-d280-4847-b882-754429f56b0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://961961696b8750c248598f4cda032b18395e5c802275dcfe6d3e9f7ef3510f1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5854028821e6a360a259427b68e7bb4c2a7220e0c68f4cc91639b2ece0540a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-krv5b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:44Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:44 crc kubenswrapper[4897]: I1121 14:09:44.159189 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"222f65f2-0b82-4760-bcf7-779244720f01\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92f1d932a83d65ba6a9a906cc489ff51e76547d07d42656630c59421f833c4db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccd51aa0faf3b1327f207b355cddf86a7353c0051f7e5ce0396f3c4bc67e3b16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8971c38daad45bda3d61545f52120dad40992dbf2e1d6e1e776500ed58d73f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfee19f82e35869a641318a0c45bc7a2ae71d80039f5790e60e277be15c9d3ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec279f2f558a8ac653006321b18cbaf13185876fe303c5cc1c65988aef09edbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de095f3c760b13d5197f624eac3806a1f4186577f2b95eb522532d453eb266e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e285a47c1193b01d8cd3cb16ca750bea2f24d3ad941dd438f96d5203f9203de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e285a47c1193b01d8cd3cb16ca750bea2f24d3ad941dd438f96d5203f9203de\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-21T14:09:38Z\\\",\\\"message\\\":\\\":776] Recording success event on pod openshift-etcd/etcd-crc\\\\nI1121 14:09:38.005810 6569 default_network_controller.go:776] Recording success event on pod openshift-image-registry/node-ca-7qfmq\\\\nI1121 14:09:38.005827 6569 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-additional-cni-plugins-nl76g\\\\nI1121 14:09:38.005832 6569 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-additional-cni-plugins-nl76g\\\\nI1121 14:09:38.005837 6569 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-additional-cni-plugins-nl76g in node crc\\\\nI1121 14:09:38.005842 6569 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-additional-cni-plugins-nl76g after 0 failed attempt(s)\\\\nI1121 14:09:38.005846 6569 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-additional-cni-plugins-nl76g\\\\nI1121 14:09:38.005834 6569 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-operator-lifecycle-manager/package-server-manager-metrics]} name:Service_openshift-operator-lifecycle-manager/package-server-manager-metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:37Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-tpfb6_openshift-ovn-kubernetes(222f65f2-0b82-4760-bcf7-779244720f01)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b486a86e0e97ab409e590f30a1bf5cb9f584a7c67f96ff1236e42ca4f8675b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7877db6f5c2bf0eb31218c64d67bca48f921ff7ac62b4e495f29024c086bf738\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7877db6f5c2bf0eb31218c64d67bca48f921ff7ac62b4e495f29024c086bf738\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tpfb6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:44Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:44 crc kubenswrapper[4897]: I1121 14:09:44.170361 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7qfmq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"223b7002-3e67-4848-b1fb-db2921deb64e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381a46c6290da2b36e572f46c16f1a50aa3e25359bcfbc02c7a5a18e1484860e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2m27\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7qfmq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:44Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:44 crc kubenswrapper[4897]: I1121 14:09:44.183282 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"402271c9-35e5-4bed-8410-c84c2e132878\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08651f8381818947818e0bb640c089be4a07022efc650befbb65ad58466bf8c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://419ab9403696491ef53a3d812d4c19d676f8fb1a1a4dfcdfcc0c6255377d2918\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://acc9629bb6b0dfbe5b4efe2f9f980e4e3865fdc6742782ebc49a5a7e0f775b9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0fc71bebc463919826d01260f38b63bfdb01199fd8dd6ae90d9ac9cd75db703\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0fc71bebc463919826d01260f38b63bfdb01199fd8dd6ae90d9ac9cd75db703\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:44Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:44 crc kubenswrapper[4897]: I1121 14:09:44.195284 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:44 crc kubenswrapper[4897]: I1121 14:09:44.195344 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:44 crc kubenswrapper[4897]: I1121 14:09:44.195358 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:44 crc kubenswrapper[4897]: I1121 14:09:44.195382 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:44 crc kubenswrapper[4897]: I1121 14:09:44.195397 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:44Z","lastTransitionTime":"2025-11-21T14:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:44 crc kubenswrapper[4897]: I1121 14:09:44.197724 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a6c0a99-565b-41f9-a8b5-c8990a0c8eb3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://737553dd1d16e2019ea89a376e3a868f9792ed5239aa7d6cc40a05d3d4f286ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6682a8886d2b9bce069cf9261d4701ef6907f19574fd035bd89bd05626023e5c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de079956e3d8de0101f3f87dcc120ef23876c4972382639e419b3f34a38cdfca\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://469b5be7467d4a58d3776ec0a967ed911be31b770b49965da4f5d2fb264d8a3d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1907bfbad7eb34da642e30d6aa81820aab457f92347333a364ac12396b0e9dae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"le observer\\\\nW1121 14:09:07.994401 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1121 14:09:07.994601 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1121 14:09:07.995290 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1461162674/tls.crt::/tmp/serving-cert-1461162674/tls.key\\\\\\\"\\\\nI1121 14:09:08.162382 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1121 14:09:08.165301 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1121 14:09:08.165318 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1121 14:09:08.165343 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1121 14:09:08.165348 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1121 14:09:08.172015 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1121 14:09:08.172083 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1121 14:09:08.172089 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1121 14:09:08.172094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1121 14:09:08.172098 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1121 14:09:08.172102 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1121 14:09:08.172106 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1121 14:09:08.172023 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1121 14:09:08.174214 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d78c87c440c7d8e30075f6143111b3f504fb99686484b8f19a9c8d61a55eccc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://101fb592fef069af009c613a1e9c9825024ed205498ab72334211ba90d6f7397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://101fb592fef069af009c613a1e9c9825024ed205498ab72334211ba90d6f7397\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:44Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:44 crc kubenswrapper[4897]: I1121 14:09:44.209763 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b78c993a-9966-4e84-85a1-b4f73aaf17f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce3a2602338a1c1e9380ee55f78360f3d9cf6ccc62678886dc5ae574a72e3d9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c256823f2fb49ecc121f6ef564cd86196f9b4a473e9bdea670173596ef14a823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c65f71f8746044f5d41687c670a0308cd09dc41a6531f9b5720e852b3d240756\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1dbe88e7e6ad63899e107ebb884a16edb9a42b21ae7c86d7fbfaf89785b0d36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:44Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:44 crc kubenswrapper[4897]: I1121 14:09:44.221180 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09877d9780caddd5f7f73e95fdf818a427ed645ed5997984113151f4b4759c13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:44Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:44 crc kubenswrapper[4897]: I1121 14:09:44.233796 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:44Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:44 crc kubenswrapper[4897]: I1121 14:09:44.246818 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6ljgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f3b70b5-9294-4f69-9528-500d28f34c89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://527cf28abdeef1b42d9c19fb88e3171dc05c4516c3a542a461b5b1daa85c6d25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78jhb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6ljgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:44Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:44 crc kubenswrapper[4897]: I1121 14:09:44.258094 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x5pws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc9e717e-af03-48bf-b005-62b3d084c6fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f968eee1d8a1888229e23abda59fd58fa5e86f02ca96ff80fbe093904b27e8df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9c4f5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0f99b9213902c003e6ca754134d2482404ace39c184fb3b431d2b600481d5b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9c4f5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-x5pws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:44Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:44 crc kubenswrapper[4897]: I1121 14:09:44.276041 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6cd330fd-e8a6-49d5-bfc9-5f6122ccd49e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1bd64b9b938bc80284bcee54141e90cad613d339c4f27ef2934b29bc2a318916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://936f59996d26e42099dcd055f87fff9e3ac4283b57e1956fe97463f2334152d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca74b28a069dc2e609b0779164c3aa470a15ed9848af0dd5fc3479aaf750157b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://270c53b9a342f26be077a01cd4e87e9ccd0eb5d2c84208ae38cbf720463245e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66c521c26d541bbddb064dd9710f1a7b41090d2f47343977922272dfc6dab7af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6eb2ac410b49098e5b3603de5e6bf5ed4484fab0b300021d0bed2e5ed8a42527\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6eb2ac410b49098e5b3603de5e6bf5ed4484fab0b300021d0bed2e5ed8a42527\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03385e0c78e4806f60c160227ef4905a11003ed3d70f0e5c0dd0a477979d1d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03385e0c78e4806f60c160227ef4905a11003ed3d70f0e5c0dd0a477979d1d4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://50aea2fa8b3b141b4330ddce2c8e0715230c6f06dab64b3abfd98a3067224bec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50aea2fa8b3b141b4330ddce2c8e0715230c6f06dab64b3abfd98a3067224bec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:44Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:44 crc kubenswrapper[4897]: I1121 14:09:44.290011 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c976d804a7ebeba2c879e920cbc389f96f7f5def4b4439e8c8980d5460ecdb65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:44Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:44 crc kubenswrapper[4897]: I1121 14:09:44.297942 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:44 crc kubenswrapper[4897]: I1121 14:09:44.297987 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:44 crc kubenswrapper[4897]: I1121 14:09:44.297999 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:44 crc kubenswrapper[4897]: I1121 14:09:44.298042 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:44 crc kubenswrapper[4897]: I1121 14:09:44.298057 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:44Z","lastTransitionTime":"2025-11-21T14:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:44 crc kubenswrapper[4897]: I1121 14:09:44.305116 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nl76g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a7d6101-d0d2-4828-b3ce-d01040c04640\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac58b6ef1bd08e5f950260c18120c72e50b90a0de182c6ab0a5dbbbe45cf7328\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e56402f62821baf335f378084f4ed7c21b60ccbafe148419a0a9a403f106ad19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e56402f62821baf335f378084f4ed7c21b60ccbafe148419a0a9a403f106ad19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6022ff28e83dafcd846a7f03b090262aa3229a0da89f1984e4b6a054a04ac24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6022ff28e83dafcd846a7f03b090262aa3229a0da89f1984e4b6a054a04ac24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89de56e32b796da11bf89dd598d04fbf2884745ddf0f5d87cccfcaf1c001d8d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89de56e32b796da11bf89dd598d04fbf2884745ddf0f5d87cccfcaf1c001d8d6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2dff3cacaba7f9238f482e67b670ea775e0f96621434b99f4e76c2dfdde0321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2dff3cacaba7f9238f482e67b670ea775e0f96621434b99f4e76c2dfdde0321\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a153389a4137ee580e5118f3248400460f9d014c992a956f508815755f583afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a153389a4137ee580e5118f3248400460f9d014c992a956f508815755f583afb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a8c051ad1c1f2373f79eb0781fa6db489aeea3d52ed6c0bb43eec11520ec2b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a8c051ad1c1f2373f79eb0781fa6db489aeea3d52ed6c0bb43eec11520ec2b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nl76g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:44Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:44 crc kubenswrapper[4897]: I1121 14:09:44.319281 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:44Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:44 crc kubenswrapper[4897]: I1121 14:09:44.400642 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:44 crc kubenswrapper[4897]: I1121 14:09:44.400686 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:44 crc kubenswrapper[4897]: I1121 14:09:44.400697 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:44 crc kubenswrapper[4897]: I1121 14:09:44.400718 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:44 crc kubenswrapper[4897]: I1121 14:09:44.400729 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:44Z","lastTransitionTime":"2025-11-21T14:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:44 crc kubenswrapper[4897]: I1121 14:09:44.503059 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:44 crc kubenswrapper[4897]: I1121 14:09:44.503104 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:44 crc kubenswrapper[4897]: I1121 14:09:44.503113 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:44 crc kubenswrapper[4897]: I1121 14:09:44.503128 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:44 crc kubenswrapper[4897]: I1121 14:09:44.503137 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:44Z","lastTransitionTime":"2025-11-21T14:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:44 crc kubenswrapper[4897]: I1121 14:09:44.606160 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:44 crc kubenswrapper[4897]: I1121 14:09:44.606223 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:44 crc kubenswrapper[4897]: I1121 14:09:44.606234 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:44 crc kubenswrapper[4897]: I1121 14:09:44.606250 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:44 crc kubenswrapper[4897]: I1121 14:09:44.606263 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:44Z","lastTransitionTime":"2025-11-21T14:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:44 crc kubenswrapper[4897]: I1121 14:09:44.708324 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:44 crc kubenswrapper[4897]: I1121 14:09:44.708371 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:44 crc kubenswrapper[4897]: I1121 14:09:44.708379 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:44 crc kubenswrapper[4897]: I1121 14:09:44.708397 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:44 crc kubenswrapper[4897]: I1121 14:09:44.708409 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:44Z","lastTransitionTime":"2025-11-21T14:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:44 crc kubenswrapper[4897]: I1121 14:09:44.810914 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:44 crc kubenswrapper[4897]: I1121 14:09:44.810967 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:44 crc kubenswrapper[4897]: I1121 14:09:44.810979 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:44 crc kubenswrapper[4897]: I1121 14:09:44.810999 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:44 crc kubenswrapper[4897]: I1121 14:09:44.811014 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:44Z","lastTransitionTime":"2025-11-21T14:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:44 crc kubenswrapper[4897]: I1121 14:09:44.914434 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:44 crc kubenswrapper[4897]: I1121 14:09:44.914498 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:44 crc kubenswrapper[4897]: I1121 14:09:44.914528 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:44 crc kubenswrapper[4897]: I1121 14:09:44.914545 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:44 crc kubenswrapper[4897]: I1121 14:09:44.914556 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:44Z","lastTransitionTime":"2025-11-21T14:09:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:45 crc kubenswrapper[4897]: I1121 14:09:45.017843 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:45 crc kubenswrapper[4897]: I1121 14:09:45.017900 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:45 crc kubenswrapper[4897]: I1121 14:09:45.017912 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:45 crc kubenswrapper[4897]: I1121 14:09:45.017936 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:45 crc kubenswrapper[4897]: I1121 14:09:45.017946 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:45Z","lastTransitionTime":"2025-11-21T14:09:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:45 crc kubenswrapper[4897]: I1121 14:09:45.088796 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 21 14:09:45 crc kubenswrapper[4897]: I1121 14:09:45.088916 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs4c8" Nov 21 14:09:45 crc kubenswrapper[4897]: I1121 14:09:45.088955 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 21 14:09:45 crc kubenswrapper[4897]: E1121 14:09:45.089030 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 21 14:09:45 crc kubenswrapper[4897]: E1121 14:09:45.089158 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs4c8" podUID="01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd" Nov 21 14:09:45 crc kubenswrapper[4897]: I1121 14:09:45.089222 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 21 14:09:45 crc kubenswrapper[4897]: E1121 14:09:45.089361 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 21 14:09:45 crc kubenswrapper[4897]: E1121 14:09:45.089596 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 21 14:09:45 crc kubenswrapper[4897]: I1121 14:09:45.121689 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:45 crc kubenswrapper[4897]: I1121 14:09:45.121770 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:45 crc kubenswrapper[4897]: I1121 14:09:45.121780 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:45 crc kubenswrapper[4897]: I1121 14:09:45.121814 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:45 crc kubenswrapper[4897]: I1121 14:09:45.121825 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:45Z","lastTransitionTime":"2025-11-21T14:09:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:45 crc kubenswrapper[4897]: I1121 14:09:45.226806 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:45 crc kubenswrapper[4897]: I1121 14:09:45.226899 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:45 crc kubenswrapper[4897]: I1121 14:09:45.226912 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:45 crc kubenswrapper[4897]: I1121 14:09:45.226935 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:45 crc kubenswrapper[4897]: I1121 14:09:45.226948 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:45Z","lastTransitionTime":"2025-11-21T14:09:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:45 crc kubenswrapper[4897]: I1121 14:09:45.329965 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:45 crc kubenswrapper[4897]: I1121 14:09:45.330014 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:45 crc kubenswrapper[4897]: I1121 14:09:45.330026 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:45 crc kubenswrapper[4897]: I1121 14:09:45.330044 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:45 crc kubenswrapper[4897]: I1121 14:09:45.330056 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:45Z","lastTransitionTime":"2025-11-21T14:09:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:45 crc kubenswrapper[4897]: I1121 14:09:45.433054 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:45 crc kubenswrapper[4897]: I1121 14:09:45.433131 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:45 crc kubenswrapper[4897]: I1121 14:09:45.433143 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:45 crc kubenswrapper[4897]: I1121 14:09:45.433163 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:45 crc kubenswrapper[4897]: I1121 14:09:45.433173 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:45Z","lastTransitionTime":"2025-11-21T14:09:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:45 crc kubenswrapper[4897]: I1121 14:09:45.536742 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:45 crc kubenswrapper[4897]: I1121 14:09:45.536795 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:45 crc kubenswrapper[4897]: I1121 14:09:45.536807 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:45 crc kubenswrapper[4897]: I1121 14:09:45.536827 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:45 crc kubenswrapper[4897]: I1121 14:09:45.536842 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:45Z","lastTransitionTime":"2025-11-21T14:09:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:45 crc kubenswrapper[4897]: I1121 14:09:45.639635 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:45 crc kubenswrapper[4897]: I1121 14:09:45.639720 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:45 crc kubenswrapper[4897]: I1121 14:09:45.639746 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:45 crc kubenswrapper[4897]: I1121 14:09:45.639774 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:45 crc kubenswrapper[4897]: I1121 14:09:45.639793 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:45Z","lastTransitionTime":"2025-11-21T14:09:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:45 crc kubenswrapper[4897]: I1121 14:09:45.742096 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:45 crc kubenswrapper[4897]: I1121 14:09:45.742291 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:45 crc kubenswrapper[4897]: I1121 14:09:45.742302 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:45 crc kubenswrapper[4897]: I1121 14:09:45.742318 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:45 crc kubenswrapper[4897]: I1121 14:09:45.742331 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:45Z","lastTransitionTime":"2025-11-21T14:09:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:45 crc kubenswrapper[4897]: I1121 14:09:45.844549 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:45 crc kubenswrapper[4897]: I1121 14:09:45.844592 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:45 crc kubenswrapper[4897]: I1121 14:09:45.844611 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:45 crc kubenswrapper[4897]: I1121 14:09:45.844627 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:45 crc kubenswrapper[4897]: I1121 14:09:45.844640 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:45Z","lastTransitionTime":"2025-11-21T14:09:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:45 crc kubenswrapper[4897]: I1121 14:09:45.946777 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:45 crc kubenswrapper[4897]: I1121 14:09:45.946828 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:45 crc kubenswrapper[4897]: I1121 14:09:45.946838 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:45 crc kubenswrapper[4897]: I1121 14:09:45.946855 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:45 crc kubenswrapper[4897]: I1121 14:09:45.946868 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:45Z","lastTransitionTime":"2025-11-21T14:09:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:46 crc kubenswrapper[4897]: I1121 14:09:46.050723 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:46 crc kubenswrapper[4897]: I1121 14:09:46.050778 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:46 crc kubenswrapper[4897]: I1121 14:09:46.050788 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:46 crc kubenswrapper[4897]: I1121 14:09:46.050803 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:46 crc kubenswrapper[4897]: I1121 14:09:46.050814 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:46Z","lastTransitionTime":"2025-11-21T14:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:46 crc kubenswrapper[4897]: I1121 14:09:46.113181 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:46 crc kubenswrapper[4897]: I1121 14:09:46.113230 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:46 crc kubenswrapper[4897]: I1121 14:09:46.113238 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:46 crc kubenswrapper[4897]: I1121 14:09:46.113256 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:46 crc kubenswrapper[4897]: I1121 14:09:46.113269 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:46Z","lastTransitionTime":"2025-11-21T14:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:46 crc kubenswrapper[4897]: E1121 14:09:46.128832 4897 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8b8bdf9e-0fef-41ac-98ad-710dc2137183\\\",\\\"systemUUID\\\":\\\"8a210af6-a71f-4ea9-90a3-aac32fb868b7\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:46Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:46 crc kubenswrapper[4897]: I1121 14:09:46.132644 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:46 crc kubenswrapper[4897]: I1121 14:09:46.132706 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:46 crc kubenswrapper[4897]: I1121 14:09:46.132720 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:46 crc kubenswrapper[4897]: I1121 14:09:46.132744 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:46 crc kubenswrapper[4897]: I1121 14:09:46.132760 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:46Z","lastTransitionTime":"2025-11-21T14:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:46 crc kubenswrapper[4897]: E1121 14:09:46.145183 4897 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8b8bdf9e-0fef-41ac-98ad-710dc2137183\\\",\\\"systemUUID\\\":\\\"8a210af6-a71f-4ea9-90a3-aac32fb868b7\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:46Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:46 crc kubenswrapper[4897]: I1121 14:09:46.150002 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:46 crc kubenswrapper[4897]: I1121 14:09:46.150085 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:46 crc kubenswrapper[4897]: I1121 14:09:46.150103 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:46 crc kubenswrapper[4897]: I1121 14:09:46.150130 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:46 crc kubenswrapper[4897]: I1121 14:09:46.150148 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:46Z","lastTransitionTime":"2025-11-21T14:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:46 crc kubenswrapper[4897]: E1121 14:09:46.173471 4897 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8b8bdf9e-0fef-41ac-98ad-710dc2137183\\\",\\\"systemUUID\\\":\\\"8a210af6-a71f-4ea9-90a3-aac32fb868b7\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:46Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:46 crc kubenswrapper[4897]: I1121 14:09:46.178335 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:46 crc kubenswrapper[4897]: I1121 14:09:46.178385 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:46 crc kubenswrapper[4897]: I1121 14:09:46.178404 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:46 crc kubenswrapper[4897]: I1121 14:09:46.178426 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:46 crc kubenswrapper[4897]: I1121 14:09:46.178439 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:46Z","lastTransitionTime":"2025-11-21T14:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:46 crc kubenswrapper[4897]: E1121 14:09:46.192371 4897 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8b8bdf9e-0fef-41ac-98ad-710dc2137183\\\",\\\"systemUUID\\\":\\\"8a210af6-a71f-4ea9-90a3-aac32fb868b7\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:46Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:46 crc kubenswrapper[4897]: I1121 14:09:46.196186 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:46 crc kubenswrapper[4897]: I1121 14:09:46.196240 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:46 crc kubenswrapper[4897]: I1121 14:09:46.196251 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:46 crc kubenswrapper[4897]: I1121 14:09:46.196268 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:46 crc kubenswrapper[4897]: I1121 14:09:46.196279 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:46Z","lastTransitionTime":"2025-11-21T14:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:46 crc kubenswrapper[4897]: E1121 14:09:46.210432 4897 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8b8bdf9e-0fef-41ac-98ad-710dc2137183\\\",\\\"systemUUID\\\":\\\"8a210af6-a71f-4ea9-90a3-aac32fb868b7\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:46Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:46 crc kubenswrapper[4897]: E1121 14:09:46.210582 4897 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 21 14:09:46 crc kubenswrapper[4897]: I1121 14:09:46.212376 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:46 crc kubenswrapper[4897]: I1121 14:09:46.212436 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:46 crc kubenswrapper[4897]: I1121 14:09:46.212448 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:46 crc kubenswrapper[4897]: I1121 14:09:46.212463 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:46 crc kubenswrapper[4897]: I1121 14:09:46.212474 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:46Z","lastTransitionTime":"2025-11-21T14:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:46 crc kubenswrapper[4897]: I1121 14:09:46.315275 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:46 crc kubenswrapper[4897]: I1121 14:09:46.315334 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:46 crc kubenswrapper[4897]: I1121 14:09:46.315343 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:46 crc kubenswrapper[4897]: I1121 14:09:46.315363 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:46 crc kubenswrapper[4897]: I1121 14:09:46.315411 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:46Z","lastTransitionTime":"2025-11-21T14:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:46 crc kubenswrapper[4897]: I1121 14:09:46.417326 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:46 crc kubenswrapper[4897]: I1121 14:09:46.417386 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:46 crc kubenswrapper[4897]: I1121 14:09:46.417398 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:46 crc kubenswrapper[4897]: I1121 14:09:46.417414 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:46 crc kubenswrapper[4897]: I1121 14:09:46.417424 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:46Z","lastTransitionTime":"2025-11-21T14:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:46 crc kubenswrapper[4897]: I1121 14:09:46.520561 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:46 crc kubenswrapper[4897]: I1121 14:09:46.520608 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:46 crc kubenswrapper[4897]: I1121 14:09:46.520620 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:46 crc kubenswrapper[4897]: I1121 14:09:46.520639 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:46 crc kubenswrapper[4897]: I1121 14:09:46.520652 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:46Z","lastTransitionTime":"2025-11-21T14:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:46 crc kubenswrapper[4897]: I1121 14:09:46.623894 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:46 crc kubenswrapper[4897]: I1121 14:09:46.623987 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:46 crc kubenswrapper[4897]: I1121 14:09:46.624007 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:46 crc kubenswrapper[4897]: I1121 14:09:46.624059 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:46 crc kubenswrapper[4897]: I1121 14:09:46.624078 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:46Z","lastTransitionTime":"2025-11-21T14:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:46 crc kubenswrapper[4897]: I1121 14:09:46.726560 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:46 crc kubenswrapper[4897]: I1121 14:09:46.726604 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:46 crc kubenswrapper[4897]: I1121 14:09:46.726616 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:46 crc kubenswrapper[4897]: I1121 14:09:46.726631 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:46 crc kubenswrapper[4897]: I1121 14:09:46.726642 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:46Z","lastTransitionTime":"2025-11-21T14:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:46 crc kubenswrapper[4897]: I1121 14:09:46.829852 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:46 crc kubenswrapper[4897]: I1121 14:09:46.830036 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:46 crc kubenswrapper[4897]: I1121 14:09:46.830056 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:46 crc kubenswrapper[4897]: I1121 14:09:46.830079 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:46 crc kubenswrapper[4897]: I1121 14:09:46.830092 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:46Z","lastTransitionTime":"2025-11-21T14:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:46 crc kubenswrapper[4897]: I1121 14:09:46.933812 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:46 crc kubenswrapper[4897]: I1121 14:09:46.933886 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:46 crc kubenswrapper[4897]: I1121 14:09:46.933898 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:46 crc kubenswrapper[4897]: I1121 14:09:46.933916 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:46 crc kubenswrapper[4897]: I1121 14:09:46.933928 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:46Z","lastTransitionTime":"2025-11-21T14:09:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:47 crc kubenswrapper[4897]: I1121 14:09:47.037122 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:47 crc kubenswrapper[4897]: I1121 14:09:47.037168 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:47 crc kubenswrapper[4897]: I1121 14:09:47.037177 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:47 crc kubenswrapper[4897]: I1121 14:09:47.037192 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:47 crc kubenswrapper[4897]: I1121 14:09:47.037204 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:47Z","lastTransitionTime":"2025-11-21T14:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:47 crc kubenswrapper[4897]: I1121 14:09:47.088616 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 21 14:09:47 crc kubenswrapper[4897]: I1121 14:09:47.088665 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 21 14:09:47 crc kubenswrapper[4897]: E1121 14:09:47.088764 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 21 14:09:47 crc kubenswrapper[4897]: I1121 14:09:47.088754 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 21 14:09:47 crc kubenswrapper[4897]: I1121 14:09:47.088840 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs4c8" Nov 21 14:09:47 crc kubenswrapper[4897]: E1121 14:09:47.088938 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 21 14:09:47 crc kubenswrapper[4897]: E1121 14:09:47.089107 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs4c8" podUID="01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd" Nov 21 14:09:47 crc kubenswrapper[4897]: E1121 14:09:47.089158 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 21 14:09:47 crc kubenswrapper[4897]: I1121 14:09:47.139789 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:47 crc kubenswrapper[4897]: I1121 14:09:47.139835 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:47 crc kubenswrapper[4897]: I1121 14:09:47.139846 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:47 crc kubenswrapper[4897]: I1121 14:09:47.139863 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:47 crc kubenswrapper[4897]: I1121 14:09:47.139874 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:47Z","lastTransitionTime":"2025-11-21T14:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:47 crc kubenswrapper[4897]: I1121 14:09:47.241846 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:47 crc kubenswrapper[4897]: I1121 14:09:47.241902 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:47 crc kubenswrapper[4897]: I1121 14:09:47.241917 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:47 crc kubenswrapper[4897]: I1121 14:09:47.241938 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:47 crc kubenswrapper[4897]: I1121 14:09:47.241957 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:47Z","lastTransitionTime":"2025-11-21T14:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:47 crc kubenswrapper[4897]: I1121 14:09:47.345101 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:47 crc kubenswrapper[4897]: I1121 14:09:47.345161 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:47 crc kubenswrapper[4897]: I1121 14:09:47.345174 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:47 crc kubenswrapper[4897]: I1121 14:09:47.345194 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:47 crc kubenswrapper[4897]: I1121 14:09:47.345208 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:47Z","lastTransitionTime":"2025-11-21T14:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:47 crc kubenswrapper[4897]: I1121 14:09:47.447725 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:47 crc kubenswrapper[4897]: I1121 14:09:47.447762 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:47 crc kubenswrapper[4897]: I1121 14:09:47.447771 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:47 crc kubenswrapper[4897]: I1121 14:09:47.447785 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:47 crc kubenswrapper[4897]: I1121 14:09:47.447797 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:47Z","lastTransitionTime":"2025-11-21T14:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:47 crc kubenswrapper[4897]: I1121 14:09:47.550315 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:47 crc kubenswrapper[4897]: I1121 14:09:47.550375 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:47 crc kubenswrapper[4897]: I1121 14:09:47.550384 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:47 crc kubenswrapper[4897]: I1121 14:09:47.550401 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:47 crc kubenswrapper[4897]: I1121 14:09:47.550414 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:47Z","lastTransitionTime":"2025-11-21T14:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:47 crc kubenswrapper[4897]: I1121 14:09:47.653371 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:47 crc kubenswrapper[4897]: I1121 14:09:47.653418 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:47 crc kubenswrapper[4897]: I1121 14:09:47.653430 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:47 crc kubenswrapper[4897]: I1121 14:09:47.653448 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:47 crc kubenswrapper[4897]: I1121 14:09:47.653460 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:47Z","lastTransitionTime":"2025-11-21T14:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:47 crc kubenswrapper[4897]: I1121 14:09:47.756796 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:47 crc kubenswrapper[4897]: I1121 14:09:47.756865 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:47 crc kubenswrapper[4897]: I1121 14:09:47.756875 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:47 crc kubenswrapper[4897]: I1121 14:09:47.756896 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:47 crc kubenswrapper[4897]: I1121 14:09:47.756909 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:47Z","lastTransitionTime":"2025-11-21T14:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:47 crc kubenswrapper[4897]: I1121 14:09:47.859555 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:47 crc kubenswrapper[4897]: I1121 14:09:47.859606 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:47 crc kubenswrapper[4897]: I1121 14:09:47.859618 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:47 crc kubenswrapper[4897]: I1121 14:09:47.859638 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:47 crc kubenswrapper[4897]: I1121 14:09:47.859650 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:47Z","lastTransitionTime":"2025-11-21T14:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:47 crc kubenswrapper[4897]: I1121 14:09:47.962797 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:47 crc kubenswrapper[4897]: I1121 14:09:47.962845 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:47 crc kubenswrapper[4897]: I1121 14:09:47.962856 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:47 crc kubenswrapper[4897]: I1121 14:09:47.962875 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:47 crc kubenswrapper[4897]: I1121 14:09:47.962888 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:47Z","lastTransitionTime":"2025-11-21T14:09:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:48 crc kubenswrapper[4897]: I1121 14:09:48.066110 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:48 crc kubenswrapper[4897]: I1121 14:09:48.066157 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:48 crc kubenswrapper[4897]: I1121 14:09:48.066170 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:48 crc kubenswrapper[4897]: I1121 14:09:48.066186 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:48 crc kubenswrapper[4897]: I1121 14:09:48.066198 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:48Z","lastTransitionTime":"2025-11-21T14:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:48 crc kubenswrapper[4897]: I1121 14:09:48.168715 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:48 crc kubenswrapper[4897]: I1121 14:09:48.168765 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:48 crc kubenswrapper[4897]: I1121 14:09:48.168777 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:48 crc kubenswrapper[4897]: I1121 14:09:48.168793 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:48 crc kubenswrapper[4897]: I1121 14:09:48.168803 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:48Z","lastTransitionTime":"2025-11-21T14:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:48 crc kubenswrapper[4897]: I1121 14:09:48.271333 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:48 crc kubenswrapper[4897]: I1121 14:09:48.271387 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:48 crc kubenswrapper[4897]: I1121 14:09:48.271400 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:48 crc kubenswrapper[4897]: I1121 14:09:48.271417 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:48 crc kubenswrapper[4897]: I1121 14:09:48.271431 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:48Z","lastTransitionTime":"2025-11-21T14:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:48 crc kubenswrapper[4897]: I1121 14:09:48.376776 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:48 crc kubenswrapper[4897]: I1121 14:09:48.376880 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:48 crc kubenswrapper[4897]: I1121 14:09:48.376890 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:48 crc kubenswrapper[4897]: I1121 14:09:48.376905 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:48 crc kubenswrapper[4897]: I1121 14:09:48.376915 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:48Z","lastTransitionTime":"2025-11-21T14:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:48 crc kubenswrapper[4897]: I1121 14:09:48.479748 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:48 crc kubenswrapper[4897]: I1121 14:09:48.479787 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:48 crc kubenswrapper[4897]: I1121 14:09:48.479795 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:48 crc kubenswrapper[4897]: I1121 14:09:48.479812 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:48 crc kubenswrapper[4897]: I1121 14:09:48.479823 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:48Z","lastTransitionTime":"2025-11-21T14:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:48 crc kubenswrapper[4897]: I1121 14:09:48.582960 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:48 crc kubenswrapper[4897]: I1121 14:09:48.583019 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:48 crc kubenswrapper[4897]: I1121 14:09:48.583031 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:48 crc kubenswrapper[4897]: I1121 14:09:48.583047 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:48 crc kubenswrapper[4897]: I1121 14:09:48.583058 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:48Z","lastTransitionTime":"2025-11-21T14:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:48 crc kubenswrapper[4897]: I1121 14:09:48.687439 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:48 crc kubenswrapper[4897]: I1121 14:09:48.687550 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:48 crc kubenswrapper[4897]: I1121 14:09:48.687574 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:48 crc kubenswrapper[4897]: I1121 14:09:48.687606 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:48 crc kubenswrapper[4897]: I1121 14:09:48.687624 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:48Z","lastTransitionTime":"2025-11-21T14:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:48 crc kubenswrapper[4897]: I1121 14:09:48.789759 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:48 crc kubenswrapper[4897]: I1121 14:09:48.789846 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:48 crc kubenswrapper[4897]: I1121 14:09:48.789864 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:48 crc kubenswrapper[4897]: I1121 14:09:48.789892 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:48 crc kubenswrapper[4897]: I1121 14:09:48.789913 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:48Z","lastTransitionTime":"2025-11-21T14:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:48 crc kubenswrapper[4897]: I1121 14:09:48.892845 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:48 crc kubenswrapper[4897]: I1121 14:09:48.892908 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:48 crc kubenswrapper[4897]: I1121 14:09:48.892918 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:48 crc kubenswrapper[4897]: I1121 14:09:48.892936 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:48 crc kubenswrapper[4897]: I1121 14:09:48.892982 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:48Z","lastTransitionTime":"2025-11-21T14:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:48 crc kubenswrapper[4897]: I1121 14:09:48.995934 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:48 crc kubenswrapper[4897]: I1121 14:09:48.995989 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:48 crc kubenswrapper[4897]: I1121 14:09:48.996002 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:48 crc kubenswrapper[4897]: I1121 14:09:48.996018 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:48 crc kubenswrapper[4897]: I1121 14:09:48.996028 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:48Z","lastTransitionTime":"2025-11-21T14:09:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:49 crc kubenswrapper[4897]: I1121 14:09:49.087919 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 21 14:09:49 crc kubenswrapper[4897]: I1121 14:09:49.087983 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs4c8" Nov 21 14:09:49 crc kubenswrapper[4897]: I1121 14:09:49.088042 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 21 14:09:49 crc kubenswrapper[4897]: E1121 14:09:49.088077 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 21 14:09:49 crc kubenswrapper[4897]: I1121 14:09:49.088295 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 21 14:09:49 crc kubenswrapper[4897]: E1121 14:09:49.088463 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 21 14:09:49 crc kubenswrapper[4897]: E1121 14:09:49.088362 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 21 14:09:49 crc kubenswrapper[4897]: E1121 14:09:49.088284 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs4c8" podUID="01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd" Nov 21 14:09:49 crc kubenswrapper[4897]: I1121 14:09:49.098630 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:49 crc kubenswrapper[4897]: I1121 14:09:49.098673 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:49 crc kubenswrapper[4897]: I1121 14:09:49.098684 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:49 crc kubenswrapper[4897]: I1121 14:09:49.098699 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:49 crc kubenswrapper[4897]: I1121 14:09:49.098732 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:49Z","lastTransitionTime":"2025-11-21T14:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:49 crc kubenswrapper[4897]: I1121 14:09:49.201433 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:49 crc kubenswrapper[4897]: I1121 14:09:49.201492 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:49 crc kubenswrapper[4897]: I1121 14:09:49.201527 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:49 crc kubenswrapper[4897]: I1121 14:09:49.201552 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:49 crc kubenswrapper[4897]: I1121 14:09:49.201566 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:49Z","lastTransitionTime":"2025-11-21T14:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:49 crc kubenswrapper[4897]: I1121 14:09:49.304849 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:49 crc kubenswrapper[4897]: I1121 14:09:49.304894 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:49 crc kubenswrapper[4897]: I1121 14:09:49.304904 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:49 crc kubenswrapper[4897]: I1121 14:09:49.304922 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:49 crc kubenswrapper[4897]: I1121 14:09:49.304932 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:49Z","lastTransitionTime":"2025-11-21T14:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:49 crc kubenswrapper[4897]: I1121 14:09:49.407366 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:49 crc kubenswrapper[4897]: I1121 14:09:49.407428 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:49 crc kubenswrapper[4897]: I1121 14:09:49.407451 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:49 crc kubenswrapper[4897]: I1121 14:09:49.407478 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:49 crc kubenswrapper[4897]: I1121 14:09:49.407497 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:49Z","lastTransitionTime":"2025-11-21T14:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:49 crc kubenswrapper[4897]: I1121 14:09:49.510976 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:49 crc kubenswrapper[4897]: I1121 14:09:49.511019 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:49 crc kubenswrapper[4897]: I1121 14:09:49.511032 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:49 crc kubenswrapper[4897]: I1121 14:09:49.511053 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:49 crc kubenswrapper[4897]: I1121 14:09:49.511065 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:49Z","lastTransitionTime":"2025-11-21T14:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:49 crc kubenswrapper[4897]: I1121 14:09:49.614245 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:49 crc kubenswrapper[4897]: I1121 14:09:49.614319 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:49 crc kubenswrapper[4897]: I1121 14:09:49.614333 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:49 crc kubenswrapper[4897]: I1121 14:09:49.614353 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:49 crc kubenswrapper[4897]: I1121 14:09:49.614366 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:49Z","lastTransitionTime":"2025-11-21T14:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:49 crc kubenswrapper[4897]: I1121 14:09:49.716843 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:49 crc kubenswrapper[4897]: I1121 14:09:49.716900 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:49 crc kubenswrapper[4897]: I1121 14:09:49.716913 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:49 crc kubenswrapper[4897]: I1121 14:09:49.716933 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:49 crc kubenswrapper[4897]: I1121 14:09:49.716946 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:49Z","lastTransitionTime":"2025-11-21T14:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:49 crc kubenswrapper[4897]: I1121 14:09:49.819267 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:49 crc kubenswrapper[4897]: I1121 14:09:49.819306 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:49 crc kubenswrapper[4897]: I1121 14:09:49.819316 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:49 crc kubenswrapper[4897]: I1121 14:09:49.819331 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:49 crc kubenswrapper[4897]: I1121 14:09:49.819340 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:49Z","lastTransitionTime":"2025-11-21T14:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:49 crc kubenswrapper[4897]: I1121 14:09:49.921912 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:49 crc kubenswrapper[4897]: I1121 14:09:49.921972 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:49 crc kubenswrapper[4897]: I1121 14:09:49.921985 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:49 crc kubenswrapper[4897]: I1121 14:09:49.922004 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:49 crc kubenswrapper[4897]: I1121 14:09:49.922018 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:49Z","lastTransitionTime":"2025-11-21T14:09:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:50 crc kubenswrapper[4897]: I1121 14:09:50.024690 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:50 crc kubenswrapper[4897]: I1121 14:09:50.024736 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:50 crc kubenswrapper[4897]: I1121 14:09:50.024748 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:50 crc kubenswrapper[4897]: I1121 14:09:50.024765 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:50 crc kubenswrapper[4897]: I1121 14:09:50.024777 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:50Z","lastTransitionTime":"2025-11-21T14:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:50 crc kubenswrapper[4897]: I1121 14:09:50.127923 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:50 crc kubenswrapper[4897]: I1121 14:09:50.127971 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:50 crc kubenswrapper[4897]: I1121 14:09:50.127980 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:50 crc kubenswrapper[4897]: I1121 14:09:50.127998 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:50 crc kubenswrapper[4897]: I1121 14:09:50.128009 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:50Z","lastTransitionTime":"2025-11-21T14:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:50 crc kubenswrapper[4897]: I1121 14:09:50.230293 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:50 crc kubenswrapper[4897]: I1121 14:09:50.230340 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:50 crc kubenswrapper[4897]: I1121 14:09:50.230350 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:50 crc kubenswrapper[4897]: I1121 14:09:50.230368 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:50 crc kubenswrapper[4897]: I1121 14:09:50.230378 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:50Z","lastTransitionTime":"2025-11-21T14:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:50 crc kubenswrapper[4897]: I1121 14:09:50.333079 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:50 crc kubenswrapper[4897]: I1121 14:09:50.333116 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:50 crc kubenswrapper[4897]: I1121 14:09:50.333126 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:50 crc kubenswrapper[4897]: I1121 14:09:50.333141 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:50 crc kubenswrapper[4897]: I1121 14:09:50.333151 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:50Z","lastTransitionTime":"2025-11-21T14:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:50 crc kubenswrapper[4897]: I1121 14:09:50.436126 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:50 crc kubenswrapper[4897]: I1121 14:09:50.436161 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:50 crc kubenswrapper[4897]: I1121 14:09:50.436172 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:50 crc kubenswrapper[4897]: I1121 14:09:50.436186 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:50 crc kubenswrapper[4897]: I1121 14:09:50.436197 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:50Z","lastTransitionTime":"2025-11-21T14:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:50 crc kubenswrapper[4897]: I1121 14:09:50.538598 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:50 crc kubenswrapper[4897]: I1121 14:09:50.538661 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:50 crc kubenswrapper[4897]: I1121 14:09:50.538678 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:50 crc kubenswrapper[4897]: I1121 14:09:50.538703 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:50 crc kubenswrapper[4897]: I1121 14:09:50.538720 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:50Z","lastTransitionTime":"2025-11-21T14:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:50 crc kubenswrapper[4897]: I1121 14:09:50.641980 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:50 crc kubenswrapper[4897]: I1121 14:09:50.642017 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:50 crc kubenswrapper[4897]: I1121 14:09:50.642027 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:50 crc kubenswrapper[4897]: I1121 14:09:50.642042 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:50 crc kubenswrapper[4897]: I1121 14:09:50.642051 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:50Z","lastTransitionTime":"2025-11-21T14:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:50 crc kubenswrapper[4897]: I1121 14:09:50.744973 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:50 crc kubenswrapper[4897]: I1121 14:09:50.745029 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:50 crc kubenswrapper[4897]: I1121 14:09:50.745042 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:50 crc kubenswrapper[4897]: I1121 14:09:50.745060 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:50 crc kubenswrapper[4897]: I1121 14:09:50.745075 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:50Z","lastTransitionTime":"2025-11-21T14:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:50 crc kubenswrapper[4897]: I1121 14:09:50.847656 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:50 crc kubenswrapper[4897]: I1121 14:09:50.847700 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:50 crc kubenswrapper[4897]: I1121 14:09:50.847710 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:50 crc kubenswrapper[4897]: I1121 14:09:50.847725 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:50 crc kubenswrapper[4897]: I1121 14:09:50.847736 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:50Z","lastTransitionTime":"2025-11-21T14:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:50 crc kubenswrapper[4897]: I1121 14:09:50.950788 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:50 crc kubenswrapper[4897]: I1121 14:09:50.950827 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:50 crc kubenswrapper[4897]: I1121 14:09:50.950837 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:50 crc kubenswrapper[4897]: I1121 14:09:50.950852 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:50 crc kubenswrapper[4897]: I1121 14:09:50.950863 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:50Z","lastTransitionTime":"2025-11-21T14:09:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:51 crc kubenswrapper[4897]: I1121 14:09:51.053194 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:51 crc kubenswrapper[4897]: I1121 14:09:51.053244 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:51 crc kubenswrapper[4897]: I1121 14:09:51.053254 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:51 crc kubenswrapper[4897]: I1121 14:09:51.053268 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:51 crc kubenswrapper[4897]: I1121 14:09:51.053277 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:51Z","lastTransitionTime":"2025-11-21T14:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:51 crc kubenswrapper[4897]: I1121 14:09:51.088938 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 21 14:09:51 crc kubenswrapper[4897]: I1121 14:09:51.089030 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 21 14:09:51 crc kubenswrapper[4897]: I1121 14:09:51.089030 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs4c8" Nov 21 14:09:51 crc kubenswrapper[4897]: I1121 14:09:51.089066 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 21 14:09:51 crc kubenswrapper[4897]: E1121 14:09:51.089151 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 21 14:09:51 crc kubenswrapper[4897]: E1121 14:09:51.089365 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 21 14:09:51 crc kubenswrapper[4897]: E1121 14:09:51.089540 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 21 14:09:51 crc kubenswrapper[4897]: E1121 14:09:51.089611 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs4c8" podUID="01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd" Nov 21 14:09:51 crc kubenswrapper[4897]: I1121 14:09:51.156200 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:51 crc kubenswrapper[4897]: I1121 14:09:51.156252 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:51 crc kubenswrapper[4897]: I1121 14:09:51.156264 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:51 crc kubenswrapper[4897]: I1121 14:09:51.156283 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:51 crc kubenswrapper[4897]: I1121 14:09:51.156297 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:51Z","lastTransitionTime":"2025-11-21T14:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:51 crc kubenswrapper[4897]: I1121 14:09:51.258637 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:51 crc kubenswrapper[4897]: I1121 14:09:51.258690 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:51 crc kubenswrapper[4897]: I1121 14:09:51.258703 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:51 crc kubenswrapper[4897]: I1121 14:09:51.258721 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:51 crc kubenswrapper[4897]: I1121 14:09:51.258734 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:51Z","lastTransitionTime":"2025-11-21T14:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:51 crc kubenswrapper[4897]: I1121 14:09:51.361770 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:51 crc kubenswrapper[4897]: I1121 14:09:51.361823 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:51 crc kubenswrapper[4897]: I1121 14:09:51.361832 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:51 crc kubenswrapper[4897]: I1121 14:09:51.361850 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:51 crc kubenswrapper[4897]: I1121 14:09:51.361862 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:51Z","lastTransitionTime":"2025-11-21T14:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:51 crc kubenswrapper[4897]: I1121 14:09:51.465080 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:51 crc kubenswrapper[4897]: I1121 14:09:51.465130 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:51 crc kubenswrapper[4897]: I1121 14:09:51.465141 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:51 crc kubenswrapper[4897]: I1121 14:09:51.465162 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:51 crc kubenswrapper[4897]: I1121 14:09:51.465175 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:51Z","lastTransitionTime":"2025-11-21T14:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:51 crc kubenswrapper[4897]: I1121 14:09:51.568171 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:51 crc kubenswrapper[4897]: I1121 14:09:51.568219 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:51 crc kubenswrapper[4897]: I1121 14:09:51.568232 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:51 crc kubenswrapper[4897]: I1121 14:09:51.568249 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:51 crc kubenswrapper[4897]: I1121 14:09:51.568261 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:51Z","lastTransitionTime":"2025-11-21T14:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:51 crc kubenswrapper[4897]: I1121 14:09:51.670152 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:51 crc kubenswrapper[4897]: I1121 14:09:51.670196 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:51 crc kubenswrapper[4897]: I1121 14:09:51.670208 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:51 crc kubenswrapper[4897]: I1121 14:09:51.670226 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:51 crc kubenswrapper[4897]: I1121 14:09:51.670241 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:51Z","lastTransitionTime":"2025-11-21T14:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:51 crc kubenswrapper[4897]: I1121 14:09:51.773342 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:51 crc kubenswrapper[4897]: I1121 14:09:51.773393 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:51 crc kubenswrapper[4897]: I1121 14:09:51.773405 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:51 crc kubenswrapper[4897]: I1121 14:09:51.773423 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:51 crc kubenswrapper[4897]: I1121 14:09:51.773437 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:51Z","lastTransitionTime":"2025-11-21T14:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:51 crc kubenswrapper[4897]: I1121 14:09:51.876184 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:51 crc kubenswrapper[4897]: I1121 14:09:51.876225 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:51 crc kubenswrapper[4897]: I1121 14:09:51.876236 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:51 crc kubenswrapper[4897]: I1121 14:09:51.876257 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:51 crc kubenswrapper[4897]: I1121 14:09:51.876269 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:51Z","lastTransitionTime":"2025-11-21T14:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:51 crc kubenswrapper[4897]: I1121 14:09:51.978968 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:51 crc kubenswrapper[4897]: I1121 14:09:51.979036 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:51 crc kubenswrapper[4897]: I1121 14:09:51.979049 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:51 crc kubenswrapper[4897]: I1121 14:09:51.979070 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:51 crc kubenswrapper[4897]: I1121 14:09:51.979084 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:51Z","lastTransitionTime":"2025-11-21T14:09:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:52 crc kubenswrapper[4897]: I1121 14:09:52.082232 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:52 crc kubenswrapper[4897]: I1121 14:09:52.082265 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:52 crc kubenswrapper[4897]: I1121 14:09:52.082273 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:52 crc kubenswrapper[4897]: I1121 14:09:52.082291 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:52 crc kubenswrapper[4897]: I1121 14:09:52.082300 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:52Z","lastTransitionTime":"2025-11-21T14:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:52 crc kubenswrapper[4897]: I1121 14:09:52.097265 4897 scope.go:117] "RemoveContainer" containerID="9e285a47c1193b01d8cd3cb16ca750bea2f24d3ad941dd438f96d5203f9203de" Nov 21 14:09:52 crc kubenswrapper[4897]: E1121 14:09:52.098348 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-tpfb6_openshift-ovn-kubernetes(222f65f2-0b82-4760-bcf7-779244720f01)\"" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" podUID="222f65f2-0b82-4760-bcf7-779244720f01" Nov 21 14:09:52 crc kubenswrapper[4897]: I1121 14:09:52.184853 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:52 crc kubenswrapper[4897]: I1121 14:09:52.184925 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:52 crc kubenswrapper[4897]: I1121 14:09:52.184947 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:52 crc kubenswrapper[4897]: I1121 14:09:52.184973 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:52 crc kubenswrapper[4897]: I1121 14:09:52.184992 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:52Z","lastTransitionTime":"2025-11-21T14:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:52 crc kubenswrapper[4897]: I1121 14:09:52.288002 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:52 crc kubenswrapper[4897]: I1121 14:09:52.288545 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:52 crc kubenswrapper[4897]: I1121 14:09:52.288559 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:52 crc kubenswrapper[4897]: I1121 14:09:52.288578 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:52 crc kubenswrapper[4897]: I1121 14:09:52.288589 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:52Z","lastTransitionTime":"2025-11-21T14:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:52 crc kubenswrapper[4897]: I1121 14:09:52.391112 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:52 crc kubenswrapper[4897]: I1121 14:09:52.391193 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:52 crc kubenswrapper[4897]: I1121 14:09:52.391208 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:52 crc kubenswrapper[4897]: I1121 14:09:52.391230 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:52 crc kubenswrapper[4897]: I1121 14:09:52.391243 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:52Z","lastTransitionTime":"2025-11-21T14:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:52 crc kubenswrapper[4897]: I1121 14:09:52.495116 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:52 crc kubenswrapper[4897]: I1121 14:09:52.495164 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:52 crc kubenswrapper[4897]: I1121 14:09:52.495173 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:52 crc kubenswrapper[4897]: I1121 14:09:52.495194 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:52 crc kubenswrapper[4897]: I1121 14:09:52.495205 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:52Z","lastTransitionTime":"2025-11-21T14:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:52 crc kubenswrapper[4897]: I1121 14:09:52.597644 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:52 crc kubenswrapper[4897]: I1121 14:09:52.597695 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:52 crc kubenswrapper[4897]: I1121 14:09:52.597711 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:52 crc kubenswrapper[4897]: I1121 14:09:52.597731 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:52 crc kubenswrapper[4897]: I1121 14:09:52.597782 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:52Z","lastTransitionTime":"2025-11-21T14:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:52 crc kubenswrapper[4897]: I1121 14:09:52.701172 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:52 crc kubenswrapper[4897]: I1121 14:09:52.701203 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:52 crc kubenswrapper[4897]: I1121 14:09:52.701212 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:52 crc kubenswrapper[4897]: I1121 14:09:52.701227 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:52 crc kubenswrapper[4897]: I1121 14:09:52.701237 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:52Z","lastTransitionTime":"2025-11-21T14:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:52 crc kubenswrapper[4897]: I1121 14:09:52.803348 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:52 crc kubenswrapper[4897]: I1121 14:09:52.803435 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:52 crc kubenswrapper[4897]: I1121 14:09:52.803454 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:52 crc kubenswrapper[4897]: I1121 14:09:52.803524 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:52 crc kubenswrapper[4897]: I1121 14:09:52.803549 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:52Z","lastTransitionTime":"2025-11-21T14:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:52 crc kubenswrapper[4897]: I1121 14:09:52.906070 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:52 crc kubenswrapper[4897]: I1121 14:09:52.906118 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:52 crc kubenswrapper[4897]: I1121 14:09:52.906151 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:52 crc kubenswrapper[4897]: I1121 14:09:52.906169 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:52 crc kubenswrapper[4897]: I1121 14:09:52.906182 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:52Z","lastTransitionTime":"2025-11-21T14:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:53 crc kubenswrapper[4897]: I1121 14:09:53.008897 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:53 crc kubenswrapper[4897]: I1121 14:09:53.008951 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:53 crc kubenswrapper[4897]: I1121 14:09:53.008964 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:53 crc kubenswrapper[4897]: I1121 14:09:53.008989 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:53 crc kubenswrapper[4897]: I1121 14:09:53.009005 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:53Z","lastTransitionTime":"2025-11-21T14:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:53 crc kubenswrapper[4897]: I1121 14:09:53.088922 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 21 14:09:53 crc kubenswrapper[4897]: I1121 14:09:53.088962 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 21 14:09:53 crc kubenswrapper[4897]: I1121 14:09:53.088998 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 21 14:09:53 crc kubenswrapper[4897]: I1121 14:09:53.089099 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs4c8" Nov 21 14:09:53 crc kubenswrapper[4897]: E1121 14:09:53.089198 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 21 14:09:53 crc kubenswrapper[4897]: E1121 14:09:53.089310 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs4c8" podUID="01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd" Nov 21 14:09:53 crc kubenswrapper[4897]: E1121 14:09:53.089467 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 21 14:09:53 crc kubenswrapper[4897]: E1121 14:09:53.089641 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 21 14:09:53 crc kubenswrapper[4897]: I1121 14:09:53.111689 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:53 crc kubenswrapper[4897]: I1121 14:09:53.111756 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:53 crc kubenswrapper[4897]: I1121 14:09:53.111765 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:53 crc kubenswrapper[4897]: I1121 14:09:53.111779 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:53 crc kubenswrapper[4897]: I1121 14:09:53.111789 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:53Z","lastTransitionTime":"2025-11-21T14:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:53 crc kubenswrapper[4897]: I1121 14:09:53.215059 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:53 crc kubenswrapper[4897]: I1121 14:09:53.215126 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:53 crc kubenswrapper[4897]: I1121 14:09:53.215140 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:53 crc kubenswrapper[4897]: I1121 14:09:53.215164 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:53 crc kubenswrapper[4897]: I1121 14:09:53.215176 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:53Z","lastTransitionTime":"2025-11-21T14:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:53 crc kubenswrapper[4897]: I1121 14:09:53.318279 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:53 crc kubenswrapper[4897]: I1121 14:09:53.318333 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:53 crc kubenswrapper[4897]: I1121 14:09:53.318346 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:53 crc kubenswrapper[4897]: I1121 14:09:53.318367 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:53 crc kubenswrapper[4897]: I1121 14:09:53.318380 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:53Z","lastTransitionTime":"2025-11-21T14:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:53 crc kubenswrapper[4897]: I1121 14:09:53.421493 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:53 crc kubenswrapper[4897]: I1121 14:09:53.421567 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:53 crc kubenswrapper[4897]: I1121 14:09:53.421577 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:53 crc kubenswrapper[4897]: I1121 14:09:53.421597 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:53 crc kubenswrapper[4897]: I1121 14:09:53.421609 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:53Z","lastTransitionTime":"2025-11-21T14:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:53 crc kubenswrapper[4897]: I1121 14:09:53.524296 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:53 crc kubenswrapper[4897]: I1121 14:09:53.524350 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:53 crc kubenswrapper[4897]: I1121 14:09:53.524361 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:53 crc kubenswrapper[4897]: I1121 14:09:53.524380 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:53 crc kubenswrapper[4897]: I1121 14:09:53.524396 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:53Z","lastTransitionTime":"2025-11-21T14:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:53 crc kubenswrapper[4897]: I1121 14:09:53.626863 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:53 crc kubenswrapper[4897]: I1121 14:09:53.626925 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:53 crc kubenswrapper[4897]: I1121 14:09:53.626941 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:53 crc kubenswrapper[4897]: I1121 14:09:53.626963 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:53 crc kubenswrapper[4897]: I1121 14:09:53.626982 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:53Z","lastTransitionTime":"2025-11-21T14:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:53 crc kubenswrapper[4897]: I1121 14:09:53.729375 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:53 crc kubenswrapper[4897]: I1121 14:09:53.729436 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:53 crc kubenswrapper[4897]: I1121 14:09:53.729455 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:53 crc kubenswrapper[4897]: I1121 14:09:53.729476 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:53 crc kubenswrapper[4897]: I1121 14:09:53.729490 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:53Z","lastTransitionTime":"2025-11-21T14:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:53 crc kubenswrapper[4897]: I1121 14:09:53.833186 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:53 crc kubenswrapper[4897]: I1121 14:09:53.833232 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:53 crc kubenswrapper[4897]: I1121 14:09:53.833242 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:53 crc kubenswrapper[4897]: I1121 14:09:53.833260 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:53 crc kubenswrapper[4897]: I1121 14:09:53.833271 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:53Z","lastTransitionTime":"2025-11-21T14:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:53 crc kubenswrapper[4897]: I1121 14:09:53.935803 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:53 crc kubenswrapper[4897]: I1121 14:09:53.935861 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:53 crc kubenswrapper[4897]: I1121 14:09:53.935876 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:53 crc kubenswrapper[4897]: I1121 14:09:53.935908 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:53 crc kubenswrapper[4897]: I1121 14:09:53.935934 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:53Z","lastTransitionTime":"2025-11-21T14:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:54 crc kubenswrapper[4897]: I1121 14:09:54.039403 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:54 crc kubenswrapper[4897]: I1121 14:09:54.039446 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:54 crc kubenswrapper[4897]: I1121 14:09:54.039459 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:54 crc kubenswrapper[4897]: I1121 14:09:54.039472 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:54 crc kubenswrapper[4897]: I1121 14:09:54.039482 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:54Z","lastTransitionTime":"2025-11-21T14:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:54 crc kubenswrapper[4897]: I1121 14:09:54.103406 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48c6b1ea7497b33c00b956e400426c1d4f3c6c66e0225133434460800eef9314\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ea6068a6b6f8123b402de354e5374a0f02a7191a922d1718ab38f8898f47c52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:54Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:54 crc kubenswrapper[4897]: I1121 14:09:54.116543 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4qwqj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f46e2ba8-9ee0-406d-a668-2ba424618b69\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f43aa89d4d9defcb50d0dc6f90293b5f426b9cfc398b780c2e10f6a5c9f1e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbsgb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4qwqj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:54Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:54 crc kubenswrapper[4897]: I1121 14:09:54.132130 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-rs4c8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7xkdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7xkdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:23Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-rs4c8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:54Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:54 crc kubenswrapper[4897]: I1121 14:09:54.141308 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:54 crc kubenswrapper[4897]: I1121 14:09:54.141342 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:54 crc kubenswrapper[4897]: I1121 14:09:54.141353 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:54 crc kubenswrapper[4897]: I1121 14:09:54.141369 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:54 crc kubenswrapper[4897]: I1121 14:09:54.141382 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:54Z","lastTransitionTime":"2025-11-21T14:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:54 crc kubenswrapper[4897]: I1121 14:09:54.143559 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7qfmq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"223b7002-3e67-4848-b1fb-db2921deb64e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381a46c6290da2b36e572f46c16f1a50aa3e25359bcfbc02c7a5a18e1484860e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2m27\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7qfmq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:54Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:54 crc kubenswrapper[4897]: I1121 14:09:54.154149 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"402271c9-35e5-4bed-8410-c84c2e132878\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08651f8381818947818e0bb640c089be4a07022efc650befbb65ad58466bf8c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://419ab9403696491ef53a3d812d4c19d676f8fb1a1a4dfcdfcc0c6255377d2918\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://acc9629bb6b0dfbe5b4efe2f9f980e4e3865fdc6742782ebc49a5a7e0f775b9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0fc71bebc463919826d01260f38b63bfdb01199fd8dd6ae90d9ac9cd75db703\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0fc71bebc463919826d01260f38b63bfdb01199fd8dd6ae90d9ac9cd75db703\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:54Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:54 crc kubenswrapper[4897]: I1121 14:09:54.165969 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:54Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:54 crc kubenswrapper[4897]: I1121 14:09:54.179464 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7670227-d280-4847-b882-754429f56b0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://961961696b8750c248598f4cda032b18395e5c802275dcfe6d3e9f7ef3510f1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5854028821e6a360a259427b68e7bb4c2a7220e0c68f4cc91639b2ece0540a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-krv5b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:54Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:54 crc kubenswrapper[4897]: I1121 14:09:54.201406 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"222f65f2-0b82-4760-bcf7-779244720f01\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92f1d932a83d65ba6a9a906cc489ff51e76547d07d42656630c59421f833c4db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccd51aa0faf3b1327f207b355cddf86a7353c0051f7e5ce0396f3c4bc67e3b16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8971c38daad45bda3d61545f52120dad40992dbf2e1d6e1e776500ed58d73f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfee19f82e35869a641318a0c45bc7a2ae71d80039f5790e60e277be15c9d3ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec279f2f558a8ac653006321b18cbaf13185876fe303c5cc1c65988aef09edbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de095f3c760b13d5197f624eac3806a1f4186577f2b95eb522532d453eb266e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e285a47c1193b01d8cd3cb16ca750bea2f24d3ad941dd438f96d5203f9203de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e285a47c1193b01d8cd3cb16ca750bea2f24d3ad941dd438f96d5203f9203de\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-21T14:09:38Z\\\",\\\"message\\\":\\\":776] Recording success event on pod openshift-etcd/etcd-crc\\\\nI1121 14:09:38.005810 6569 default_network_controller.go:776] Recording success event on pod openshift-image-registry/node-ca-7qfmq\\\\nI1121 14:09:38.005827 6569 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-additional-cni-plugins-nl76g\\\\nI1121 14:09:38.005832 6569 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-additional-cni-plugins-nl76g\\\\nI1121 14:09:38.005837 6569 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-additional-cni-plugins-nl76g in node crc\\\\nI1121 14:09:38.005842 6569 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-additional-cni-plugins-nl76g after 0 failed attempt(s)\\\\nI1121 14:09:38.005846 6569 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-additional-cni-plugins-nl76g\\\\nI1121 14:09:38.005834 6569 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-operator-lifecycle-manager/package-server-manager-metrics]} name:Service_openshift-operator-lifecycle-manager/package-server-manager-metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:37Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-tpfb6_openshift-ovn-kubernetes(222f65f2-0b82-4760-bcf7-779244720f01)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b486a86e0e97ab409e590f30a1bf5cb9f584a7c67f96ff1236e42ca4f8675b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7877db6f5c2bf0eb31218c64d67bca48f921ff7ac62b4e495f29024c086bf738\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7877db6f5c2bf0eb31218c64d67bca48f921ff7ac62b4e495f29024c086bf738\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tpfb6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:54Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:54 crc kubenswrapper[4897]: I1121 14:09:54.213120 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:54Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:54 crc kubenswrapper[4897]: I1121 14:09:54.225970 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6ljgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f3b70b5-9294-4f69-9528-500d28f34c89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://527cf28abdeef1b42d9c19fb88e3171dc05c4516c3a542a461b5b1daa85c6d25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78jhb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6ljgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:54Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:54 crc kubenswrapper[4897]: I1121 14:09:54.243032 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:54 crc kubenswrapper[4897]: I1121 14:09:54.243061 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:54 crc kubenswrapper[4897]: I1121 14:09:54.243071 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:54 crc kubenswrapper[4897]: I1121 14:09:54.243087 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:54 crc kubenswrapper[4897]: I1121 14:09:54.243099 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:54Z","lastTransitionTime":"2025-11-21T14:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:54 crc kubenswrapper[4897]: I1121 14:09:54.244807 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x5pws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc9e717e-af03-48bf-b005-62b3d084c6fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f968eee1d8a1888229e23abda59fd58fa5e86f02ca96ff80fbe093904b27e8df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9c4f5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0f99b9213902c003e6ca754134d2482404ace39c184fb3b431d2b600481d5b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9c4f5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-x5pws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:54Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:54 crc kubenswrapper[4897]: I1121 14:09:54.266902 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6cd330fd-e8a6-49d5-bfc9-5f6122ccd49e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1bd64b9b938bc80284bcee54141e90cad613d339c4f27ef2934b29bc2a318916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://936f59996d26e42099dcd055f87fff9e3ac4283b57e1956fe97463f2334152d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca74b28a069dc2e609b0779164c3aa470a15ed9848af0dd5fc3479aaf750157b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://270c53b9a342f26be077a01cd4e87e9ccd0eb5d2c84208ae38cbf720463245e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66c521c26d541bbddb064dd9710f1a7b41090d2f47343977922272dfc6dab7af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6eb2ac410b49098e5b3603de5e6bf5ed4484fab0b300021d0bed2e5ed8a42527\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6eb2ac410b49098e5b3603de5e6bf5ed4484fab0b300021d0bed2e5ed8a42527\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03385e0c78e4806f60c160227ef4905a11003ed3d70f0e5c0dd0a477979d1d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03385e0c78e4806f60c160227ef4905a11003ed3d70f0e5c0dd0a477979d1d4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://50aea2fa8b3b141b4330ddce2c8e0715230c6f06dab64b3abfd98a3067224bec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50aea2fa8b3b141b4330ddce2c8e0715230c6f06dab64b3abfd98a3067224bec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:54Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:54 crc kubenswrapper[4897]: I1121 14:09:54.281574 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a6c0a99-565b-41f9-a8b5-c8990a0c8eb3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://737553dd1d16e2019ea89a376e3a868f9792ed5239aa7d6cc40a05d3d4f286ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6682a8886d2b9bce069cf9261d4701ef6907f19574fd035bd89bd05626023e5c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de079956e3d8de0101f3f87dcc120ef23876c4972382639e419b3f34a38cdfca\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://469b5be7467d4a58d3776ec0a967ed911be31b770b49965da4f5d2fb264d8a3d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1907bfbad7eb34da642e30d6aa81820aab457f92347333a364ac12396b0e9dae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"le observer\\\\nW1121 14:09:07.994401 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1121 14:09:07.994601 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1121 14:09:07.995290 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1461162674/tls.crt::/tmp/serving-cert-1461162674/tls.key\\\\\\\"\\\\nI1121 14:09:08.162382 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1121 14:09:08.165301 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1121 14:09:08.165318 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1121 14:09:08.165343 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1121 14:09:08.165348 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1121 14:09:08.172015 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1121 14:09:08.172083 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1121 14:09:08.172089 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1121 14:09:08.172094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1121 14:09:08.172098 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1121 14:09:08.172102 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1121 14:09:08.172106 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1121 14:09:08.172023 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1121 14:09:08.174214 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d78c87c440c7d8e30075f6143111b3f504fb99686484b8f19a9c8d61a55eccc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://101fb592fef069af009c613a1e9c9825024ed205498ab72334211ba90d6f7397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://101fb592fef069af009c613a1e9c9825024ed205498ab72334211ba90d6f7397\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:54Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:54 crc kubenswrapper[4897]: I1121 14:09:54.295537 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b78c993a-9966-4e84-85a1-b4f73aaf17f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce3a2602338a1c1e9380ee55f78360f3d9cf6ccc62678886dc5ae574a72e3d9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c256823f2fb49ecc121f6ef564cd86196f9b4a473e9bdea670173596ef14a823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c65f71f8746044f5d41687c670a0308cd09dc41a6531f9b5720e852b3d240756\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1dbe88e7e6ad63899e107ebb884a16edb9a42b21ae7c86d7fbfaf89785b0d36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:54Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:54 crc kubenswrapper[4897]: I1121 14:09:54.306605 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09877d9780caddd5f7f73e95fdf818a427ed645ed5997984113151f4b4759c13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:54Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:54 crc kubenswrapper[4897]: I1121 14:09:54.319924 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:54Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:54 crc kubenswrapper[4897]: I1121 14:09:54.334553 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c976d804a7ebeba2c879e920cbc389f96f7f5def4b4439e8c8980d5460ecdb65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:54Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:54 crc kubenswrapper[4897]: I1121 14:09:54.345780 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:54 crc kubenswrapper[4897]: I1121 14:09:54.345832 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:54 crc kubenswrapper[4897]: I1121 14:09:54.345843 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:54 crc kubenswrapper[4897]: I1121 14:09:54.345863 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:54 crc kubenswrapper[4897]: I1121 14:09:54.345876 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:54Z","lastTransitionTime":"2025-11-21T14:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:54 crc kubenswrapper[4897]: I1121 14:09:54.348973 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nl76g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a7d6101-d0d2-4828-b3ce-d01040c04640\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac58b6ef1bd08e5f950260c18120c72e50b90a0de182c6ab0a5dbbbe45cf7328\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e56402f62821baf335f378084f4ed7c21b60ccbafe148419a0a9a403f106ad19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e56402f62821baf335f378084f4ed7c21b60ccbafe148419a0a9a403f106ad19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6022ff28e83dafcd846a7f03b090262aa3229a0da89f1984e4b6a054a04ac24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6022ff28e83dafcd846a7f03b090262aa3229a0da89f1984e4b6a054a04ac24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89de56e32b796da11bf89dd598d04fbf2884745ddf0f5d87cccfcaf1c001d8d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89de56e32b796da11bf89dd598d04fbf2884745ddf0f5d87cccfcaf1c001d8d6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2dff3cacaba7f9238f482e67b670ea775e0f96621434b99f4e76c2dfdde0321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2dff3cacaba7f9238f482e67b670ea775e0f96621434b99f4e76c2dfdde0321\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a153389a4137ee580e5118f3248400460f9d014c992a956f508815755f583afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a153389a4137ee580e5118f3248400460f9d014c992a956f508815755f583afb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a8c051ad1c1f2373f79eb0781fa6db489aeea3d52ed6c0bb43eec11520ec2b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a8c051ad1c1f2373f79eb0781fa6db489aeea3d52ed6c0bb43eec11520ec2b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nl76g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:54Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:54 crc kubenswrapper[4897]: I1121 14:09:54.448296 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:54 crc kubenswrapper[4897]: I1121 14:09:54.448347 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:54 crc kubenswrapper[4897]: I1121 14:09:54.448358 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:54 crc kubenswrapper[4897]: I1121 14:09:54.448375 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:54 crc kubenswrapper[4897]: I1121 14:09:54.448386 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:54Z","lastTransitionTime":"2025-11-21T14:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:54 crc kubenswrapper[4897]: I1121 14:09:54.550675 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:54 crc kubenswrapper[4897]: I1121 14:09:54.550730 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:54 crc kubenswrapper[4897]: I1121 14:09:54.550743 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:54 crc kubenswrapper[4897]: I1121 14:09:54.550759 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:54 crc kubenswrapper[4897]: I1121 14:09:54.550769 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:54Z","lastTransitionTime":"2025-11-21T14:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:54 crc kubenswrapper[4897]: I1121 14:09:54.653324 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:54 crc kubenswrapper[4897]: I1121 14:09:54.653364 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:54 crc kubenswrapper[4897]: I1121 14:09:54.653372 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:54 crc kubenswrapper[4897]: I1121 14:09:54.653386 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:54 crc kubenswrapper[4897]: I1121 14:09:54.653396 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:54Z","lastTransitionTime":"2025-11-21T14:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:54 crc kubenswrapper[4897]: I1121 14:09:54.755468 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:54 crc kubenswrapper[4897]: I1121 14:09:54.755531 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:54 crc kubenswrapper[4897]: I1121 14:09:54.755540 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:54 crc kubenswrapper[4897]: I1121 14:09:54.755554 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:54 crc kubenswrapper[4897]: I1121 14:09:54.755564 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:54Z","lastTransitionTime":"2025-11-21T14:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:54 crc kubenswrapper[4897]: I1121 14:09:54.858486 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:54 crc kubenswrapper[4897]: I1121 14:09:54.858565 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:54 crc kubenswrapper[4897]: I1121 14:09:54.858578 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:54 crc kubenswrapper[4897]: I1121 14:09:54.858597 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:54 crc kubenswrapper[4897]: I1121 14:09:54.858608 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:54Z","lastTransitionTime":"2025-11-21T14:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:54 crc kubenswrapper[4897]: I1121 14:09:54.961086 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:54 crc kubenswrapper[4897]: I1121 14:09:54.961169 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:54 crc kubenswrapper[4897]: I1121 14:09:54.961183 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:54 crc kubenswrapper[4897]: I1121 14:09:54.961203 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:54 crc kubenswrapper[4897]: I1121 14:09:54.961216 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:54Z","lastTransitionTime":"2025-11-21T14:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:55 crc kubenswrapper[4897]: I1121 14:09:55.063153 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:55 crc kubenswrapper[4897]: I1121 14:09:55.063190 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:55 crc kubenswrapper[4897]: I1121 14:09:55.063199 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:55 crc kubenswrapper[4897]: I1121 14:09:55.063216 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:55 crc kubenswrapper[4897]: I1121 14:09:55.063227 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:55Z","lastTransitionTime":"2025-11-21T14:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:55 crc kubenswrapper[4897]: I1121 14:09:55.087921 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 21 14:09:55 crc kubenswrapper[4897]: I1121 14:09:55.087960 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 21 14:09:55 crc kubenswrapper[4897]: E1121 14:09:55.088068 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 21 14:09:55 crc kubenswrapper[4897]: I1121 14:09:55.088106 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs4c8" Nov 21 14:09:55 crc kubenswrapper[4897]: I1121 14:09:55.087921 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 21 14:09:55 crc kubenswrapper[4897]: E1121 14:09:55.088383 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs4c8" podUID="01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd" Nov 21 14:09:55 crc kubenswrapper[4897]: E1121 14:09:55.088472 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 21 14:09:55 crc kubenswrapper[4897]: E1121 14:09:55.088608 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 21 14:09:55 crc kubenswrapper[4897]: I1121 14:09:55.165654 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:55 crc kubenswrapper[4897]: I1121 14:09:55.165701 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:55 crc kubenswrapper[4897]: I1121 14:09:55.165711 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:55 crc kubenswrapper[4897]: I1121 14:09:55.165728 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:55 crc kubenswrapper[4897]: I1121 14:09:55.165740 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:55Z","lastTransitionTime":"2025-11-21T14:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:55 crc kubenswrapper[4897]: I1121 14:09:55.268555 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:55 crc kubenswrapper[4897]: I1121 14:09:55.268600 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:55 crc kubenswrapper[4897]: I1121 14:09:55.268611 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:55 crc kubenswrapper[4897]: I1121 14:09:55.268628 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:55 crc kubenswrapper[4897]: I1121 14:09:55.268640 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:55Z","lastTransitionTime":"2025-11-21T14:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:55 crc kubenswrapper[4897]: I1121 14:09:55.371114 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:55 crc kubenswrapper[4897]: I1121 14:09:55.371169 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:55 crc kubenswrapper[4897]: I1121 14:09:55.371180 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:55 crc kubenswrapper[4897]: I1121 14:09:55.371198 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:55 crc kubenswrapper[4897]: I1121 14:09:55.371212 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:55Z","lastTransitionTime":"2025-11-21T14:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:55 crc kubenswrapper[4897]: I1121 14:09:55.474018 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:55 crc kubenswrapper[4897]: I1121 14:09:55.474051 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:55 crc kubenswrapper[4897]: I1121 14:09:55.474060 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:55 crc kubenswrapper[4897]: I1121 14:09:55.474073 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:55 crc kubenswrapper[4897]: I1121 14:09:55.474083 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:55Z","lastTransitionTime":"2025-11-21T14:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:55 crc kubenswrapper[4897]: I1121 14:09:55.576163 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:55 crc kubenswrapper[4897]: I1121 14:09:55.576210 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:55 crc kubenswrapper[4897]: I1121 14:09:55.576220 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:55 crc kubenswrapper[4897]: I1121 14:09:55.576234 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:55 crc kubenswrapper[4897]: I1121 14:09:55.576245 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:55Z","lastTransitionTime":"2025-11-21T14:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:55 crc kubenswrapper[4897]: I1121 14:09:55.668875 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd-metrics-certs\") pod \"network-metrics-daemon-rs4c8\" (UID: \"01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd\") " pod="openshift-multus/network-metrics-daemon-rs4c8" Nov 21 14:09:55 crc kubenswrapper[4897]: E1121 14:09:55.669065 4897 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 21 14:09:55 crc kubenswrapper[4897]: E1121 14:09:55.669198 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd-metrics-certs podName:01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd nodeName:}" failed. No retries permitted until 2025-11-21 14:10:27.669167827 +0000 UTC m=+104.953761352 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd-metrics-certs") pod "network-metrics-daemon-rs4c8" (UID: "01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 21 14:09:55 crc kubenswrapper[4897]: I1121 14:09:55.678478 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:55 crc kubenswrapper[4897]: I1121 14:09:55.678538 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:55 crc kubenswrapper[4897]: I1121 14:09:55.678547 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:55 crc kubenswrapper[4897]: I1121 14:09:55.678564 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:55 crc kubenswrapper[4897]: I1121 14:09:55.678579 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:55Z","lastTransitionTime":"2025-11-21T14:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:55 crc kubenswrapper[4897]: I1121 14:09:55.782036 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:55 crc kubenswrapper[4897]: I1121 14:09:55.782091 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:55 crc kubenswrapper[4897]: I1121 14:09:55.782103 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:55 crc kubenswrapper[4897]: I1121 14:09:55.782122 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:55 crc kubenswrapper[4897]: I1121 14:09:55.782137 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:55Z","lastTransitionTime":"2025-11-21T14:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:55 crc kubenswrapper[4897]: I1121 14:09:55.885233 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:55 crc kubenswrapper[4897]: I1121 14:09:55.885268 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:55 crc kubenswrapper[4897]: I1121 14:09:55.885277 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:55 crc kubenswrapper[4897]: I1121 14:09:55.885290 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:55 crc kubenswrapper[4897]: I1121 14:09:55.885300 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:55Z","lastTransitionTime":"2025-11-21T14:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:55 crc kubenswrapper[4897]: I1121 14:09:55.988268 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:55 crc kubenswrapper[4897]: I1121 14:09:55.988329 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:55 crc kubenswrapper[4897]: I1121 14:09:55.988341 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:55 crc kubenswrapper[4897]: I1121 14:09:55.988361 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:55 crc kubenswrapper[4897]: I1121 14:09:55.988375 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:55Z","lastTransitionTime":"2025-11-21T14:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:56 crc kubenswrapper[4897]: I1121 14:09:56.090778 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:56 crc kubenswrapper[4897]: I1121 14:09:56.090849 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:56 crc kubenswrapper[4897]: I1121 14:09:56.090866 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:56 crc kubenswrapper[4897]: I1121 14:09:56.090883 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:56 crc kubenswrapper[4897]: I1121 14:09:56.090893 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:56Z","lastTransitionTime":"2025-11-21T14:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:56 crc kubenswrapper[4897]: I1121 14:09:56.193166 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:56 crc kubenswrapper[4897]: I1121 14:09:56.193218 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:56 crc kubenswrapper[4897]: I1121 14:09:56.193228 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:56 crc kubenswrapper[4897]: I1121 14:09:56.193245 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:56 crc kubenswrapper[4897]: I1121 14:09:56.193260 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:56Z","lastTransitionTime":"2025-11-21T14:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:56 crc kubenswrapper[4897]: I1121 14:09:56.296200 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:56 crc kubenswrapper[4897]: I1121 14:09:56.296255 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:56 crc kubenswrapper[4897]: I1121 14:09:56.296268 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:56 crc kubenswrapper[4897]: I1121 14:09:56.296284 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:56 crc kubenswrapper[4897]: I1121 14:09:56.296295 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:56Z","lastTransitionTime":"2025-11-21T14:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:56 crc kubenswrapper[4897]: I1121 14:09:56.399635 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:56 crc kubenswrapper[4897]: I1121 14:09:56.399679 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:56 crc kubenswrapper[4897]: I1121 14:09:56.399697 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:56 crc kubenswrapper[4897]: I1121 14:09:56.399716 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:56 crc kubenswrapper[4897]: I1121 14:09:56.399728 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:56Z","lastTransitionTime":"2025-11-21T14:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:56 crc kubenswrapper[4897]: I1121 14:09:56.502009 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:56 crc kubenswrapper[4897]: I1121 14:09:56.502057 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:56 crc kubenswrapper[4897]: I1121 14:09:56.502068 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:56 crc kubenswrapper[4897]: I1121 14:09:56.502085 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:56 crc kubenswrapper[4897]: I1121 14:09:56.502094 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:56Z","lastTransitionTime":"2025-11-21T14:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:56 crc kubenswrapper[4897]: I1121 14:09:56.594682 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:56 crc kubenswrapper[4897]: I1121 14:09:56.594729 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:56 crc kubenswrapper[4897]: I1121 14:09:56.594742 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:56 crc kubenswrapper[4897]: I1121 14:09:56.594759 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:56 crc kubenswrapper[4897]: I1121 14:09:56.594771 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:56Z","lastTransitionTime":"2025-11-21T14:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:56 crc kubenswrapper[4897]: E1121 14:09:56.607487 4897 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8b8bdf9e-0fef-41ac-98ad-710dc2137183\\\",\\\"systemUUID\\\":\\\"8a210af6-a71f-4ea9-90a3-aac32fb868b7\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:56Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:56 crc kubenswrapper[4897]: I1121 14:09:56.611319 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:56 crc kubenswrapper[4897]: I1121 14:09:56.611365 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:56 crc kubenswrapper[4897]: I1121 14:09:56.611377 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:56 crc kubenswrapper[4897]: I1121 14:09:56.611395 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:56 crc kubenswrapper[4897]: I1121 14:09:56.611409 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:56Z","lastTransitionTime":"2025-11-21T14:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:56 crc kubenswrapper[4897]: E1121 14:09:56.626621 4897 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8b8bdf9e-0fef-41ac-98ad-710dc2137183\\\",\\\"systemUUID\\\":\\\"8a210af6-a71f-4ea9-90a3-aac32fb868b7\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:56Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:56 crc kubenswrapper[4897]: I1121 14:09:56.631285 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:56 crc kubenswrapper[4897]: I1121 14:09:56.631342 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:56 crc kubenswrapper[4897]: I1121 14:09:56.631357 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:56 crc kubenswrapper[4897]: I1121 14:09:56.631376 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:56 crc kubenswrapper[4897]: I1121 14:09:56.631396 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:56Z","lastTransitionTime":"2025-11-21T14:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:56 crc kubenswrapper[4897]: E1121 14:09:56.644875 4897 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8b8bdf9e-0fef-41ac-98ad-710dc2137183\\\",\\\"systemUUID\\\":\\\"8a210af6-a71f-4ea9-90a3-aac32fb868b7\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:56Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:56 crc kubenswrapper[4897]: I1121 14:09:56.649792 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:56 crc kubenswrapper[4897]: I1121 14:09:56.649833 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:56 crc kubenswrapper[4897]: I1121 14:09:56.649860 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:56 crc kubenswrapper[4897]: I1121 14:09:56.649877 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:56 crc kubenswrapper[4897]: I1121 14:09:56.649887 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:56Z","lastTransitionTime":"2025-11-21T14:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:56 crc kubenswrapper[4897]: E1121 14:09:56.664851 4897 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8b8bdf9e-0fef-41ac-98ad-710dc2137183\\\",\\\"systemUUID\\\":\\\"8a210af6-a71f-4ea9-90a3-aac32fb868b7\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:56Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:56 crc kubenswrapper[4897]: I1121 14:09:56.669002 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:56 crc kubenswrapper[4897]: I1121 14:09:56.669064 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:56 crc kubenswrapper[4897]: I1121 14:09:56.669078 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:56 crc kubenswrapper[4897]: I1121 14:09:56.669095 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:56 crc kubenswrapper[4897]: I1121 14:09:56.669110 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:56Z","lastTransitionTime":"2025-11-21T14:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:56 crc kubenswrapper[4897]: E1121 14:09:56.682825 4897 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:09:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8b8bdf9e-0fef-41ac-98ad-710dc2137183\\\",\\\"systemUUID\\\":\\\"8a210af6-a71f-4ea9-90a3-aac32fb868b7\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:56Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:56 crc kubenswrapper[4897]: E1121 14:09:56.682949 4897 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 21 14:09:56 crc kubenswrapper[4897]: I1121 14:09:56.684602 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:56 crc kubenswrapper[4897]: I1121 14:09:56.684636 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:56 crc kubenswrapper[4897]: I1121 14:09:56.684647 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:56 crc kubenswrapper[4897]: I1121 14:09:56.684659 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:56 crc kubenswrapper[4897]: I1121 14:09:56.684668 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:56Z","lastTransitionTime":"2025-11-21T14:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:56 crc kubenswrapper[4897]: I1121 14:09:56.787279 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:56 crc kubenswrapper[4897]: I1121 14:09:56.787329 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:56 crc kubenswrapper[4897]: I1121 14:09:56.787339 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:56 crc kubenswrapper[4897]: I1121 14:09:56.787353 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:56 crc kubenswrapper[4897]: I1121 14:09:56.787364 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:56Z","lastTransitionTime":"2025-11-21T14:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:56 crc kubenswrapper[4897]: I1121 14:09:56.889877 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:56 crc kubenswrapper[4897]: I1121 14:09:56.889913 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:56 crc kubenswrapper[4897]: I1121 14:09:56.889922 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:56 crc kubenswrapper[4897]: I1121 14:09:56.889936 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:56 crc kubenswrapper[4897]: I1121 14:09:56.889945 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:56Z","lastTransitionTime":"2025-11-21T14:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:56 crc kubenswrapper[4897]: I1121 14:09:56.992779 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:56 crc kubenswrapper[4897]: I1121 14:09:56.992830 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:56 crc kubenswrapper[4897]: I1121 14:09:56.992841 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:56 crc kubenswrapper[4897]: I1121 14:09:56.992859 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:56 crc kubenswrapper[4897]: I1121 14:09:56.992872 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:56Z","lastTransitionTime":"2025-11-21T14:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:57 crc kubenswrapper[4897]: I1121 14:09:57.088903 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 21 14:09:57 crc kubenswrapper[4897]: I1121 14:09:57.088932 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 21 14:09:57 crc kubenswrapper[4897]: I1121 14:09:57.088945 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 21 14:09:57 crc kubenswrapper[4897]: I1121 14:09:57.088900 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs4c8" Nov 21 14:09:57 crc kubenswrapper[4897]: E1121 14:09:57.089070 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 21 14:09:57 crc kubenswrapper[4897]: E1121 14:09:57.089216 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs4c8" podUID="01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd" Nov 21 14:09:57 crc kubenswrapper[4897]: E1121 14:09:57.089307 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 21 14:09:57 crc kubenswrapper[4897]: E1121 14:09:57.089397 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 21 14:09:57 crc kubenswrapper[4897]: I1121 14:09:57.094985 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:57 crc kubenswrapper[4897]: I1121 14:09:57.095016 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:57 crc kubenswrapper[4897]: I1121 14:09:57.095027 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:57 crc kubenswrapper[4897]: I1121 14:09:57.095044 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:57 crc kubenswrapper[4897]: I1121 14:09:57.095055 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:57Z","lastTransitionTime":"2025-11-21T14:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:57 crc kubenswrapper[4897]: I1121 14:09:57.197623 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:57 crc kubenswrapper[4897]: I1121 14:09:57.197672 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:57 crc kubenswrapper[4897]: I1121 14:09:57.197683 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:57 crc kubenswrapper[4897]: I1121 14:09:57.197700 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:57 crc kubenswrapper[4897]: I1121 14:09:57.197711 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:57Z","lastTransitionTime":"2025-11-21T14:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:57 crc kubenswrapper[4897]: I1121 14:09:57.299895 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:57 crc kubenswrapper[4897]: I1121 14:09:57.299930 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:57 crc kubenswrapper[4897]: I1121 14:09:57.299940 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:57 crc kubenswrapper[4897]: I1121 14:09:57.299956 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:57 crc kubenswrapper[4897]: I1121 14:09:57.299968 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:57Z","lastTransitionTime":"2025-11-21T14:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:57 crc kubenswrapper[4897]: I1121 14:09:57.403116 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:57 crc kubenswrapper[4897]: I1121 14:09:57.403174 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:57 crc kubenswrapper[4897]: I1121 14:09:57.403191 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:57 crc kubenswrapper[4897]: I1121 14:09:57.403214 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:57 crc kubenswrapper[4897]: I1121 14:09:57.403234 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:57Z","lastTransitionTime":"2025-11-21T14:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:57 crc kubenswrapper[4897]: I1121 14:09:57.508290 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:57 crc kubenswrapper[4897]: I1121 14:09:57.508345 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:57 crc kubenswrapper[4897]: I1121 14:09:57.508357 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:57 crc kubenswrapper[4897]: I1121 14:09:57.508382 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:57 crc kubenswrapper[4897]: I1121 14:09:57.508393 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:57Z","lastTransitionTime":"2025-11-21T14:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:57 crc kubenswrapper[4897]: I1121 14:09:57.610736 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:57 crc kubenswrapper[4897]: I1121 14:09:57.610808 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:57 crc kubenswrapper[4897]: I1121 14:09:57.610818 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:57 crc kubenswrapper[4897]: I1121 14:09:57.610831 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:57 crc kubenswrapper[4897]: I1121 14:09:57.610841 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:57Z","lastTransitionTime":"2025-11-21T14:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:57 crc kubenswrapper[4897]: I1121 14:09:57.714061 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:57 crc kubenswrapper[4897]: I1121 14:09:57.714106 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:57 crc kubenswrapper[4897]: I1121 14:09:57.714116 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:57 crc kubenswrapper[4897]: I1121 14:09:57.714132 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:57 crc kubenswrapper[4897]: I1121 14:09:57.714142 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:57Z","lastTransitionTime":"2025-11-21T14:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:57 crc kubenswrapper[4897]: I1121 14:09:57.816774 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:57 crc kubenswrapper[4897]: I1121 14:09:57.816818 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:57 crc kubenswrapper[4897]: I1121 14:09:57.816828 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:57 crc kubenswrapper[4897]: I1121 14:09:57.816844 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:57 crc kubenswrapper[4897]: I1121 14:09:57.816854 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:57Z","lastTransitionTime":"2025-11-21T14:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:57 crc kubenswrapper[4897]: I1121 14:09:57.920178 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:57 crc kubenswrapper[4897]: I1121 14:09:57.920229 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:57 crc kubenswrapper[4897]: I1121 14:09:57.920240 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:57 crc kubenswrapper[4897]: I1121 14:09:57.920257 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:57 crc kubenswrapper[4897]: I1121 14:09:57.920269 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:57Z","lastTransitionTime":"2025-11-21T14:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:58 crc kubenswrapper[4897]: I1121 14:09:58.022837 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:58 crc kubenswrapper[4897]: I1121 14:09:58.022874 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:58 crc kubenswrapper[4897]: I1121 14:09:58.022883 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:58 crc kubenswrapper[4897]: I1121 14:09:58.022897 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:58 crc kubenswrapper[4897]: I1121 14:09:58.022908 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:58Z","lastTransitionTime":"2025-11-21T14:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:58 crc kubenswrapper[4897]: I1121 14:09:58.125611 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:58 crc kubenswrapper[4897]: I1121 14:09:58.125654 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:58 crc kubenswrapper[4897]: I1121 14:09:58.125663 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:58 crc kubenswrapper[4897]: I1121 14:09:58.125675 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:58 crc kubenswrapper[4897]: I1121 14:09:58.125684 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:58Z","lastTransitionTime":"2025-11-21T14:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:58 crc kubenswrapper[4897]: I1121 14:09:58.229124 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:58 crc kubenswrapper[4897]: I1121 14:09:58.229156 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:58 crc kubenswrapper[4897]: I1121 14:09:58.229163 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:58 crc kubenswrapper[4897]: I1121 14:09:58.229176 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:58 crc kubenswrapper[4897]: I1121 14:09:58.229188 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:58Z","lastTransitionTime":"2025-11-21T14:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:58 crc kubenswrapper[4897]: I1121 14:09:58.331224 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:58 crc kubenswrapper[4897]: I1121 14:09:58.331291 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:58 crc kubenswrapper[4897]: I1121 14:09:58.331303 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:58 crc kubenswrapper[4897]: I1121 14:09:58.331325 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:58 crc kubenswrapper[4897]: I1121 14:09:58.331339 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:58Z","lastTransitionTime":"2025-11-21T14:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:58 crc kubenswrapper[4897]: I1121 14:09:58.433118 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:58 crc kubenswrapper[4897]: I1121 14:09:58.433159 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:58 crc kubenswrapper[4897]: I1121 14:09:58.433169 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:58 crc kubenswrapper[4897]: I1121 14:09:58.433192 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:58 crc kubenswrapper[4897]: I1121 14:09:58.433203 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:58Z","lastTransitionTime":"2025-11-21T14:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:58 crc kubenswrapper[4897]: I1121 14:09:58.535767 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:58 crc kubenswrapper[4897]: I1121 14:09:58.535812 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:58 crc kubenswrapper[4897]: I1121 14:09:58.535825 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:58 crc kubenswrapper[4897]: I1121 14:09:58.535841 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:58 crc kubenswrapper[4897]: I1121 14:09:58.535853 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:58Z","lastTransitionTime":"2025-11-21T14:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:58 crc kubenswrapper[4897]: I1121 14:09:58.638836 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:58 crc kubenswrapper[4897]: I1121 14:09:58.638905 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:58 crc kubenswrapper[4897]: I1121 14:09:58.638928 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:58 crc kubenswrapper[4897]: I1121 14:09:58.638951 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:58 crc kubenswrapper[4897]: I1121 14:09:58.638968 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:58Z","lastTransitionTime":"2025-11-21T14:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:58 crc kubenswrapper[4897]: I1121 14:09:58.741689 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:58 crc kubenswrapper[4897]: I1121 14:09:58.741723 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:58 crc kubenswrapper[4897]: I1121 14:09:58.741733 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:58 crc kubenswrapper[4897]: I1121 14:09:58.741756 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:58 crc kubenswrapper[4897]: I1121 14:09:58.741769 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:58Z","lastTransitionTime":"2025-11-21T14:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:58 crc kubenswrapper[4897]: I1121 14:09:58.844433 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:58 crc kubenswrapper[4897]: I1121 14:09:58.844487 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:58 crc kubenswrapper[4897]: I1121 14:09:58.844522 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:58 crc kubenswrapper[4897]: I1121 14:09:58.844544 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:58 crc kubenswrapper[4897]: I1121 14:09:58.844559 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:58Z","lastTransitionTime":"2025-11-21T14:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:58 crc kubenswrapper[4897]: I1121 14:09:58.947634 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:58 crc kubenswrapper[4897]: I1121 14:09:58.947688 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:58 crc kubenswrapper[4897]: I1121 14:09:58.947699 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:58 crc kubenswrapper[4897]: I1121 14:09:58.947719 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:58 crc kubenswrapper[4897]: I1121 14:09:58.947734 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:58Z","lastTransitionTime":"2025-11-21T14:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:59 crc kubenswrapper[4897]: I1121 14:09:59.050483 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:59 crc kubenswrapper[4897]: I1121 14:09:59.050557 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:59 crc kubenswrapper[4897]: I1121 14:09:59.050568 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:59 crc kubenswrapper[4897]: I1121 14:09:59.050583 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:59 crc kubenswrapper[4897]: I1121 14:09:59.050596 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:59Z","lastTransitionTime":"2025-11-21T14:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:59 crc kubenswrapper[4897]: I1121 14:09:59.088117 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 21 14:09:59 crc kubenswrapper[4897]: I1121 14:09:59.088145 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs4c8" Nov 21 14:09:59 crc kubenswrapper[4897]: I1121 14:09:59.088145 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 21 14:09:59 crc kubenswrapper[4897]: E1121 14:09:59.088282 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 21 14:09:59 crc kubenswrapper[4897]: I1121 14:09:59.088321 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 21 14:09:59 crc kubenswrapper[4897]: E1121 14:09:59.088363 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 21 14:09:59 crc kubenswrapper[4897]: E1121 14:09:59.088463 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs4c8" podUID="01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd" Nov 21 14:09:59 crc kubenswrapper[4897]: E1121 14:09:59.088519 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 21 14:09:59 crc kubenswrapper[4897]: I1121 14:09:59.153452 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:59 crc kubenswrapper[4897]: I1121 14:09:59.153518 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:59 crc kubenswrapper[4897]: I1121 14:09:59.153532 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:59 crc kubenswrapper[4897]: I1121 14:09:59.153549 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:59 crc kubenswrapper[4897]: I1121 14:09:59.153561 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:59Z","lastTransitionTime":"2025-11-21T14:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:59 crc kubenswrapper[4897]: I1121 14:09:59.256056 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:59 crc kubenswrapper[4897]: I1121 14:09:59.256108 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:59 crc kubenswrapper[4897]: I1121 14:09:59.256119 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:59 crc kubenswrapper[4897]: I1121 14:09:59.256137 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:59 crc kubenswrapper[4897]: I1121 14:09:59.256152 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:59Z","lastTransitionTime":"2025-11-21T14:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:59 crc kubenswrapper[4897]: I1121 14:09:59.358780 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:59 crc kubenswrapper[4897]: I1121 14:09:59.358828 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:59 crc kubenswrapper[4897]: I1121 14:09:59.358840 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:59 crc kubenswrapper[4897]: I1121 14:09:59.358856 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:59 crc kubenswrapper[4897]: I1121 14:09:59.358866 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:59Z","lastTransitionTime":"2025-11-21T14:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:59 crc kubenswrapper[4897]: I1121 14:09:59.461285 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:59 crc kubenswrapper[4897]: I1121 14:09:59.461324 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:59 crc kubenswrapper[4897]: I1121 14:09:59.461335 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:59 crc kubenswrapper[4897]: I1121 14:09:59.461357 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:59 crc kubenswrapper[4897]: I1121 14:09:59.461402 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:59Z","lastTransitionTime":"2025-11-21T14:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:59 crc kubenswrapper[4897]: I1121 14:09:59.563950 4897 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-6ljgx_5f3b70b5-9294-4f69-9528-500d28f34c89/kube-multus/0.log" Nov 21 14:09:59 crc kubenswrapper[4897]: I1121 14:09:59.564076 4897 generic.go:334] "Generic (PLEG): container finished" podID="5f3b70b5-9294-4f69-9528-500d28f34c89" containerID="527cf28abdeef1b42d9c19fb88e3171dc05c4516c3a542a461b5b1daa85c6d25" exitCode=1 Nov 21 14:09:59 crc kubenswrapper[4897]: I1121 14:09:59.564118 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:59 crc kubenswrapper[4897]: I1121 14:09:59.564120 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-6ljgx" event={"ID":"5f3b70b5-9294-4f69-9528-500d28f34c89","Type":"ContainerDied","Data":"527cf28abdeef1b42d9c19fb88e3171dc05c4516c3a542a461b5b1daa85c6d25"} Nov 21 14:09:59 crc kubenswrapper[4897]: I1121 14:09:59.564152 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:59 crc kubenswrapper[4897]: I1121 14:09:59.564458 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:59 crc kubenswrapper[4897]: I1121 14:09:59.564481 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:59 crc kubenswrapper[4897]: I1121 14:09:59.564492 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:59Z","lastTransitionTime":"2025-11-21T14:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:59 crc kubenswrapper[4897]: I1121 14:09:59.564632 4897 scope.go:117] "RemoveContainer" containerID="527cf28abdeef1b42d9c19fb88e3171dc05c4516c3a542a461b5b1daa85c6d25" Nov 21 14:09:59 crc kubenswrapper[4897]: I1121 14:09:59.578518 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7670227-d280-4847-b882-754429f56b0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://961961696b8750c248598f4cda032b18395e5c802275dcfe6d3e9f7ef3510f1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5854028821e6a360a259427b68e7bb4c2a7220e0c68f4cc91639b2ece0540a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-krv5b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:59Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:59 crc kubenswrapper[4897]: I1121 14:09:59.604155 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"222f65f2-0b82-4760-bcf7-779244720f01\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92f1d932a83d65ba6a9a906cc489ff51e76547d07d42656630c59421f833c4db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccd51aa0faf3b1327f207b355cddf86a7353c0051f7e5ce0396f3c4bc67e3b16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8971c38daad45bda3d61545f52120dad40992dbf2e1d6e1e776500ed58d73f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfee19f82e35869a641318a0c45bc7a2ae71d80039f5790e60e277be15c9d3ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec279f2f558a8ac653006321b18cbaf13185876fe303c5cc1c65988aef09edbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de095f3c760b13d5197f624eac3806a1f4186577f2b95eb522532d453eb266e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e285a47c1193b01d8cd3cb16ca750bea2f24d3ad941dd438f96d5203f9203de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e285a47c1193b01d8cd3cb16ca750bea2f24d3ad941dd438f96d5203f9203de\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-21T14:09:38Z\\\",\\\"message\\\":\\\":776] Recording success event on pod openshift-etcd/etcd-crc\\\\nI1121 14:09:38.005810 6569 default_network_controller.go:776] Recording success event on pod openshift-image-registry/node-ca-7qfmq\\\\nI1121 14:09:38.005827 6569 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-additional-cni-plugins-nl76g\\\\nI1121 14:09:38.005832 6569 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-additional-cni-plugins-nl76g\\\\nI1121 14:09:38.005837 6569 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-additional-cni-plugins-nl76g in node crc\\\\nI1121 14:09:38.005842 6569 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-additional-cni-plugins-nl76g after 0 failed attempt(s)\\\\nI1121 14:09:38.005846 6569 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-additional-cni-plugins-nl76g\\\\nI1121 14:09:38.005834 6569 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-operator-lifecycle-manager/package-server-manager-metrics]} name:Service_openshift-operator-lifecycle-manager/package-server-manager-metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:37Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-tpfb6_openshift-ovn-kubernetes(222f65f2-0b82-4760-bcf7-779244720f01)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b486a86e0e97ab409e590f30a1bf5cb9f584a7c67f96ff1236e42ca4f8675b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7877db6f5c2bf0eb31218c64d67bca48f921ff7ac62b4e495f29024c086bf738\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7877db6f5c2bf0eb31218c64d67bca48f921ff7ac62b4e495f29024c086bf738\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tpfb6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:59Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:59 crc kubenswrapper[4897]: I1121 14:09:59.618569 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7qfmq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"223b7002-3e67-4848-b1fb-db2921deb64e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381a46c6290da2b36e572f46c16f1a50aa3e25359bcfbc02c7a5a18e1484860e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2m27\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7qfmq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:59Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:59 crc kubenswrapper[4897]: I1121 14:09:59.629896 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"402271c9-35e5-4bed-8410-c84c2e132878\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08651f8381818947818e0bb640c089be4a07022efc650befbb65ad58466bf8c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://419ab9403696491ef53a3d812d4c19d676f8fb1a1a4dfcdfcc0c6255377d2918\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://acc9629bb6b0dfbe5b4efe2f9f980e4e3865fdc6742782ebc49a5a7e0f775b9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0fc71bebc463919826d01260f38b63bfdb01199fd8dd6ae90d9ac9cd75db703\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0fc71bebc463919826d01260f38b63bfdb01199fd8dd6ae90d9ac9cd75db703\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:59Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:59 crc kubenswrapper[4897]: I1121 14:09:59.646222 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:59Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:59 crc kubenswrapper[4897]: I1121 14:09:59.658093 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b78c993a-9966-4e84-85a1-b4f73aaf17f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce3a2602338a1c1e9380ee55f78360f3d9cf6ccc62678886dc5ae574a72e3d9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c256823f2fb49ecc121f6ef564cd86196f9b4a473e9bdea670173596ef14a823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c65f71f8746044f5d41687c670a0308cd09dc41a6531f9b5720e852b3d240756\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1dbe88e7e6ad63899e107ebb884a16edb9a42b21ae7c86d7fbfaf89785b0d36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:59Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:59 crc kubenswrapper[4897]: I1121 14:09:59.666635 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:59 crc kubenswrapper[4897]: I1121 14:09:59.666672 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:59 crc kubenswrapper[4897]: I1121 14:09:59.666684 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:59 crc kubenswrapper[4897]: I1121 14:09:59.666756 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:59 crc kubenswrapper[4897]: I1121 14:09:59.666771 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:59Z","lastTransitionTime":"2025-11-21T14:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:59 crc kubenswrapper[4897]: I1121 14:09:59.670829 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09877d9780caddd5f7f73e95fdf818a427ed645ed5997984113151f4b4759c13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:59Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:59 crc kubenswrapper[4897]: I1121 14:09:59.684642 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:59Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:59 crc kubenswrapper[4897]: I1121 14:09:59.700765 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6ljgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f3b70b5-9294-4f69-9528-500d28f34c89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://527cf28abdeef1b42d9c19fb88e3171dc05c4516c3a542a461b5b1daa85c6d25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://527cf28abdeef1b42d9c19fb88e3171dc05c4516c3a542a461b5b1daa85c6d25\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-21T14:09:58Z\\\",\\\"message\\\":\\\"2025-11-21T14:09:13+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_a63cd511-2e05-4faf-b47b-8037110976d1\\\\n2025-11-21T14:09:13+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_a63cd511-2e05-4faf-b47b-8037110976d1 to /host/opt/cni/bin/\\\\n2025-11-21T14:09:13Z [verbose] multus-daemon started\\\\n2025-11-21T14:09:13Z [verbose] Readiness Indicator file check\\\\n2025-11-21T14:09:58Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78jhb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6ljgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:59Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:59 crc kubenswrapper[4897]: I1121 14:09:59.716327 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x5pws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc9e717e-af03-48bf-b005-62b3d084c6fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f968eee1d8a1888229e23abda59fd58fa5e86f02ca96ff80fbe093904b27e8df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9c4f5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0f99b9213902c003e6ca754134d2482404ace39c184fb3b431d2b600481d5b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9c4f5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-x5pws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:59Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:59 crc kubenswrapper[4897]: I1121 14:09:59.743552 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6cd330fd-e8a6-49d5-bfc9-5f6122ccd49e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1bd64b9b938bc80284bcee54141e90cad613d339c4f27ef2934b29bc2a318916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://936f59996d26e42099dcd055f87fff9e3ac4283b57e1956fe97463f2334152d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca74b28a069dc2e609b0779164c3aa470a15ed9848af0dd5fc3479aaf750157b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://270c53b9a342f26be077a01cd4e87e9ccd0eb5d2c84208ae38cbf720463245e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66c521c26d541bbddb064dd9710f1a7b41090d2f47343977922272dfc6dab7af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6eb2ac410b49098e5b3603de5e6bf5ed4484fab0b300021d0bed2e5ed8a42527\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6eb2ac410b49098e5b3603de5e6bf5ed4484fab0b300021d0bed2e5ed8a42527\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03385e0c78e4806f60c160227ef4905a11003ed3d70f0e5c0dd0a477979d1d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03385e0c78e4806f60c160227ef4905a11003ed3d70f0e5c0dd0a477979d1d4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://50aea2fa8b3b141b4330ddce2c8e0715230c6f06dab64b3abfd98a3067224bec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50aea2fa8b3b141b4330ddce2c8e0715230c6f06dab64b3abfd98a3067224bec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:59Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:59 crc kubenswrapper[4897]: I1121 14:09:59.758052 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a6c0a99-565b-41f9-a8b5-c8990a0c8eb3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://737553dd1d16e2019ea89a376e3a868f9792ed5239aa7d6cc40a05d3d4f286ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6682a8886d2b9bce069cf9261d4701ef6907f19574fd035bd89bd05626023e5c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de079956e3d8de0101f3f87dcc120ef23876c4972382639e419b3f34a38cdfca\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://469b5be7467d4a58d3776ec0a967ed911be31b770b49965da4f5d2fb264d8a3d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1907bfbad7eb34da642e30d6aa81820aab457f92347333a364ac12396b0e9dae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"le observer\\\\nW1121 14:09:07.994401 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1121 14:09:07.994601 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1121 14:09:07.995290 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1461162674/tls.crt::/tmp/serving-cert-1461162674/tls.key\\\\\\\"\\\\nI1121 14:09:08.162382 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1121 14:09:08.165301 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1121 14:09:08.165318 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1121 14:09:08.165343 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1121 14:09:08.165348 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1121 14:09:08.172015 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1121 14:09:08.172083 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1121 14:09:08.172089 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1121 14:09:08.172094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1121 14:09:08.172098 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1121 14:09:08.172102 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1121 14:09:08.172106 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1121 14:09:08.172023 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1121 14:09:08.174214 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d78c87c440c7d8e30075f6143111b3f504fb99686484b8f19a9c8d61a55eccc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://101fb592fef069af009c613a1e9c9825024ed205498ab72334211ba90d6f7397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://101fb592fef069af009c613a1e9c9825024ed205498ab72334211ba90d6f7397\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:59Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:59 crc kubenswrapper[4897]: I1121 14:09:59.768639 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:59 crc kubenswrapper[4897]: I1121 14:09:59.768675 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:59 crc kubenswrapper[4897]: I1121 14:09:59.768682 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:59 crc kubenswrapper[4897]: I1121 14:09:59.768695 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:59 crc kubenswrapper[4897]: I1121 14:09:59.768704 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:59Z","lastTransitionTime":"2025-11-21T14:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:59 crc kubenswrapper[4897]: I1121 14:09:59.774467 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nl76g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a7d6101-d0d2-4828-b3ce-d01040c04640\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac58b6ef1bd08e5f950260c18120c72e50b90a0de182c6ab0a5dbbbe45cf7328\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e56402f62821baf335f378084f4ed7c21b60ccbafe148419a0a9a403f106ad19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e56402f62821baf335f378084f4ed7c21b60ccbafe148419a0a9a403f106ad19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6022ff28e83dafcd846a7f03b090262aa3229a0da89f1984e4b6a054a04ac24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6022ff28e83dafcd846a7f03b090262aa3229a0da89f1984e4b6a054a04ac24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89de56e32b796da11bf89dd598d04fbf2884745ddf0f5d87cccfcaf1c001d8d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89de56e32b796da11bf89dd598d04fbf2884745ddf0f5d87cccfcaf1c001d8d6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2dff3cacaba7f9238f482e67b670ea775e0f96621434b99f4e76c2dfdde0321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2dff3cacaba7f9238f482e67b670ea775e0f96621434b99f4e76c2dfdde0321\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a153389a4137ee580e5118f3248400460f9d014c992a956f508815755f583afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a153389a4137ee580e5118f3248400460f9d014c992a956f508815755f583afb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a8c051ad1c1f2373f79eb0781fa6db489aeea3d52ed6c0bb43eec11520ec2b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a8c051ad1c1f2373f79eb0781fa6db489aeea3d52ed6c0bb43eec11520ec2b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nl76g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:59Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:59 crc kubenswrapper[4897]: I1121 14:09:59.787914 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:59Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:59 crc kubenswrapper[4897]: I1121 14:09:59.802336 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c976d804a7ebeba2c879e920cbc389f96f7f5def4b4439e8c8980d5460ecdb65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:59Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:59 crc kubenswrapper[4897]: I1121 14:09:59.812690 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4qwqj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f46e2ba8-9ee0-406d-a668-2ba424618b69\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f43aa89d4d9defcb50d0dc6f90293b5f426b9cfc398b780c2e10f6a5c9f1e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbsgb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4qwqj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:59Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:59 crc kubenswrapper[4897]: I1121 14:09:59.824092 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-rs4c8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7xkdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7xkdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:23Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-rs4c8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:59Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:59 crc kubenswrapper[4897]: I1121 14:09:59.837093 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48c6b1ea7497b33c00b956e400426c1d4f3c6c66e0225133434460800eef9314\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ea6068a6b6f8123b402de354e5374a0f02a7191a922d1718ab38f8898f47c52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:09:59Z is after 2025-08-24T17:21:41Z" Nov 21 14:09:59 crc kubenswrapper[4897]: I1121 14:09:59.871190 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:59 crc kubenswrapper[4897]: I1121 14:09:59.871224 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:59 crc kubenswrapper[4897]: I1121 14:09:59.871234 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:59 crc kubenswrapper[4897]: I1121 14:09:59.871248 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:59 crc kubenswrapper[4897]: I1121 14:09:59.871257 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:59Z","lastTransitionTime":"2025-11-21T14:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:09:59 crc kubenswrapper[4897]: I1121 14:09:59.973714 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:09:59 crc kubenswrapper[4897]: I1121 14:09:59.973780 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:09:59 crc kubenswrapper[4897]: I1121 14:09:59.973798 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:09:59 crc kubenswrapper[4897]: I1121 14:09:59.973826 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:09:59 crc kubenswrapper[4897]: I1121 14:09:59.973845 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:09:59Z","lastTransitionTime":"2025-11-21T14:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:00 crc kubenswrapper[4897]: I1121 14:10:00.076695 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:00 crc kubenswrapper[4897]: I1121 14:10:00.076746 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:00 crc kubenswrapper[4897]: I1121 14:10:00.076758 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:00 crc kubenswrapper[4897]: I1121 14:10:00.076775 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:00 crc kubenswrapper[4897]: I1121 14:10:00.076788 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:00Z","lastTransitionTime":"2025-11-21T14:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:00 crc kubenswrapper[4897]: I1121 14:10:00.179558 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:00 crc kubenswrapper[4897]: I1121 14:10:00.179610 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:00 crc kubenswrapper[4897]: I1121 14:10:00.179623 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:00 crc kubenswrapper[4897]: I1121 14:10:00.179641 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:00 crc kubenswrapper[4897]: I1121 14:10:00.179653 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:00Z","lastTransitionTime":"2025-11-21T14:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:00 crc kubenswrapper[4897]: I1121 14:10:00.282223 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:00 crc kubenswrapper[4897]: I1121 14:10:00.282268 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:00 crc kubenswrapper[4897]: I1121 14:10:00.282279 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:00 crc kubenswrapper[4897]: I1121 14:10:00.282296 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:00 crc kubenswrapper[4897]: I1121 14:10:00.282309 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:00Z","lastTransitionTime":"2025-11-21T14:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:00 crc kubenswrapper[4897]: I1121 14:10:00.384640 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:00 crc kubenswrapper[4897]: I1121 14:10:00.384695 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:00 crc kubenswrapper[4897]: I1121 14:10:00.384711 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:00 crc kubenswrapper[4897]: I1121 14:10:00.384733 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:00 crc kubenswrapper[4897]: I1121 14:10:00.384750 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:00Z","lastTransitionTime":"2025-11-21T14:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:00 crc kubenswrapper[4897]: I1121 14:10:00.487969 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:00 crc kubenswrapper[4897]: I1121 14:10:00.488004 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:00 crc kubenswrapper[4897]: I1121 14:10:00.488014 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:00 crc kubenswrapper[4897]: I1121 14:10:00.488030 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:00 crc kubenswrapper[4897]: I1121 14:10:00.488043 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:00Z","lastTransitionTime":"2025-11-21T14:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:00 crc kubenswrapper[4897]: I1121 14:10:00.569187 4897 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-6ljgx_5f3b70b5-9294-4f69-9528-500d28f34c89/kube-multus/0.log" Nov 21 14:10:00 crc kubenswrapper[4897]: I1121 14:10:00.569256 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-6ljgx" event={"ID":"5f3b70b5-9294-4f69-9528-500d28f34c89","Type":"ContainerStarted","Data":"bf92ab70c5a3196bf83f427ccf03acf188d1995766a9154bcd6267e3926c3ac8"} Nov 21 14:10:00 crc kubenswrapper[4897]: I1121 14:10:00.584645 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:00Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:00 crc kubenswrapper[4897]: I1121 14:10:00.590672 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:00 crc kubenswrapper[4897]: I1121 14:10:00.590716 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:00 crc kubenswrapper[4897]: I1121 14:10:00.590728 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:00 crc kubenswrapper[4897]: I1121 14:10:00.590744 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:00 crc kubenswrapper[4897]: I1121 14:10:00.590755 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:00Z","lastTransitionTime":"2025-11-21T14:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:00 crc kubenswrapper[4897]: I1121 14:10:00.597803 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c976d804a7ebeba2c879e920cbc389f96f7f5def4b4439e8c8980d5460ecdb65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:00Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:00 crc kubenswrapper[4897]: I1121 14:10:00.613268 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nl76g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a7d6101-d0d2-4828-b3ce-d01040c04640\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac58b6ef1bd08e5f950260c18120c72e50b90a0de182c6ab0a5dbbbe45cf7328\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e56402f62821baf335f378084f4ed7c21b60ccbafe148419a0a9a403f106ad19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e56402f62821baf335f378084f4ed7c21b60ccbafe148419a0a9a403f106ad19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6022ff28e83dafcd846a7f03b090262aa3229a0da89f1984e4b6a054a04ac24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6022ff28e83dafcd846a7f03b090262aa3229a0da89f1984e4b6a054a04ac24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89de56e32b796da11bf89dd598d04fbf2884745ddf0f5d87cccfcaf1c001d8d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89de56e32b796da11bf89dd598d04fbf2884745ddf0f5d87cccfcaf1c001d8d6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2dff3cacaba7f9238f482e67b670ea775e0f96621434b99f4e76c2dfdde0321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2dff3cacaba7f9238f482e67b670ea775e0f96621434b99f4e76c2dfdde0321\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a153389a4137ee580e5118f3248400460f9d014c992a956f508815755f583afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a153389a4137ee580e5118f3248400460f9d014c992a956f508815755f583afb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a8c051ad1c1f2373f79eb0781fa6db489aeea3d52ed6c0bb43eec11520ec2b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a8c051ad1c1f2373f79eb0781fa6db489aeea3d52ed6c0bb43eec11520ec2b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nl76g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:00Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:00 crc kubenswrapper[4897]: I1121 14:10:00.624834 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48c6b1ea7497b33c00b956e400426c1d4f3c6c66e0225133434460800eef9314\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ea6068a6b6f8123b402de354e5374a0f02a7191a922d1718ab38f8898f47c52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:00Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:00 crc kubenswrapper[4897]: I1121 14:10:00.635200 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4qwqj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f46e2ba8-9ee0-406d-a668-2ba424618b69\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f43aa89d4d9defcb50d0dc6f90293b5f426b9cfc398b780c2e10f6a5c9f1e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbsgb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4qwqj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:00Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:00 crc kubenswrapper[4897]: I1121 14:10:00.645397 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-rs4c8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7xkdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7xkdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:23Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-rs4c8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:00Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:00 crc kubenswrapper[4897]: I1121 14:10:00.656986 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"402271c9-35e5-4bed-8410-c84c2e132878\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08651f8381818947818e0bb640c089be4a07022efc650befbb65ad58466bf8c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://419ab9403696491ef53a3d812d4c19d676f8fb1a1a4dfcdfcc0c6255377d2918\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://acc9629bb6b0dfbe5b4efe2f9f980e4e3865fdc6742782ebc49a5a7e0f775b9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0fc71bebc463919826d01260f38b63bfdb01199fd8dd6ae90d9ac9cd75db703\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0fc71bebc463919826d01260f38b63bfdb01199fd8dd6ae90d9ac9cd75db703\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:00Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:00 crc kubenswrapper[4897]: I1121 14:10:00.669408 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:00Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:00 crc kubenswrapper[4897]: I1121 14:10:00.682355 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7670227-d280-4847-b882-754429f56b0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://961961696b8750c248598f4cda032b18395e5c802275dcfe6d3e9f7ef3510f1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5854028821e6a360a259427b68e7bb4c2a7220e0c68f4cc91639b2ece0540a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-krv5b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:00Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:00 crc kubenswrapper[4897]: I1121 14:10:00.693690 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:00 crc kubenswrapper[4897]: I1121 14:10:00.693747 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:00 crc kubenswrapper[4897]: I1121 14:10:00.693764 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:00 crc kubenswrapper[4897]: I1121 14:10:00.693788 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:00 crc kubenswrapper[4897]: I1121 14:10:00.693808 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:00Z","lastTransitionTime":"2025-11-21T14:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:00 crc kubenswrapper[4897]: I1121 14:10:00.701234 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"222f65f2-0b82-4760-bcf7-779244720f01\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92f1d932a83d65ba6a9a906cc489ff51e76547d07d42656630c59421f833c4db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccd51aa0faf3b1327f207b355cddf86a7353c0051f7e5ce0396f3c4bc67e3b16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8971c38daad45bda3d61545f52120dad40992dbf2e1d6e1e776500ed58d73f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfee19f82e35869a641318a0c45bc7a2ae71d80039f5790e60e277be15c9d3ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec279f2f558a8ac653006321b18cbaf13185876fe303c5cc1c65988aef09edbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de095f3c760b13d5197f624eac3806a1f4186577f2b95eb522532d453eb266e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e285a47c1193b01d8cd3cb16ca750bea2f24d3ad941dd438f96d5203f9203de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e285a47c1193b01d8cd3cb16ca750bea2f24d3ad941dd438f96d5203f9203de\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-21T14:09:38Z\\\",\\\"message\\\":\\\":776] Recording success event on pod openshift-etcd/etcd-crc\\\\nI1121 14:09:38.005810 6569 default_network_controller.go:776] Recording success event on pod openshift-image-registry/node-ca-7qfmq\\\\nI1121 14:09:38.005827 6569 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-additional-cni-plugins-nl76g\\\\nI1121 14:09:38.005832 6569 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-additional-cni-plugins-nl76g\\\\nI1121 14:09:38.005837 6569 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-additional-cni-plugins-nl76g in node crc\\\\nI1121 14:09:38.005842 6569 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-additional-cni-plugins-nl76g after 0 failed attempt(s)\\\\nI1121 14:09:38.005846 6569 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-additional-cni-plugins-nl76g\\\\nI1121 14:09:38.005834 6569 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-operator-lifecycle-manager/package-server-manager-metrics]} name:Service_openshift-operator-lifecycle-manager/package-server-manager-metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:37Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-tpfb6_openshift-ovn-kubernetes(222f65f2-0b82-4760-bcf7-779244720f01)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b486a86e0e97ab409e590f30a1bf5cb9f584a7c67f96ff1236e42ca4f8675b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7877db6f5c2bf0eb31218c64d67bca48f921ff7ac62b4e495f29024c086bf738\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7877db6f5c2bf0eb31218c64d67bca48f921ff7ac62b4e495f29024c086bf738\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tpfb6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:00Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:00 crc kubenswrapper[4897]: I1121 14:10:00.712913 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7qfmq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"223b7002-3e67-4848-b1fb-db2921deb64e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381a46c6290da2b36e572f46c16f1a50aa3e25359bcfbc02c7a5a18e1484860e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2m27\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7qfmq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:00Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:00 crc kubenswrapper[4897]: I1121 14:10:00.724836 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x5pws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc9e717e-af03-48bf-b005-62b3d084c6fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f968eee1d8a1888229e23abda59fd58fa5e86f02ca96ff80fbe093904b27e8df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9c4f5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0f99b9213902c003e6ca754134d2482404ace39c184fb3b431d2b600481d5b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9c4f5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-x5pws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:00Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:00 crc kubenswrapper[4897]: I1121 14:10:00.745080 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6cd330fd-e8a6-49d5-bfc9-5f6122ccd49e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1bd64b9b938bc80284bcee54141e90cad613d339c4f27ef2934b29bc2a318916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://936f59996d26e42099dcd055f87fff9e3ac4283b57e1956fe97463f2334152d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca74b28a069dc2e609b0779164c3aa470a15ed9848af0dd5fc3479aaf750157b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://270c53b9a342f26be077a01cd4e87e9ccd0eb5d2c84208ae38cbf720463245e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66c521c26d541bbddb064dd9710f1a7b41090d2f47343977922272dfc6dab7af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6eb2ac410b49098e5b3603de5e6bf5ed4484fab0b300021d0bed2e5ed8a42527\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6eb2ac410b49098e5b3603de5e6bf5ed4484fab0b300021d0bed2e5ed8a42527\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03385e0c78e4806f60c160227ef4905a11003ed3d70f0e5c0dd0a477979d1d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03385e0c78e4806f60c160227ef4905a11003ed3d70f0e5c0dd0a477979d1d4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://50aea2fa8b3b141b4330ddce2c8e0715230c6f06dab64b3abfd98a3067224bec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50aea2fa8b3b141b4330ddce2c8e0715230c6f06dab64b3abfd98a3067224bec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:00Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:00 crc kubenswrapper[4897]: I1121 14:10:00.760441 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a6c0a99-565b-41f9-a8b5-c8990a0c8eb3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://737553dd1d16e2019ea89a376e3a868f9792ed5239aa7d6cc40a05d3d4f286ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6682a8886d2b9bce069cf9261d4701ef6907f19574fd035bd89bd05626023e5c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de079956e3d8de0101f3f87dcc120ef23876c4972382639e419b3f34a38cdfca\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://469b5be7467d4a58d3776ec0a967ed911be31b770b49965da4f5d2fb264d8a3d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1907bfbad7eb34da642e30d6aa81820aab457f92347333a364ac12396b0e9dae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"le observer\\\\nW1121 14:09:07.994401 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1121 14:09:07.994601 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1121 14:09:07.995290 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1461162674/tls.crt::/tmp/serving-cert-1461162674/tls.key\\\\\\\"\\\\nI1121 14:09:08.162382 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1121 14:09:08.165301 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1121 14:09:08.165318 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1121 14:09:08.165343 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1121 14:09:08.165348 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1121 14:09:08.172015 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1121 14:09:08.172083 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1121 14:09:08.172089 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1121 14:09:08.172094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1121 14:09:08.172098 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1121 14:09:08.172102 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1121 14:09:08.172106 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1121 14:09:08.172023 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1121 14:09:08.174214 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d78c87c440c7d8e30075f6143111b3f504fb99686484b8f19a9c8d61a55eccc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://101fb592fef069af009c613a1e9c9825024ed205498ab72334211ba90d6f7397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://101fb592fef069af009c613a1e9c9825024ed205498ab72334211ba90d6f7397\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:00Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:00 crc kubenswrapper[4897]: I1121 14:10:00.774345 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b78c993a-9966-4e84-85a1-b4f73aaf17f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce3a2602338a1c1e9380ee55f78360f3d9cf6ccc62678886dc5ae574a72e3d9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c256823f2fb49ecc121f6ef564cd86196f9b4a473e9bdea670173596ef14a823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c65f71f8746044f5d41687c670a0308cd09dc41a6531f9b5720e852b3d240756\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1dbe88e7e6ad63899e107ebb884a16edb9a42b21ae7c86d7fbfaf89785b0d36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:00Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:00 crc kubenswrapper[4897]: I1121 14:10:00.788559 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09877d9780caddd5f7f73e95fdf818a427ed645ed5997984113151f4b4759c13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:00Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:00 crc kubenswrapper[4897]: I1121 14:10:00.796668 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:00 crc kubenswrapper[4897]: I1121 14:10:00.796715 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:00 crc kubenswrapper[4897]: I1121 14:10:00.796725 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:00 crc kubenswrapper[4897]: I1121 14:10:00.796743 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:00 crc kubenswrapper[4897]: I1121 14:10:00.796754 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:00Z","lastTransitionTime":"2025-11-21T14:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:00 crc kubenswrapper[4897]: I1121 14:10:00.805434 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:00Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:00 crc kubenswrapper[4897]: I1121 14:10:00.818976 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6ljgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f3b70b5-9294-4f69-9528-500d28f34c89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:10:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:10:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf92ab70c5a3196bf83f427ccf03acf188d1995766a9154bcd6267e3926c3ac8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://527cf28abdeef1b42d9c19fb88e3171dc05c4516c3a542a461b5b1daa85c6d25\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-21T14:09:58Z\\\",\\\"message\\\":\\\"2025-11-21T14:09:13+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_a63cd511-2e05-4faf-b47b-8037110976d1\\\\n2025-11-21T14:09:13+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_a63cd511-2e05-4faf-b47b-8037110976d1 to /host/opt/cni/bin/\\\\n2025-11-21T14:09:13Z [verbose] multus-daemon started\\\\n2025-11-21T14:09:13Z [verbose] Readiness Indicator file check\\\\n2025-11-21T14:09:58Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78jhb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6ljgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:00Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:00 crc kubenswrapper[4897]: I1121 14:10:00.899476 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:00 crc kubenswrapper[4897]: I1121 14:10:00.899538 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:00 crc kubenswrapper[4897]: I1121 14:10:00.899553 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:00 crc kubenswrapper[4897]: I1121 14:10:00.899569 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:00 crc kubenswrapper[4897]: I1121 14:10:00.899583 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:00Z","lastTransitionTime":"2025-11-21T14:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:01 crc kubenswrapper[4897]: I1121 14:10:01.002038 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:01 crc kubenswrapper[4897]: I1121 14:10:01.002072 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:01 crc kubenswrapper[4897]: I1121 14:10:01.002080 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:01 crc kubenswrapper[4897]: I1121 14:10:01.002095 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:01 crc kubenswrapper[4897]: I1121 14:10:01.002107 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:01Z","lastTransitionTime":"2025-11-21T14:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:01 crc kubenswrapper[4897]: I1121 14:10:01.087910 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 21 14:10:01 crc kubenswrapper[4897]: I1121 14:10:01.087946 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs4c8" Nov 21 14:10:01 crc kubenswrapper[4897]: I1121 14:10:01.087990 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 21 14:10:01 crc kubenswrapper[4897]: E1121 14:10:01.088039 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 21 14:10:01 crc kubenswrapper[4897]: E1121 14:10:01.088185 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs4c8" podUID="01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd" Nov 21 14:10:01 crc kubenswrapper[4897]: I1121 14:10:01.088223 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 21 14:10:01 crc kubenswrapper[4897]: E1121 14:10:01.088279 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 21 14:10:01 crc kubenswrapper[4897]: E1121 14:10:01.088340 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 21 14:10:01 crc kubenswrapper[4897]: I1121 14:10:01.104679 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:01 crc kubenswrapper[4897]: I1121 14:10:01.104725 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:01 crc kubenswrapper[4897]: I1121 14:10:01.104737 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:01 crc kubenswrapper[4897]: I1121 14:10:01.104754 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:01 crc kubenswrapper[4897]: I1121 14:10:01.104764 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:01Z","lastTransitionTime":"2025-11-21T14:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:01 crc kubenswrapper[4897]: I1121 14:10:01.207064 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:01 crc kubenswrapper[4897]: I1121 14:10:01.207100 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:01 crc kubenswrapper[4897]: I1121 14:10:01.207109 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:01 crc kubenswrapper[4897]: I1121 14:10:01.207122 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:01 crc kubenswrapper[4897]: I1121 14:10:01.207131 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:01Z","lastTransitionTime":"2025-11-21T14:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:01 crc kubenswrapper[4897]: I1121 14:10:01.309469 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:01 crc kubenswrapper[4897]: I1121 14:10:01.309563 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:01 crc kubenswrapper[4897]: I1121 14:10:01.309576 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:01 crc kubenswrapper[4897]: I1121 14:10:01.309593 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:01 crc kubenswrapper[4897]: I1121 14:10:01.309605 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:01Z","lastTransitionTime":"2025-11-21T14:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:01 crc kubenswrapper[4897]: I1121 14:10:01.411345 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:01 crc kubenswrapper[4897]: I1121 14:10:01.411383 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:01 crc kubenswrapper[4897]: I1121 14:10:01.411392 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:01 crc kubenswrapper[4897]: I1121 14:10:01.411405 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:01 crc kubenswrapper[4897]: I1121 14:10:01.411415 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:01Z","lastTransitionTime":"2025-11-21T14:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:01 crc kubenswrapper[4897]: I1121 14:10:01.514143 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:01 crc kubenswrapper[4897]: I1121 14:10:01.514263 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:01 crc kubenswrapper[4897]: I1121 14:10:01.514303 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:01 crc kubenswrapper[4897]: I1121 14:10:01.514349 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:01 crc kubenswrapper[4897]: I1121 14:10:01.514374 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:01Z","lastTransitionTime":"2025-11-21T14:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:01 crc kubenswrapper[4897]: I1121 14:10:01.616210 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:01 crc kubenswrapper[4897]: I1121 14:10:01.616246 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:01 crc kubenswrapper[4897]: I1121 14:10:01.616255 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:01 crc kubenswrapper[4897]: I1121 14:10:01.616272 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:01 crc kubenswrapper[4897]: I1121 14:10:01.616286 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:01Z","lastTransitionTime":"2025-11-21T14:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:01 crc kubenswrapper[4897]: I1121 14:10:01.718706 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:01 crc kubenswrapper[4897]: I1121 14:10:01.718767 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:01 crc kubenswrapper[4897]: I1121 14:10:01.718780 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:01 crc kubenswrapper[4897]: I1121 14:10:01.718800 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:01 crc kubenswrapper[4897]: I1121 14:10:01.718815 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:01Z","lastTransitionTime":"2025-11-21T14:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:01 crc kubenswrapper[4897]: I1121 14:10:01.821646 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:01 crc kubenswrapper[4897]: I1121 14:10:01.821721 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:01 crc kubenswrapper[4897]: I1121 14:10:01.821740 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:01 crc kubenswrapper[4897]: I1121 14:10:01.821764 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:01 crc kubenswrapper[4897]: I1121 14:10:01.821784 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:01Z","lastTransitionTime":"2025-11-21T14:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:01 crc kubenswrapper[4897]: I1121 14:10:01.924570 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:01 crc kubenswrapper[4897]: I1121 14:10:01.925079 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:01 crc kubenswrapper[4897]: I1121 14:10:01.925196 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:01 crc kubenswrapper[4897]: I1121 14:10:01.925304 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:01 crc kubenswrapper[4897]: I1121 14:10:01.925389 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:01Z","lastTransitionTime":"2025-11-21T14:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:02 crc kubenswrapper[4897]: I1121 14:10:02.028658 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:02 crc kubenswrapper[4897]: I1121 14:10:02.028718 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:02 crc kubenswrapper[4897]: I1121 14:10:02.028734 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:02 crc kubenswrapper[4897]: I1121 14:10:02.028754 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:02 crc kubenswrapper[4897]: I1121 14:10:02.028766 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:02Z","lastTransitionTime":"2025-11-21T14:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:02 crc kubenswrapper[4897]: I1121 14:10:02.130982 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:02 crc kubenswrapper[4897]: I1121 14:10:02.131218 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:02 crc kubenswrapper[4897]: I1121 14:10:02.131330 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:02 crc kubenswrapper[4897]: I1121 14:10:02.131408 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:02 crc kubenswrapper[4897]: I1121 14:10:02.131483 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:02Z","lastTransitionTime":"2025-11-21T14:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:02 crc kubenswrapper[4897]: I1121 14:10:02.234404 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:02 crc kubenswrapper[4897]: I1121 14:10:02.234672 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:02 crc kubenswrapper[4897]: I1121 14:10:02.234705 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:02 crc kubenswrapper[4897]: I1121 14:10:02.234729 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:02 crc kubenswrapper[4897]: I1121 14:10:02.234743 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:02Z","lastTransitionTime":"2025-11-21T14:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:02 crc kubenswrapper[4897]: I1121 14:10:02.337311 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:02 crc kubenswrapper[4897]: I1121 14:10:02.337356 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:02 crc kubenswrapper[4897]: I1121 14:10:02.337366 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:02 crc kubenswrapper[4897]: I1121 14:10:02.337381 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:02 crc kubenswrapper[4897]: I1121 14:10:02.337391 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:02Z","lastTransitionTime":"2025-11-21T14:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:02 crc kubenswrapper[4897]: I1121 14:10:02.439441 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:02 crc kubenswrapper[4897]: I1121 14:10:02.439493 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:02 crc kubenswrapper[4897]: I1121 14:10:02.439533 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:02 crc kubenswrapper[4897]: I1121 14:10:02.439549 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:02 crc kubenswrapper[4897]: I1121 14:10:02.439559 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:02Z","lastTransitionTime":"2025-11-21T14:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:02 crc kubenswrapper[4897]: I1121 14:10:02.542712 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:02 crc kubenswrapper[4897]: I1121 14:10:02.542784 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:02 crc kubenswrapper[4897]: I1121 14:10:02.542830 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:02 crc kubenswrapper[4897]: I1121 14:10:02.542848 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:02 crc kubenswrapper[4897]: I1121 14:10:02.542862 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:02Z","lastTransitionTime":"2025-11-21T14:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:02 crc kubenswrapper[4897]: I1121 14:10:02.644750 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:02 crc kubenswrapper[4897]: I1121 14:10:02.644793 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:02 crc kubenswrapper[4897]: I1121 14:10:02.644806 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:02 crc kubenswrapper[4897]: I1121 14:10:02.644822 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:02 crc kubenswrapper[4897]: I1121 14:10:02.644833 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:02Z","lastTransitionTime":"2025-11-21T14:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:02 crc kubenswrapper[4897]: I1121 14:10:02.747927 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:02 crc kubenswrapper[4897]: I1121 14:10:02.747985 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:02 crc kubenswrapper[4897]: I1121 14:10:02.748000 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:02 crc kubenswrapper[4897]: I1121 14:10:02.748021 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:02 crc kubenswrapper[4897]: I1121 14:10:02.748038 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:02Z","lastTransitionTime":"2025-11-21T14:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:02 crc kubenswrapper[4897]: I1121 14:10:02.850961 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:02 crc kubenswrapper[4897]: I1121 14:10:02.851031 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:02 crc kubenswrapper[4897]: I1121 14:10:02.851053 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:02 crc kubenswrapper[4897]: I1121 14:10:02.851085 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:02 crc kubenswrapper[4897]: I1121 14:10:02.851106 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:02Z","lastTransitionTime":"2025-11-21T14:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:02 crc kubenswrapper[4897]: I1121 14:10:02.953886 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:02 crc kubenswrapper[4897]: I1121 14:10:02.953946 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:02 crc kubenswrapper[4897]: I1121 14:10:02.953964 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:02 crc kubenswrapper[4897]: I1121 14:10:02.953987 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:02 crc kubenswrapper[4897]: I1121 14:10:02.954004 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:02Z","lastTransitionTime":"2025-11-21T14:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:03 crc kubenswrapper[4897]: I1121 14:10:03.056742 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:03 crc kubenswrapper[4897]: I1121 14:10:03.056821 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:03 crc kubenswrapper[4897]: I1121 14:10:03.056846 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:03 crc kubenswrapper[4897]: I1121 14:10:03.056874 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:03 crc kubenswrapper[4897]: I1121 14:10:03.056893 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:03Z","lastTransitionTime":"2025-11-21T14:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:03 crc kubenswrapper[4897]: I1121 14:10:03.088651 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 21 14:10:03 crc kubenswrapper[4897]: I1121 14:10:03.088734 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 21 14:10:03 crc kubenswrapper[4897]: I1121 14:10:03.088784 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs4c8" Nov 21 14:10:03 crc kubenswrapper[4897]: E1121 14:10:03.088848 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 21 14:10:03 crc kubenswrapper[4897]: I1121 14:10:03.088878 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 21 14:10:03 crc kubenswrapper[4897]: E1121 14:10:03.088969 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs4c8" podUID="01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd" Nov 21 14:10:03 crc kubenswrapper[4897]: E1121 14:10:03.089119 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 21 14:10:03 crc kubenswrapper[4897]: E1121 14:10:03.089164 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 21 14:10:03 crc kubenswrapper[4897]: I1121 14:10:03.163204 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:03 crc kubenswrapper[4897]: I1121 14:10:03.163268 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:03 crc kubenswrapper[4897]: I1121 14:10:03.163281 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:03 crc kubenswrapper[4897]: I1121 14:10:03.163308 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:03 crc kubenswrapper[4897]: I1121 14:10:03.163319 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:03Z","lastTransitionTime":"2025-11-21T14:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:03 crc kubenswrapper[4897]: I1121 14:10:03.266304 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:03 crc kubenswrapper[4897]: I1121 14:10:03.266340 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:03 crc kubenswrapper[4897]: I1121 14:10:03.266348 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:03 crc kubenswrapper[4897]: I1121 14:10:03.266360 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:03 crc kubenswrapper[4897]: I1121 14:10:03.266370 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:03Z","lastTransitionTime":"2025-11-21T14:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:03 crc kubenswrapper[4897]: I1121 14:10:03.368657 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:03 crc kubenswrapper[4897]: I1121 14:10:03.368690 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:03 crc kubenswrapper[4897]: I1121 14:10:03.368698 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:03 crc kubenswrapper[4897]: I1121 14:10:03.368713 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:03 crc kubenswrapper[4897]: I1121 14:10:03.368722 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:03Z","lastTransitionTime":"2025-11-21T14:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:03 crc kubenswrapper[4897]: I1121 14:10:03.471670 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:03 crc kubenswrapper[4897]: I1121 14:10:03.471727 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:03 crc kubenswrapper[4897]: I1121 14:10:03.471745 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:03 crc kubenswrapper[4897]: I1121 14:10:03.471764 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:03 crc kubenswrapper[4897]: I1121 14:10:03.471777 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:03Z","lastTransitionTime":"2025-11-21T14:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:03 crc kubenswrapper[4897]: I1121 14:10:03.574220 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:03 crc kubenswrapper[4897]: I1121 14:10:03.574301 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:03 crc kubenswrapper[4897]: I1121 14:10:03.574314 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:03 crc kubenswrapper[4897]: I1121 14:10:03.574331 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:03 crc kubenswrapper[4897]: I1121 14:10:03.574342 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:03Z","lastTransitionTime":"2025-11-21T14:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:03 crc kubenswrapper[4897]: I1121 14:10:03.677413 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:03 crc kubenswrapper[4897]: I1121 14:10:03.677472 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:03 crc kubenswrapper[4897]: I1121 14:10:03.677483 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:03 crc kubenswrapper[4897]: I1121 14:10:03.677530 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:03 crc kubenswrapper[4897]: I1121 14:10:03.677544 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:03Z","lastTransitionTime":"2025-11-21T14:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:03 crc kubenswrapper[4897]: I1121 14:10:03.780354 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:03 crc kubenswrapper[4897]: I1121 14:10:03.780401 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:03 crc kubenswrapper[4897]: I1121 14:10:03.780412 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:03 crc kubenswrapper[4897]: I1121 14:10:03.780433 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:03 crc kubenswrapper[4897]: I1121 14:10:03.780444 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:03Z","lastTransitionTime":"2025-11-21T14:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:03 crc kubenswrapper[4897]: I1121 14:10:03.883326 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:03 crc kubenswrapper[4897]: I1121 14:10:03.883603 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:03 crc kubenswrapper[4897]: I1121 14:10:03.883744 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:03 crc kubenswrapper[4897]: I1121 14:10:03.883827 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:03 crc kubenswrapper[4897]: I1121 14:10:03.883891 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:03Z","lastTransitionTime":"2025-11-21T14:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:03 crc kubenswrapper[4897]: I1121 14:10:03.986682 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:03 crc kubenswrapper[4897]: I1121 14:10:03.986726 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:03 crc kubenswrapper[4897]: I1121 14:10:03.986735 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:03 crc kubenswrapper[4897]: I1121 14:10:03.986752 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:03 crc kubenswrapper[4897]: I1121 14:10:03.986762 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:03Z","lastTransitionTime":"2025-11-21T14:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:04 crc kubenswrapper[4897]: I1121 14:10:04.089341 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:04 crc kubenswrapper[4897]: I1121 14:10:04.089857 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:04 crc kubenswrapper[4897]: I1121 14:10:04.090012 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:04 crc kubenswrapper[4897]: I1121 14:10:04.090135 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:04 crc kubenswrapper[4897]: I1121 14:10:04.090244 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:04Z","lastTransitionTime":"2025-11-21T14:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:04 crc kubenswrapper[4897]: I1121 14:10:04.100123 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c976d804a7ebeba2c879e920cbc389f96f7f5def4b4439e8c8980d5460ecdb65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:04Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:04 crc kubenswrapper[4897]: I1121 14:10:04.112547 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nl76g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a7d6101-d0d2-4828-b3ce-d01040c04640\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac58b6ef1bd08e5f950260c18120c72e50b90a0de182c6ab0a5dbbbe45cf7328\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e56402f62821baf335f378084f4ed7c21b60ccbafe148419a0a9a403f106ad19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e56402f62821baf335f378084f4ed7c21b60ccbafe148419a0a9a403f106ad19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6022ff28e83dafcd846a7f03b090262aa3229a0da89f1984e4b6a054a04ac24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6022ff28e83dafcd846a7f03b090262aa3229a0da89f1984e4b6a054a04ac24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89de56e32b796da11bf89dd598d04fbf2884745ddf0f5d87cccfcaf1c001d8d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89de56e32b796da11bf89dd598d04fbf2884745ddf0f5d87cccfcaf1c001d8d6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2dff3cacaba7f9238f482e67b670ea775e0f96621434b99f4e76c2dfdde0321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2dff3cacaba7f9238f482e67b670ea775e0f96621434b99f4e76c2dfdde0321\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a153389a4137ee580e5118f3248400460f9d014c992a956f508815755f583afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a153389a4137ee580e5118f3248400460f9d014c992a956f508815755f583afb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a8c051ad1c1f2373f79eb0781fa6db489aeea3d52ed6c0bb43eec11520ec2b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a8c051ad1c1f2373f79eb0781fa6db489aeea3d52ed6c0bb43eec11520ec2b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nl76g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:04Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:04 crc kubenswrapper[4897]: I1121 14:10:04.122885 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:04Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:04 crc kubenswrapper[4897]: I1121 14:10:04.132932 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48c6b1ea7497b33c00b956e400426c1d4f3c6c66e0225133434460800eef9314\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ea6068a6b6f8123b402de354e5374a0f02a7191a922d1718ab38f8898f47c52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:04Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:04 crc kubenswrapper[4897]: I1121 14:10:04.140603 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4qwqj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f46e2ba8-9ee0-406d-a668-2ba424618b69\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f43aa89d4d9defcb50d0dc6f90293b5f426b9cfc398b780c2e10f6a5c9f1e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbsgb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4qwqj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:04Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:04 crc kubenswrapper[4897]: I1121 14:10:04.149102 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-rs4c8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7xkdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7xkdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:23Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-rs4c8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:04Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:04 crc kubenswrapper[4897]: I1121 14:10:04.159539 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:04Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:04 crc kubenswrapper[4897]: I1121 14:10:04.169908 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7670227-d280-4847-b882-754429f56b0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://961961696b8750c248598f4cda032b18395e5c802275dcfe6d3e9f7ef3510f1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5854028821e6a360a259427b68e7bb4c2a7220e0c68f4cc91639b2ece0540a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-krv5b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:04Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:04 crc kubenswrapper[4897]: I1121 14:10:04.185634 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"222f65f2-0b82-4760-bcf7-779244720f01\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92f1d932a83d65ba6a9a906cc489ff51e76547d07d42656630c59421f833c4db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccd51aa0faf3b1327f207b355cddf86a7353c0051f7e5ce0396f3c4bc67e3b16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8971c38daad45bda3d61545f52120dad40992dbf2e1d6e1e776500ed58d73f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfee19f82e35869a641318a0c45bc7a2ae71d80039f5790e60e277be15c9d3ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec279f2f558a8ac653006321b18cbaf13185876fe303c5cc1c65988aef09edbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de095f3c760b13d5197f624eac3806a1f4186577f2b95eb522532d453eb266e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e285a47c1193b01d8cd3cb16ca750bea2f24d3ad941dd438f96d5203f9203de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e285a47c1193b01d8cd3cb16ca750bea2f24d3ad941dd438f96d5203f9203de\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-21T14:09:38Z\\\",\\\"message\\\":\\\":776] Recording success event on pod openshift-etcd/etcd-crc\\\\nI1121 14:09:38.005810 6569 default_network_controller.go:776] Recording success event on pod openshift-image-registry/node-ca-7qfmq\\\\nI1121 14:09:38.005827 6569 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-additional-cni-plugins-nl76g\\\\nI1121 14:09:38.005832 6569 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-additional-cni-plugins-nl76g\\\\nI1121 14:09:38.005837 6569 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-additional-cni-plugins-nl76g in node crc\\\\nI1121 14:09:38.005842 6569 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-additional-cni-plugins-nl76g after 0 failed attempt(s)\\\\nI1121 14:09:38.005846 6569 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-additional-cni-plugins-nl76g\\\\nI1121 14:09:38.005834 6569 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-operator-lifecycle-manager/package-server-manager-metrics]} name:Service_openshift-operator-lifecycle-manager/package-server-manager-metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:37Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-tpfb6_openshift-ovn-kubernetes(222f65f2-0b82-4760-bcf7-779244720f01)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b486a86e0e97ab409e590f30a1bf5cb9f584a7c67f96ff1236e42ca4f8675b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7877db6f5c2bf0eb31218c64d67bca48f921ff7ac62b4e495f29024c086bf738\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7877db6f5c2bf0eb31218c64d67bca48f921ff7ac62b4e495f29024c086bf738\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tpfb6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:04Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:04 crc kubenswrapper[4897]: I1121 14:10:04.192976 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:04 crc kubenswrapper[4897]: I1121 14:10:04.193084 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:04 crc kubenswrapper[4897]: I1121 14:10:04.193267 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:04 crc kubenswrapper[4897]: I1121 14:10:04.193348 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:04 crc kubenswrapper[4897]: I1121 14:10:04.193413 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:04Z","lastTransitionTime":"2025-11-21T14:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:04 crc kubenswrapper[4897]: I1121 14:10:04.195885 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7qfmq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"223b7002-3e67-4848-b1fb-db2921deb64e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381a46c6290da2b36e572f46c16f1a50aa3e25359bcfbc02c7a5a18e1484860e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2m27\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7qfmq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:04Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:04 crc kubenswrapper[4897]: I1121 14:10:04.208141 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"402271c9-35e5-4bed-8410-c84c2e132878\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08651f8381818947818e0bb640c089be4a07022efc650befbb65ad58466bf8c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://419ab9403696491ef53a3d812d4c19d676f8fb1a1a4dfcdfcc0c6255377d2918\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://acc9629bb6b0dfbe5b4efe2f9f980e4e3865fdc6742782ebc49a5a7e0f775b9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0fc71bebc463919826d01260f38b63bfdb01199fd8dd6ae90d9ac9cd75db703\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0fc71bebc463919826d01260f38b63bfdb01199fd8dd6ae90d9ac9cd75db703\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:04Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:04 crc kubenswrapper[4897]: I1121 14:10:04.222130 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a6c0a99-565b-41f9-a8b5-c8990a0c8eb3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://737553dd1d16e2019ea89a376e3a868f9792ed5239aa7d6cc40a05d3d4f286ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6682a8886d2b9bce069cf9261d4701ef6907f19574fd035bd89bd05626023e5c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de079956e3d8de0101f3f87dcc120ef23876c4972382639e419b3f34a38cdfca\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://469b5be7467d4a58d3776ec0a967ed911be31b770b49965da4f5d2fb264d8a3d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1907bfbad7eb34da642e30d6aa81820aab457f92347333a364ac12396b0e9dae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"le observer\\\\nW1121 14:09:07.994401 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1121 14:09:07.994601 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1121 14:09:07.995290 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1461162674/tls.crt::/tmp/serving-cert-1461162674/tls.key\\\\\\\"\\\\nI1121 14:09:08.162382 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1121 14:09:08.165301 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1121 14:09:08.165318 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1121 14:09:08.165343 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1121 14:09:08.165348 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1121 14:09:08.172015 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1121 14:09:08.172083 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1121 14:09:08.172089 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1121 14:09:08.172094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1121 14:09:08.172098 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1121 14:09:08.172102 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1121 14:09:08.172106 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1121 14:09:08.172023 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1121 14:09:08.174214 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d78c87c440c7d8e30075f6143111b3f504fb99686484b8f19a9c8d61a55eccc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://101fb592fef069af009c613a1e9c9825024ed205498ab72334211ba90d6f7397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://101fb592fef069af009c613a1e9c9825024ed205498ab72334211ba90d6f7397\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:04Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:04 crc kubenswrapper[4897]: I1121 14:10:04.234889 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b78c993a-9966-4e84-85a1-b4f73aaf17f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce3a2602338a1c1e9380ee55f78360f3d9cf6ccc62678886dc5ae574a72e3d9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c256823f2fb49ecc121f6ef564cd86196f9b4a473e9bdea670173596ef14a823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c65f71f8746044f5d41687c670a0308cd09dc41a6531f9b5720e852b3d240756\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1dbe88e7e6ad63899e107ebb884a16edb9a42b21ae7c86d7fbfaf89785b0d36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:04Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:04 crc kubenswrapper[4897]: I1121 14:10:04.250649 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09877d9780caddd5f7f73e95fdf818a427ed645ed5997984113151f4b4759c13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:04Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:04 crc kubenswrapper[4897]: I1121 14:10:04.263803 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:04Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:04 crc kubenswrapper[4897]: I1121 14:10:04.278121 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6ljgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f3b70b5-9294-4f69-9528-500d28f34c89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:10:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:10:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf92ab70c5a3196bf83f427ccf03acf188d1995766a9154bcd6267e3926c3ac8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://527cf28abdeef1b42d9c19fb88e3171dc05c4516c3a542a461b5b1daa85c6d25\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-21T14:09:58Z\\\",\\\"message\\\":\\\"2025-11-21T14:09:13+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_a63cd511-2e05-4faf-b47b-8037110976d1\\\\n2025-11-21T14:09:13+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_a63cd511-2e05-4faf-b47b-8037110976d1 to /host/opt/cni/bin/\\\\n2025-11-21T14:09:13Z [verbose] multus-daemon started\\\\n2025-11-21T14:09:13Z [verbose] Readiness Indicator file check\\\\n2025-11-21T14:09:58Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78jhb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6ljgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:04Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:04 crc kubenswrapper[4897]: I1121 14:10:04.288711 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x5pws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc9e717e-af03-48bf-b005-62b3d084c6fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f968eee1d8a1888229e23abda59fd58fa5e86f02ca96ff80fbe093904b27e8df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9c4f5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0f99b9213902c003e6ca754134d2482404ace39c184fb3b431d2b600481d5b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9c4f5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-x5pws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:04Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:04 crc kubenswrapper[4897]: I1121 14:10:04.295860 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:04 crc kubenswrapper[4897]: I1121 14:10:04.295921 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:04 crc kubenswrapper[4897]: I1121 14:10:04.295930 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:04 crc kubenswrapper[4897]: I1121 14:10:04.295947 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:04 crc kubenswrapper[4897]: I1121 14:10:04.295961 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:04Z","lastTransitionTime":"2025-11-21T14:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:04 crc kubenswrapper[4897]: I1121 14:10:04.307764 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6cd330fd-e8a6-49d5-bfc9-5f6122ccd49e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1bd64b9b938bc80284bcee54141e90cad613d339c4f27ef2934b29bc2a318916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://936f59996d26e42099dcd055f87fff9e3ac4283b57e1956fe97463f2334152d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca74b28a069dc2e609b0779164c3aa470a15ed9848af0dd5fc3479aaf750157b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://270c53b9a342f26be077a01cd4e87e9ccd0eb5d2c84208ae38cbf720463245e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66c521c26d541bbddb064dd9710f1a7b41090d2f47343977922272dfc6dab7af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6eb2ac410b49098e5b3603de5e6bf5ed4484fab0b300021d0bed2e5ed8a42527\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6eb2ac410b49098e5b3603de5e6bf5ed4484fab0b300021d0bed2e5ed8a42527\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03385e0c78e4806f60c160227ef4905a11003ed3d70f0e5c0dd0a477979d1d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03385e0c78e4806f60c160227ef4905a11003ed3d70f0e5c0dd0a477979d1d4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://50aea2fa8b3b141b4330ddce2c8e0715230c6f06dab64b3abfd98a3067224bec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50aea2fa8b3b141b4330ddce2c8e0715230c6f06dab64b3abfd98a3067224bec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:04Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:04 crc kubenswrapper[4897]: I1121 14:10:04.399592 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:04 crc kubenswrapper[4897]: I1121 14:10:04.399673 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:04 crc kubenswrapper[4897]: I1121 14:10:04.399699 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:04 crc kubenswrapper[4897]: I1121 14:10:04.399735 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:04 crc kubenswrapper[4897]: I1121 14:10:04.399759 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:04Z","lastTransitionTime":"2025-11-21T14:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:04 crc kubenswrapper[4897]: I1121 14:10:04.501940 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:04 crc kubenswrapper[4897]: I1121 14:10:04.502308 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:04 crc kubenswrapper[4897]: I1121 14:10:04.502319 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:04 crc kubenswrapper[4897]: I1121 14:10:04.502336 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:04 crc kubenswrapper[4897]: I1121 14:10:04.502349 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:04Z","lastTransitionTime":"2025-11-21T14:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:04 crc kubenswrapper[4897]: I1121 14:10:04.604307 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:04 crc kubenswrapper[4897]: I1121 14:10:04.604348 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:04 crc kubenswrapper[4897]: I1121 14:10:04.604360 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:04 crc kubenswrapper[4897]: I1121 14:10:04.604394 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:04 crc kubenswrapper[4897]: I1121 14:10:04.604423 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:04Z","lastTransitionTime":"2025-11-21T14:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:04 crc kubenswrapper[4897]: I1121 14:10:04.706676 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:04 crc kubenswrapper[4897]: I1121 14:10:04.706711 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:04 crc kubenswrapper[4897]: I1121 14:10:04.706719 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:04 crc kubenswrapper[4897]: I1121 14:10:04.706733 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:04 crc kubenswrapper[4897]: I1121 14:10:04.706742 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:04Z","lastTransitionTime":"2025-11-21T14:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:04 crc kubenswrapper[4897]: I1121 14:10:04.809055 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:04 crc kubenswrapper[4897]: I1121 14:10:04.809102 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:04 crc kubenswrapper[4897]: I1121 14:10:04.809111 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:04 crc kubenswrapper[4897]: I1121 14:10:04.809129 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:04 crc kubenswrapper[4897]: I1121 14:10:04.809140 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:04Z","lastTransitionTime":"2025-11-21T14:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:04 crc kubenswrapper[4897]: I1121 14:10:04.912002 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:04 crc kubenswrapper[4897]: I1121 14:10:04.912052 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:04 crc kubenswrapper[4897]: I1121 14:10:04.912065 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:04 crc kubenswrapper[4897]: I1121 14:10:04.912082 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:04 crc kubenswrapper[4897]: I1121 14:10:04.912095 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:04Z","lastTransitionTime":"2025-11-21T14:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:05 crc kubenswrapper[4897]: I1121 14:10:05.021908 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:05 crc kubenswrapper[4897]: I1121 14:10:05.021948 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:05 crc kubenswrapper[4897]: I1121 14:10:05.021958 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:05 crc kubenswrapper[4897]: I1121 14:10:05.021976 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:05 crc kubenswrapper[4897]: I1121 14:10:05.021988 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:05Z","lastTransitionTime":"2025-11-21T14:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:05 crc kubenswrapper[4897]: I1121 14:10:05.088489 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs4c8" Nov 21 14:10:05 crc kubenswrapper[4897]: I1121 14:10:05.088580 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 21 14:10:05 crc kubenswrapper[4897]: E1121 14:10:05.088686 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs4c8" podUID="01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd" Nov 21 14:10:05 crc kubenswrapper[4897]: E1121 14:10:05.088764 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 21 14:10:05 crc kubenswrapper[4897]: I1121 14:10:05.088852 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 21 14:10:05 crc kubenswrapper[4897]: E1121 14:10:05.088901 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 21 14:10:05 crc kubenswrapper[4897]: I1121 14:10:05.088946 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 21 14:10:05 crc kubenswrapper[4897]: E1121 14:10:05.088996 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 21 14:10:05 crc kubenswrapper[4897]: I1121 14:10:05.124254 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:05 crc kubenswrapper[4897]: I1121 14:10:05.124303 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:05 crc kubenswrapper[4897]: I1121 14:10:05.124317 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:05 crc kubenswrapper[4897]: I1121 14:10:05.124336 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:05 crc kubenswrapper[4897]: I1121 14:10:05.124353 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:05Z","lastTransitionTime":"2025-11-21T14:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:05 crc kubenswrapper[4897]: I1121 14:10:05.226764 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:05 crc kubenswrapper[4897]: I1121 14:10:05.226795 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:05 crc kubenswrapper[4897]: I1121 14:10:05.226802 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:05 crc kubenswrapper[4897]: I1121 14:10:05.226818 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:05 crc kubenswrapper[4897]: I1121 14:10:05.226833 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:05Z","lastTransitionTime":"2025-11-21T14:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:05 crc kubenswrapper[4897]: I1121 14:10:05.329711 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:05 crc kubenswrapper[4897]: I1121 14:10:05.329749 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:05 crc kubenswrapper[4897]: I1121 14:10:05.329759 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:05 crc kubenswrapper[4897]: I1121 14:10:05.329775 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:05 crc kubenswrapper[4897]: I1121 14:10:05.329787 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:05Z","lastTransitionTime":"2025-11-21T14:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:05 crc kubenswrapper[4897]: I1121 14:10:05.432027 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:05 crc kubenswrapper[4897]: I1121 14:10:05.432076 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:05 crc kubenswrapper[4897]: I1121 14:10:05.432087 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:05 crc kubenswrapper[4897]: I1121 14:10:05.432104 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:05 crc kubenswrapper[4897]: I1121 14:10:05.432117 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:05Z","lastTransitionTime":"2025-11-21T14:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:05 crc kubenswrapper[4897]: I1121 14:10:05.535115 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:05 crc kubenswrapper[4897]: I1121 14:10:05.535169 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:05 crc kubenswrapper[4897]: I1121 14:10:05.535185 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:05 crc kubenswrapper[4897]: I1121 14:10:05.535203 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:05 crc kubenswrapper[4897]: I1121 14:10:05.535216 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:05Z","lastTransitionTime":"2025-11-21T14:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:05 crc kubenswrapper[4897]: I1121 14:10:05.638572 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:05 crc kubenswrapper[4897]: I1121 14:10:05.638638 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:05 crc kubenswrapper[4897]: I1121 14:10:05.638664 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:05 crc kubenswrapper[4897]: I1121 14:10:05.638693 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:05 crc kubenswrapper[4897]: I1121 14:10:05.638714 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:05Z","lastTransitionTime":"2025-11-21T14:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:05 crc kubenswrapper[4897]: I1121 14:10:05.741294 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:05 crc kubenswrapper[4897]: I1121 14:10:05.741336 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:05 crc kubenswrapper[4897]: I1121 14:10:05.741348 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:05 crc kubenswrapper[4897]: I1121 14:10:05.741367 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:05 crc kubenswrapper[4897]: I1121 14:10:05.741378 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:05Z","lastTransitionTime":"2025-11-21T14:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:05 crc kubenswrapper[4897]: I1121 14:10:05.844021 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:05 crc kubenswrapper[4897]: I1121 14:10:05.844093 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:05 crc kubenswrapper[4897]: I1121 14:10:05.844116 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:05 crc kubenswrapper[4897]: I1121 14:10:05.844147 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:05 crc kubenswrapper[4897]: I1121 14:10:05.844172 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:05Z","lastTransitionTime":"2025-11-21T14:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:05 crc kubenswrapper[4897]: I1121 14:10:05.946947 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:05 crc kubenswrapper[4897]: I1121 14:10:05.947104 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:05 crc kubenswrapper[4897]: I1121 14:10:05.947130 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:05 crc kubenswrapper[4897]: I1121 14:10:05.947154 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:05 crc kubenswrapper[4897]: I1121 14:10:05.947171 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:05Z","lastTransitionTime":"2025-11-21T14:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:06 crc kubenswrapper[4897]: I1121 14:10:06.050059 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:06 crc kubenswrapper[4897]: I1121 14:10:06.050120 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:06 crc kubenswrapper[4897]: I1121 14:10:06.050130 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:06 crc kubenswrapper[4897]: I1121 14:10:06.050158 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:06 crc kubenswrapper[4897]: I1121 14:10:06.050170 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:06Z","lastTransitionTime":"2025-11-21T14:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:06 crc kubenswrapper[4897]: I1121 14:10:06.088725 4897 scope.go:117] "RemoveContainer" containerID="9e285a47c1193b01d8cd3cb16ca750bea2f24d3ad941dd438f96d5203f9203de" Nov 21 14:10:06 crc kubenswrapper[4897]: I1121 14:10:06.152541 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:06 crc kubenswrapper[4897]: I1121 14:10:06.152607 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:06 crc kubenswrapper[4897]: I1121 14:10:06.152630 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:06 crc kubenswrapper[4897]: I1121 14:10:06.152660 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:06 crc kubenswrapper[4897]: I1121 14:10:06.152681 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:06Z","lastTransitionTime":"2025-11-21T14:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:06 crc kubenswrapper[4897]: I1121 14:10:06.256133 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:06 crc kubenswrapper[4897]: I1121 14:10:06.256187 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:06 crc kubenswrapper[4897]: I1121 14:10:06.256201 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:06 crc kubenswrapper[4897]: I1121 14:10:06.256224 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:06 crc kubenswrapper[4897]: I1121 14:10:06.256242 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:06Z","lastTransitionTime":"2025-11-21T14:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:06 crc kubenswrapper[4897]: I1121 14:10:06.359122 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:06 crc kubenswrapper[4897]: I1121 14:10:06.359166 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:06 crc kubenswrapper[4897]: I1121 14:10:06.359175 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:06 crc kubenswrapper[4897]: I1121 14:10:06.359190 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:06 crc kubenswrapper[4897]: I1121 14:10:06.359202 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:06Z","lastTransitionTime":"2025-11-21T14:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:06 crc kubenswrapper[4897]: I1121 14:10:06.461388 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:06 crc kubenswrapper[4897]: I1121 14:10:06.461426 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:06 crc kubenswrapper[4897]: I1121 14:10:06.461436 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:06 crc kubenswrapper[4897]: I1121 14:10:06.461450 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:06 crc kubenswrapper[4897]: I1121 14:10:06.461459 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:06Z","lastTransitionTime":"2025-11-21T14:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:06 crc kubenswrapper[4897]: I1121 14:10:06.564063 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:06 crc kubenswrapper[4897]: I1121 14:10:06.564094 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:06 crc kubenswrapper[4897]: I1121 14:10:06.564103 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:06 crc kubenswrapper[4897]: I1121 14:10:06.564118 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:06 crc kubenswrapper[4897]: I1121 14:10:06.564127 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:06Z","lastTransitionTime":"2025-11-21T14:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:06 crc kubenswrapper[4897]: I1121 14:10:06.587324 4897 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tpfb6_222f65f2-0b82-4760-bcf7-779244720f01/ovnkube-controller/2.log" Nov 21 14:10:06 crc kubenswrapper[4897]: I1121 14:10:06.589402 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" event={"ID":"222f65f2-0b82-4760-bcf7-779244720f01","Type":"ContainerStarted","Data":"40b0bddce48b3626ad889ae2f43deb866acd521a8aab71253506364bfb47c723"} Nov 21 14:10:06 crc kubenswrapper[4897]: I1121 14:10:06.589993 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" Nov 21 14:10:06 crc kubenswrapper[4897]: I1121 14:10:06.610203 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6cd330fd-e8a6-49d5-bfc9-5f6122ccd49e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1bd64b9b938bc80284bcee54141e90cad613d339c4f27ef2934b29bc2a318916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://936f59996d26e42099dcd055f87fff9e3ac4283b57e1956fe97463f2334152d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca74b28a069dc2e609b0779164c3aa470a15ed9848af0dd5fc3479aaf750157b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://270c53b9a342f26be077a01cd4e87e9ccd0eb5d2c84208ae38cbf720463245e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66c521c26d541bbddb064dd9710f1a7b41090d2f47343977922272dfc6dab7af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6eb2ac410b49098e5b3603de5e6bf5ed4484fab0b300021d0bed2e5ed8a42527\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6eb2ac410b49098e5b3603de5e6bf5ed4484fab0b300021d0bed2e5ed8a42527\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03385e0c78e4806f60c160227ef4905a11003ed3d70f0e5c0dd0a477979d1d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03385e0c78e4806f60c160227ef4905a11003ed3d70f0e5c0dd0a477979d1d4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://50aea2fa8b3b141b4330ddce2c8e0715230c6f06dab64b3abfd98a3067224bec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50aea2fa8b3b141b4330ddce2c8e0715230c6f06dab64b3abfd98a3067224bec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:06Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:06 crc kubenswrapper[4897]: I1121 14:10:06.622853 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a6c0a99-565b-41f9-a8b5-c8990a0c8eb3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://737553dd1d16e2019ea89a376e3a868f9792ed5239aa7d6cc40a05d3d4f286ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6682a8886d2b9bce069cf9261d4701ef6907f19574fd035bd89bd05626023e5c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de079956e3d8de0101f3f87dcc120ef23876c4972382639e419b3f34a38cdfca\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://469b5be7467d4a58d3776ec0a967ed911be31b770b49965da4f5d2fb264d8a3d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1907bfbad7eb34da642e30d6aa81820aab457f92347333a364ac12396b0e9dae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"le observer\\\\nW1121 14:09:07.994401 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1121 14:09:07.994601 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1121 14:09:07.995290 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1461162674/tls.crt::/tmp/serving-cert-1461162674/tls.key\\\\\\\"\\\\nI1121 14:09:08.162382 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1121 14:09:08.165301 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1121 14:09:08.165318 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1121 14:09:08.165343 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1121 14:09:08.165348 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1121 14:09:08.172015 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1121 14:09:08.172083 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1121 14:09:08.172089 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1121 14:09:08.172094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1121 14:09:08.172098 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1121 14:09:08.172102 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1121 14:09:08.172106 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1121 14:09:08.172023 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1121 14:09:08.174214 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d78c87c440c7d8e30075f6143111b3f504fb99686484b8f19a9c8d61a55eccc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://101fb592fef069af009c613a1e9c9825024ed205498ab72334211ba90d6f7397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://101fb592fef069af009c613a1e9c9825024ed205498ab72334211ba90d6f7397\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:06Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:06 crc kubenswrapper[4897]: I1121 14:10:06.632839 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b78c993a-9966-4e84-85a1-b4f73aaf17f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce3a2602338a1c1e9380ee55f78360f3d9cf6ccc62678886dc5ae574a72e3d9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c256823f2fb49ecc121f6ef564cd86196f9b4a473e9bdea670173596ef14a823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c65f71f8746044f5d41687c670a0308cd09dc41a6531f9b5720e852b3d240756\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1dbe88e7e6ad63899e107ebb884a16edb9a42b21ae7c86d7fbfaf89785b0d36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:06Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:06 crc kubenswrapper[4897]: I1121 14:10:06.642906 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09877d9780caddd5f7f73e95fdf818a427ed645ed5997984113151f4b4759c13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:06Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:06 crc kubenswrapper[4897]: I1121 14:10:06.652765 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:06Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:06 crc kubenswrapper[4897]: I1121 14:10:06.663303 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6ljgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f3b70b5-9294-4f69-9528-500d28f34c89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:10:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:10:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf92ab70c5a3196bf83f427ccf03acf188d1995766a9154bcd6267e3926c3ac8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://527cf28abdeef1b42d9c19fb88e3171dc05c4516c3a542a461b5b1daa85c6d25\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-21T14:09:58Z\\\",\\\"message\\\":\\\"2025-11-21T14:09:13+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_a63cd511-2e05-4faf-b47b-8037110976d1\\\\n2025-11-21T14:09:13+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_a63cd511-2e05-4faf-b47b-8037110976d1 to /host/opt/cni/bin/\\\\n2025-11-21T14:09:13Z [verbose] multus-daemon started\\\\n2025-11-21T14:09:13Z [verbose] Readiness Indicator file check\\\\n2025-11-21T14:09:58Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78jhb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6ljgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:06Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:06 crc kubenswrapper[4897]: I1121 14:10:06.665788 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:06 crc kubenswrapper[4897]: I1121 14:10:06.665844 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:06 crc kubenswrapper[4897]: I1121 14:10:06.665858 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:06 crc kubenswrapper[4897]: I1121 14:10:06.665888 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:06 crc kubenswrapper[4897]: I1121 14:10:06.665902 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:06Z","lastTransitionTime":"2025-11-21T14:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:06 crc kubenswrapper[4897]: I1121 14:10:06.675644 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x5pws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc9e717e-af03-48bf-b005-62b3d084c6fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f968eee1d8a1888229e23abda59fd58fa5e86f02ca96ff80fbe093904b27e8df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9c4f5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0f99b9213902c003e6ca754134d2482404ace39c184fb3b431d2b600481d5b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9c4f5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-x5pws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:06Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:06 crc kubenswrapper[4897]: I1121 14:10:06.690685 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:06Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:06 crc kubenswrapper[4897]: I1121 14:10:06.706115 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c976d804a7ebeba2c879e920cbc389f96f7f5def4b4439e8c8980d5460ecdb65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:06Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:06 crc kubenswrapper[4897]: I1121 14:10:06.723580 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nl76g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a7d6101-d0d2-4828-b3ce-d01040c04640\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac58b6ef1bd08e5f950260c18120c72e50b90a0de182c6ab0a5dbbbe45cf7328\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e56402f62821baf335f378084f4ed7c21b60ccbafe148419a0a9a403f106ad19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e56402f62821baf335f378084f4ed7c21b60ccbafe148419a0a9a403f106ad19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6022ff28e83dafcd846a7f03b090262aa3229a0da89f1984e4b6a054a04ac24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6022ff28e83dafcd846a7f03b090262aa3229a0da89f1984e4b6a054a04ac24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89de56e32b796da11bf89dd598d04fbf2884745ddf0f5d87cccfcaf1c001d8d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89de56e32b796da11bf89dd598d04fbf2884745ddf0f5d87cccfcaf1c001d8d6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2dff3cacaba7f9238f482e67b670ea775e0f96621434b99f4e76c2dfdde0321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2dff3cacaba7f9238f482e67b670ea775e0f96621434b99f4e76c2dfdde0321\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a153389a4137ee580e5118f3248400460f9d014c992a956f508815755f583afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a153389a4137ee580e5118f3248400460f9d014c992a956f508815755f583afb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a8c051ad1c1f2373f79eb0781fa6db489aeea3d52ed6c0bb43eec11520ec2b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a8c051ad1c1f2373f79eb0781fa6db489aeea3d52ed6c0bb43eec11520ec2b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nl76g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:06Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:06 crc kubenswrapper[4897]: I1121 14:10:06.737703 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48c6b1ea7497b33c00b956e400426c1d4f3c6c66e0225133434460800eef9314\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ea6068a6b6f8123b402de354e5374a0f02a7191a922d1718ab38f8898f47c52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:06Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:06 crc kubenswrapper[4897]: I1121 14:10:06.752344 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4qwqj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f46e2ba8-9ee0-406d-a668-2ba424618b69\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f43aa89d4d9defcb50d0dc6f90293b5f426b9cfc398b780c2e10f6a5c9f1e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbsgb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4qwqj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:06Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:06 crc kubenswrapper[4897]: I1121 14:10:06.763962 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-rs4c8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7xkdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7xkdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:23Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-rs4c8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:06Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:06 crc kubenswrapper[4897]: I1121 14:10:06.768242 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:06 crc kubenswrapper[4897]: I1121 14:10:06.768268 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:06 crc kubenswrapper[4897]: I1121 14:10:06.768277 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:06 crc kubenswrapper[4897]: I1121 14:10:06.768291 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:06 crc kubenswrapper[4897]: I1121 14:10:06.768301 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:06Z","lastTransitionTime":"2025-11-21T14:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:06 crc kubenswrapper[4897]: I1121 14:10:06.778260 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"402271c9-35e5-4bed-8410-c84c2e132878\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08651f8381818947818e0bb640c089be4a07022efc650befbb65ad58466bf8c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://419ab9403696491ef53a3d812d4c19d676f8fb1a1a4dfcdfcc0c6255377d2918\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://acc9629bb6b0dfbe5b4efe2f9f980e4e3865fdc6742782ebc49a5a7e0f775b9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0fc71bebc463919826d01260f38b63bfdb01199fd8dd6ae90d9ac9cd75db703\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0fc71bebc463919826d01260f38b63bfdb01199fd8dd6ae90d9ac9cd75db703\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:06Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:06 crc kubenswrapper[4897]: I1121 14:10:06.792326 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:06Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:06 crc kubenswrapper[4897]: I1121 14:10:06.804256 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7670227-d280-4847-b882-754429f56b0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://961961696b8750c248598f4cda032b18395e5c802275dcfe6d3e9f7ef3510f1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5854028821e6a360a259427b68e7bb4c2a7220e0c68f4cc91639b2ece0540a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-krv5b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:06Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:06 crc kubenswrapper[4897]: I1121 14:10:06.823407 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"222f65f2-0b82-4760-bcf7-779244720f01\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92f1d932a83d65ba6a9a906cc489ff51e76547d07d42656630c59421f833c4db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccd51aa0faf3b1327f207b355cddf86a7353c0051f7e5ce0396f3c4bc67e3b16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8971c38daad45bda3d61545f52120dad40992dbf2e1d6e1e776500ed58d73f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfee19f82e35869a641318a0c45bc7a2ae71d80039f5790e60e277be15c9d3ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec279f2f558a8ac653006321b18cbaf13185876fe303c5cc1c65988aef09edbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de095f3c760b13d5197f624eac3806a1f4186577f2b95eb522532d453eb266e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40b0bddce48b3626ad889ae2f43deb866acd521a8aab71253506364bfb47c723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e285a47c1193b01d8cd3cb16ca750bea2f24d3ad941dd438f96d5203f9203de\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-21T14:09:38Z\\\",\\\"message\\\":\\\":776] Recording success event on pod openshift-etcd/etcd-crc\\\\nI1121 14:09:38.005810 6569 default_network_controller.go:776] Recording success event on pod openshift-image-registry/node-ca-7qfmq\\\\nI1121 14:09:38.005827 6569 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-additional-cni-plugins-nl76g\\\\nI1121 14:09:38.005832 6569 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-additional-cni-plugins-nl76g\\\\nI1121 14:09:38.005837 6569 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-additional-cni-plugins-nl76g in node crc\\\\nI1121 14:09:38.005842 6569 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-additional-cni-plugins-nl76g after 0 failed attempt(s)\\\\nI1121 14:09:38.005846 6569 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-additional-cni-plugins-nl76g\\\\nI1121 14:09:38.005834 6569 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-operator-lifecycle-manager/package-server-manager-metrics]} name:Service_openshift-operator-lifecycle-manager/package-server-manager-metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:37Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b486a86e0e97ab409e590f30a1bf5cb9f584a7c67f96ff1236e42ca4f8675b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7877db6f5c2bf0eb31218c64d67bca48f921ff7ac62b4e495f29024c086bf738\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7877db6f5c2bf0eb31218c64d67bca48f921ff7ac62b4e495f29024c086bf738\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tpfb6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:06Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:06 crc kubenswrapper[4897]: I1121 14:10:06.836845 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7qfmq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"223b7002-3e67-4848-b1fb-db2921deb64e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381a46c6290da2b36e572f46c16f1a50aa3e25359bcfbc02c7a5a18e1484860e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2m27\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7qfmq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:06Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:06 crc kubenswrapper[4897]: I1121 14:10:06.870774 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:06 crc kubenswrapper[4897]: I1121 14:10:06.870819 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:06 crc kubenswrapper[4897]: I1121 14:10:06.870828 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:06 crc kubenswrapper[4897]: I1121 14:10:06.870845 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:06 crc kubenswrapper[4897]: I1121 14:10:06.870856 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:06Z","lastTransitionTime":"2025-11-21T14:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:06 crc kubenswrapper[4897]: I1121 14:10:06.914813 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:06 crc kubenswrapper[4897]: I1121 14:10:06.914848 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:06 crc kubenswrapper[4897]: I1121 14:10:06.914856 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:06 crc kubenswrapper[4897]: I1121 14:10:06.914872 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:06 crc kubenswrapper[4897]: I1121 14:10:06.914881 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:06Z","lastTransitionTime":"2025-11-21T14:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:06 crc kubenswrapper[4897]: E1121 14:10:06.926879 4897 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:10:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:10:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:10:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:10:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:10:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:10:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:10:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:10:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8b8bdf9e-0fef-41ac-98ad-710dc2137183\\\",\\\"systemUUID\\\":\\\"8a210af6-a71f-4ea9-90a3-aac32fb868b7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:06Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:06 crc kubenswrapper[4897]: I1121 14:10:06.930909 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:06 crc kubenswrapper[4897]: I1121 14:10:06.930939 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:06 crc kubenswrapper[4897]: I1121 14:10:06.930948 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:06 crc kubenswrapper[4897]: I1121 14:10:06.930962 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:06 crc kubenswrapper[4897]: I1121 14:10:06.930972 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:06Z","lastTransitionTime":"2025-11-21T14:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:06 crc kubenswrapper[4897]: E1121 14:10:06.942092 4897 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:10:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:10:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:10:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:10:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:10:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:10:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:10:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:10:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8b8bdf9e-0fef-41ac-98ad-710dc2137183\\\",\\\"systemUUID\\\":\\\"8a210af6-a71f-4ea9-90a3-aac32fb868b7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:06Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:06 crc kubenswrapper[4897]: I1121 14:10:06.945727 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:06 crc kubenswrapper[4897]: I1121 14:10:06.945774 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:06 crc kubenswrapper[4897]: I1121 14:10:06.945787 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:06 crc kubenswrapper[4897]: I1121 14:10:06.945805 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:06 crc kubenswrapper[4897]: I1121 14:10:06.945819 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:06Z","lastTransitionTime":"2025-11-21T14:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:06 crc kubenswrapper[4897]: E1121 14:10:06.958631 4897 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:10:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:10:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:10:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:10:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:10:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:10:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:10:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:10:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8b8bdf9e-0fef-41ac-98ad-710dc2137183\\\",\\\"systemUUID\\\":\\\"8a210af6-a71f-4ea9-90a3-aac32fb868b7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:06Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:06 crc kubenswrapper[4897]: I1121 14:10:06.961897 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:06 crc kubenswrapper[4897]: I1121 14:10:06.961951 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:06 crc kubenswrapper[4897]: I1121 14:10:06.961963 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:06 crc kubenswrapper[4897]: I1121 14:10:06.961982 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:06 crc kubenswrapper[4897]: I1121 14:10:06.961999 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:06Z","lastTransitionTime":"2025-11-21T14:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:06 crc kubenswrapper[4897]: E1121 14:10:06.974163 4897 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:10:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:10:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:10:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:10:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:10:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:10:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:10:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:10:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8b8bdf9e-0fef-41ac-98ad-710dc2137183\\\",\\\"systemUUID\\\":\\\"8a210af6-a71f-4ea9-90a3-aac32fb868b7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:06Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:06 crc kubenswrapper[4897]: I1121 14:10:06.981221 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:06 crc kubenswrapper[4897]: I1121 14:10:06.981251 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:06 crc kubenswrapper[4897]: I1121 14:10:06.981262 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:06 crc kubenswrapper[4897]: I1121 14:10:06.981278 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:06 crc kubenswrapper[4897]: I1121 14:10:06.981290 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:06Z","lastTransitionTime":"2025-11-21T14:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:06 crc kubenswrapper[4897]: E1121 14:10:06.993023 4897 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:10:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:10:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:10:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:10:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:10:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:10:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:10:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:10:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8b8bdf9e-0fef-41ac-98ad-710dc2137183\\\",\\\"systemUUID\\\":\\\"8a210af6-a71f-4ea9-90a3-aac32fb868b7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:06Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:06 crc kubenswrapper[4897]: E1121 14:10:06.993152 4897 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 21 14:10:06 crc kubenswrapper[4897]: I1121 14:10:06.994833 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:06 crc kubenswrapper[4897]: I1121 14:10:06.994871 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:06 crc kubenswrapper[4897]: I1121 14:10:06.994881 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:06 crc kubenswrapper[4897]: I1121 14:10:06.994895 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:06 crc kubenswrapper[4897]: I1121 14:10:06.994907 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:06Z","lastTransitionTime":"2025-11-21T14:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:07 crc kubenswrapper[4897]: I1121 14:10:07.088809 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 21 14:10:07 crc kubenswrapper[4897]: I1121 14:10:07.088877 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 21 14:10:07 crc kubenswrapper[4897]: I1121 14:10:07.088877 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 21 14:10:07 crc kubenswrapper[4897]: I1121 14:10:07.088956 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs4c8" Nov 21 14:10:07 crc kubenswrapper[4897]: E1121 14:10:07.089085 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 21 14:10:07 crc kubenswrapper[4897]: E1121 14:10:07.089341 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 21 14:10:07 crc kubenswrapper[4897]: E1121 14:10:07.089426 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 21 14:10:07 crc kubenswrapper[4897]: E1121 14:10:07.089475 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs4c8" podUID="01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd" Nov 21 14:10:07 crc kubenswrapper[4897]: I1121 14:10:07.097348 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:07 crc kubenswrapper[4897]: I1121 14:10:07.097383 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:07 crc kubenswrapper[4897]: I1121 14:10:07.097395 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:07 crc kubenswrapper[4897]: I1121 14:10:07.097410 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:07 crc kubenswrapper[4897]: I1121 14:10:07.097421 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:07Z","lastTransitionTime":"2025-11-21T14:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:07 crc kubenswrapper[4897]: I1121 14:10:07.200729 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:07 crc kubenswrapper[4897]: I1121 14:10:07.201037 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:07 crc kubenswrapper[4897]: I1121 14:10:07.201119 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:07 crc kubenswrapper[4897]: I1121 14:10:07.201202 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:07 crc kubenswrapper[4897]: I1121 14:10:07.201269 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:07Z","lastTransitionTime":"2025-11-21T14:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:07 crc kubenswrapper[4897]: I1121 14:10:07.303807 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:07 crc kubenswrapper[4897]: I1121 14:10:07.303877 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:07 crc kubenswrapper[4897]: I1121 14:10:07.303900 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:07 crc kubenswrapper[4897]: I1121 14:10:07.303929 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:07 crc kubenswrapper[4897]: I1121 14:10:07.303950 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:07Z","lastTransitionTime":"2025-11-21T14:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:07 crc kubenswrapper[4897]: I1121 14:10:07.406622 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:07 crc kubenswrapper[4897]: I1121 14:10:07.406946 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:07 crc kubenswrapper[4897]: I1121 14:10:07.407011 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:07 crc kubenswrapper[4897]: I1121 14:10:07.407092 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:07 crc kubenswrapper[4897]: I1121 14:10:07.407166 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:07Z","lastTransitionTime":"2025-11-21T14:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:07 crc kubenswrapper[4897]: I1121 14:10:07.510440 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:07 crc kubenswrapper[4897]: I1121 14:10:07.510486 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:07 crc kubenswrapper[4897]: I1121 14:10:07.510496 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:07 crc kubenswrapper[4897]: I1121 14:10:07.510528 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:07 crc kubenswrapper[4897]: I1121 14:10:07.510538 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:07Z","lastTransitionTime":"2025-11-21T14:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:07 crc kubenswrapper[4897]: I1121 14:10:07.595709 4897 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tpfb6_222f65f2-0b82-4760-bcf7-779244720f01/ovnkube-controller/3.log" Nov 21 14:10:07 crc kubenswrapper[4897]: I1121 14:10:07.596580 4897 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tpfb6_222f65f2-0b82-4760-bcf7-779244720f01/ovnkube-controller/2.log" Nov 21 14:10:07 crc kubenswrapper[4897]: I1121 14:10:07.598898 4897 generic.go:334] "Generic (PLEG): container finished" podID="222f65f2-0b82-4760-bcf7-779244720f01" containerID="40b0bddce48b3626ad889ae2f43deb866acd521a8aab71253506364bfb47c723" exitCode=1 Nov 21 14:10:07 crc kubenswrapper[4897]: I1121 14:10:07.598951 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" event={"ID":"222f65f2-0b82-4760-bcf7-779244720f01","Type":"ContainerDied","Data":"40b0bddce48b3626ad889ae2f43deb866acd521a8aab71253506364bfb47c723"} Nov 21 14:10:07 crc kubenswrapper[4897]: I1121 14:10:07.598993 4897 scope.go:117] "RemoveContainer" containerID="9e285a47c1193b01d8cd3cb16ca750bea2f24d3ad941dd438f96d5203f9203de" Nov 21 14:10:07 crc kubenswrapper[4897]: I1121 14:10:07.599967 4897 scope.go:117] "RemoveContainer" containerID="40b0bddce48b3626ad889ae2f43deb866acd521a8aab71253506364bfb47c723" Nov 21 14:10:07 crc kubenswrapper[4897]: E1121 14:10:07.600122 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-tpfb6_openshift-ovn-kubernetes(222f65f2-0b82-4760-bcf7-779244720f01)\"" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" podUID="222f65f2-0b82-4760-bcf7-779244720f01" Nov 21 14:10:07 crc kubenswrapper[4897]: I1121 14:10:07.613249 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:07 crc kubenswrapper[4897]: I1121 14:10:07.613285 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:07 crc kubenswrapper[4897]: I1121 14:10:07.613296 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:07 crc kubenswrapper[4897]: I1121 14:10:07.613315 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:07 crc kubenswrapper[4897]: I1121 14:10:07.613329 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:07Z","lastTransitionTime":"2025-11-21T14:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:07 crc kubenswrapper[4897]: I1121 14:10:07.618341 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:07Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:07 crc kubenswrapper[4897]: I1121 14:10:07.632177 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6ljgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f3b70b5-9294-4f69-9528-500d28f34c89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:10:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:10:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf92ab70c5a3196bf83f427ccf03acf188d1995766a9154bcd6267e3926c3ac8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://527cf28abdeef1b42d9c19fb88e3171dc05c4516c3a542a461b5b1daa85c6d25\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-21T14:09:58Z\\\",\\\"message\\\":\\\"2025-11-21T14:09:13+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_a63cd511-2e05-4faf-b47b-8037110976d1\\\\n2025-11-21T14:09:13+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_a63cd511-2e05-4faf-b47b-8037110976d1 to /host/opt/cni/bin/\\\\n2025-11-21T14:09:13Z [verbose] multus-daemon started\\\\n2025-11-21T14:09:13Z [verbose] Readiness Indicator file check\\\\n2025-11-21T14:09:58Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78jhb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6ljgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:07Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:07 crc kubenswrapper[4897]: I1121 14:10:07.643974 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x5pws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc9e717e-af03-48bf-b005-62b3d084c6fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f968eee1d8a1888229e23abda59fd58fa5e86f02ca96ff80fbe093904b27e8df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9c4f5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0f99b9213902c003e6ca754134d2482404ace39c184fb3b431d2b600481d5b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9c4f5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-x5pws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:07Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:07 crc kubenswrapper[4897]: I1121 14:10:07.662886 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6cd330fd-e8a6-49d5-bfc9-5f6122ccd49e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1bd64b9b938bc80284bcee54141e90cad613d339c4f27ef2934b29bc2a318916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://936f59996d26e42099dcd055f87fff9e3ac4283b57e1956fe97463f2334152d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca74b28a069dc2e609b0779164c3aa470a15ed9848af0dd5fc3479aaf750157b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://270c53b9a342f26be077a01cd4e87e9ccd0eb5d2c84208ae38cbf720463245e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66c521c26d541bbddb064dd9710f1a7b41090d2f47343977922272dfc6dab7af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6eb2ac410b49098e5b3603de5e6bf5ed4484fab0b300021d0bed2e5ed8a42527\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6eb2ac410b49098e5b3603de5e6bf5ed4484fab0b300021d0bed2e5ed8a42527\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03385e0c78e4806f60c160227ef4905a11003ed3d70f0e5c0dd0a477979d1d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03385e0c78e4806f60c160227ef4905a11003ed3d70f0e5c0dd0a477979d1d4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://50aea2fa8b3b141b4330ddce2c8e0715230c6f06dab64b3abfd98a3067224bec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50aea2fa8b3b141b4330ddce2c8e0715230c6f06dab64b3abfd98a3067224bec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:07Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:07 crc kubenswrapper[4897]: I1121 14:10:07.675289 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a6c0a99-565b-41f9-a8b5-c8990a0c8eb3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://737553dd1d16e2019ea89a376e3a868f9792ed5239aa7d6cc40a05d3d4f286ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6682a8886d2b9bce069cf9261d4701ef6907f19574fd035bd89bd05626023e5c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de079956e3d8de0101f3f87dcc120ef23876c4972382639e419b3f34a38cdfca\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://469b5be7467d4a58d3776ec0a967ed911be31b770b49965da4f5d2fb264d8a3d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1907bfbad7eb34da642e30d6aa81820aab457f92347333a364ac12396b0e9dae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"le observer\\\\nW1121 14:09:07.994401 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1121 14:09:07.994601 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1121 14:09:07.995290 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1461162674/tls.crt::/tmp/serving-cert-1461162674/tls.key\\\\\\\"\\\\nI1121 14:09:08.162382 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1121 14:09:08.165301 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1121 14:09:08.165318 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1121 14:09:08.165343 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1121 14:09:08.165348 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1121 14:09:08.172015 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1121 14:09:08.172083 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1121 14:09:08.172089 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1121 14:09:08.172094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1121 14:09:08.172098 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1121 14:09:08.172102 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1121 14:09:08.172106 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1121 14:09:08.172023 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1121 14:09:08.174214 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d78c87c440c7d8e30075f6143111b3f504fb99686484b8f19a9c8d61a55eccc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://101fb592fef069af009c613a1e9c9825024ed205498ab72334211ba90d6f7397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://101fb592fef069af009c613a1e9c9825024ed205498ab72334211ba90d6f7397\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:07Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:07 crc kubenswrapper[4897]: I1121 14:10:07.687956 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b78c993a-9966-4e84-85a1-b4f73aaf17f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce3a2602338a1c1e9380ee55f78360f3d9cf6ccc62678886dc5ae574a72e3d9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c256823f2fb49ecc121f6ef564cd86196f9b4a473e9bdea670173596ef14a823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c65f71f8746044f5d41687c670a0308cd09dc41a6531f9b5720e852b3d240756\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1dbe88e7e6ad63899e107ebb884a16edb9a42b21ae7c86d7fbfaf89785b0d36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:07Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:07 crc kubenswrapper[4897]: I1121 14:10:07.698974 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09877d9780caddd5f7f73e95fdf818a427ed645ed5997984113151f4b4759c13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:07Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:07 crc kubenswrapper[4897]: I1121 14:10:07.712359 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:07Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:07 crc kubenswrapper[4897]: I1121 14:10:07.716074 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:07 crc kubenswrapper[4897]: I1121 14:10:07.716317 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:07 crc kubenswrapper[4897]: I1121 14:10:07.716492 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:07 crc kubenswrapper[4897]: I1121 14:10:07.718649 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:07 crc kubenswrapper[4897]: I1121 14:10:07.718690 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:07Z","lastTransitionTime":"2025-11-21T14:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:07 crc kubenswrapper[4897]: I1121 14:10:07.730406 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c976d804a7ebeba2c879e920cbc389f96f7f5def4b4439e8c8980d5460ecdb65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:07Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:07 crc kubenswrapper[4897]: I1121 14:10:07.743032 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nl76g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a7d6101-d0d2-4828-b3ce-d01040c04640\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac58b6ef1bd08e5f950260c18120c72e50b90a0de182c6ab0a5dbbbe45cf7328\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e56402f62821baf335f378084f4ed7c21b60ccbafe148419a0a9a403f106ad19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e56402f62821baf335f378084f4ed7c21b60ccbafe148419a0a9a403f106ad19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6022ff28e83dafcd846a7f03b090262aa3229a0da89f1984e4b6a054a04ac24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6022ff28e83dafcd846a7f03b090262aa3229a0da89f1984e4b6a054a04ac24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89de56e32b796da11bf89dd598d04fbf2884745ddf0f5d87cccfcaf1c001d8d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89de56e32b796da11bf89dd598d04fbf2884745ddf0f5d87cccfcaf1c001d8d6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2dff3cacaba7f9238f482e67b670ea775e0f96621434b99f4e76c2dfdde0321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2dff3cacaba7f9238f482e67b670ea775e0f96621434b99f4e76c2dfdde0321\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a153389a4137ee580e5118f3248400460f9d014c992a956f508815755f583afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a153389a4137ee580e5118f3248400460f9d014c992a956f508815755f583afb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a8c051ad1c1f2373f79eb0781fa6db489aeea3d52ed6c0bb43eec11520ec2b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a8c051ad1c1f2373f79eb0781fa6db489aeea3d52ed6c0bb43eec11520ec2b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nl76g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:07Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:07 crc kubenswrapper[4897]: I1121 14:10:07.755591 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48c6b1ea7497b33c00b956e400426c1d4f3c6c66e0225133434460800eef9314\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ea6068a6b6f8123b402de354e5374a0f02a7191a922d1718ab38f8898f47c52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:07Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:07 crc kubenswrapper[4897]: I1121 14:10:07.766338 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4qwqj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f46e2ba8-9ee0-406d-a668-2ba424618b69\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f43aa89d4d9defcb50d0dc6f90293b5f426b9cfc398b780c2e10f6a5c9f1e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbsgb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4qwqj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:07Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:07 crc kubenswrapper[4897]: I1121 14:10:07.775580 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-rs4c8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7xkdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7xkdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:23Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-rs4c8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:07Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:07 crc kubenswrapper[4897]: I1121 14:10:07.785093 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7qfmq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"223b7002-3e67-4848-b1fb-db2921deb64e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381a46c6290da2b36e572f46c16f1a50aa3e25359bcfbc02c7a5a18e1484860e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2m27\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7qfmq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:07Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:07 crc kubenswrapper[4897]: I1121 14:10:07.798343 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"402271c9-35e5-4bed-8410-c84c2e132878\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08651f8381818947818e0bb640c089be4a07022efc650befbb65ad58466bf8c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://419ab9403696491ef53a3d812d4c19d676f8fb1a1a4dfcdfcc0c6255377d2918\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://acc9629bb6b0dfbe5b4efe2f9f980e4e3865fdc6742782ebc49a5a7e0f775b9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0fc71bebc463919826d01260f38b63bfdb01199fd8dd6ae90d9ac9cd75db703\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0fc71bebc463919826d01260f38b63bfdb01199fd8dd6ae90d9ac9cd75db703\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:07Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:07 crc kubenswrapper[4897]: I1121 14:10:07.811480 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:07Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:07 crc kubenswrapper[4897]: I1121 14:10:07.820930 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:07 crc kubenswrapper[4897]: I1121 14:10:07.820979 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:07 crc kubenswrapper[4897]: I1121 14:10:07.820991 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:07 crc kubenswrapper[4897]: I1121 14:10:07.821009 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:07 crc kubenswrapper[4897]: I1121 14:10:07.821021 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:07Z","lastTransitionTime":"2025-11-21T14:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:07 crc kubenswrapper[4897]: I1121 14:10:07.824685 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7670227-d280-4847-b882-754429f56b0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://961961696b8750c248598f4cda032b18395e5c802275dcfe6d3e9f7ef3510f1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5854028821e6a360a259427b68e7bb4c2a7220e0c68f4cc91639b2ece0540a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-krv5b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:07Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:07 crc kubenswrapper[4897]: I1121 14:10:07.842444 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"222f65f2-0b82-4760-bcf7-779244720f01\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92f1d932a83d65ba6a9a906cc489ff51e76547d07d42656630c59421f833c4db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccd51aa0faf3b1327f207b355cddf86a7353c0051f7e5ce0396f3c4bc67e3b16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8971c38daad45bda3d61545f52120dad40992dbf2e1d6e1e776500ed58d73f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfee19f82e35869a641318a0c45bc7a2ae71d80039f5790e60e277be15c9d3ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec279f2f558a8ac653006321b18cbaf13185876fe303c5cc1c65988aef09edbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de095f3c760b13d5197f624eac3806a1f4186577f2b95eb522532d453eb266e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40b0bddce48b3626ad889ae2f43deb866acd521a8aab71253506364bfb47c723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e285a47c1193b01d8cd3cb16ca750bea2f24d3ad941dd438f96d5203f9203de\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-21T14:09:38Z\\\",\\\"message\\\":\\\":776] Recording success event on pod openshift-etcd/etcd-crc\\\\nI1121 14:09:38.005810 6569 default_network_controller.go:776] Recording success event on pod openshift-image-registry/node-ca-7qfmq\\\\nI1121 14:09:38.005827 6569 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-additional-cni-plugins-nl76g\\\\nI1121 14:09:38.005832 6569 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-additional-cni-plugins-nl76g\\\\nI1121 14:09:38.005837 6569 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-additional-cni-plugins-nl76g in node crc\\\\nI1121 14:09:38.005842 6569 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-additional-cni-plugins-nl76g after 0 failed attempt(s)\\\\nI1121 14:09:38.005846 6569 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-additional-cni-plugins-nl76g\\\\nI1121 14:09:38.005834 6569 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-operator-lifecycle-manager/package-server-manager-metrics]} name:Service_openshift-operator-lifecycle-manager/package-server-manager-metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:37Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://40b0bddce48b3626ad889ae2f43deb866acd521a8aab71253506364bfb47c723\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-21T14:10:06Z\\\",\\\"message\\\":\\\"25947 6970 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: failed to add event handler: handler {0x1e60340 0x1e60020 0x1e5ffc0} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:06Z is after 2025-08-24T17:21:41Z]\\\\nI1121 14:10:06.825948 6970 services_controller.go:452] Built service openshift-kube-scheduler/scheduler per-node LB for network=default: []services.LB{}\\\\nI1121 14:10:06.825939 6970 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-ingress-operator/metrics\\\\\\\"}\\\\nI1121 14:10:06.825960 6970 services_controller.go:453] Built service openshift-kube-scheduler/scheduler template LB for network=default: []services.LB{}\\\\nI1121 14:10:06.825946 6970 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-21T14:10:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b486a86e0e97ab409e590f30a1bf5cb9f584a7c67f96ff1236e42ca4f8675b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7877db6f5c2bf0eb31218c64d67bca48f921ff7ac62b4e495f29024c086bf738\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7877db6f5c2bf0eb31218c64d67bca48f921ff7ac62b4e495f29024c086bf738\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tpfb6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:07Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:07 crc kubenswrapper[4897]: I1121 14:10:07.923646 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:07 crc kubenswrapper[4897]: I1121 14:10:07.923695 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:07 crc kubenswrapper[4897]: I1121 14:10:07.923708 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:07 crc kubenswrapper[4897]: I1121 14:10:07.923725 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:07 crc kubenswrapper[4897]: I1121 14:10:07.923736 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:07Z","lastTransitionTime":"2025-11-21T14:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:08 crc kubenswrapper[4897]: I1121 14:10:08.027980 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:08 crc kubenswrapper[4897]: I1121 14:10:08.028030 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:08 crc kubenswrapper[4897]: I1121 14:10:08.028041 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:08 crc kubenswrapper[4897]: I1121 14:10:08.028058 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:08 crc kubenswrapper[4897]: I1121 14:10:08.028070 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:08Z","lastTransitionTime":"2025-11-21T14:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:08 crc kubenswrapper[4897]: I1121 14:10:08.130287 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:08 crc kubenswrapper[4897]: I1121 14:10:08.130338 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:08 crc kubenswrapper[4897]: I1121 14:10:08.130350 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:08 crc kubenswrapper[4897]: I1121 14:10:08.130370 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:08 crc kubenswrapper[4897]: I1121 14:10:08.130382 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:08Z","lastTransitionTime":"2025-11-21T14:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:08 crc kubenswrapper[4897]: I1121 14:10:08.232624 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:08 crc kubenswrapper[4897]: I1121 14:10:08.232654 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:08 crc kubenswrapper[4897]: I1121 14:10:08.232663 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:08 crc kubenswrapper[4897]: I1121 14:10:08.232676 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:08 crc kubenswrapper[4897]: I1121 14:10:08.232685 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:08Z","lastTransitionTime":"2025-11-21T14:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:08 crc kubenswrapper[4897]: I1121 14:10:08.335484 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:08 crc kubenswrapper[4897]: I1121 14:10:08.335556 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:08 crc kubenswrapper[4897]: I1121 14:10:08.335573 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:08 crc kubenswrapper[4897]: I1121 14:10:08.335593 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:08 crc kubenswrapper[4897]: I1121 14:10:08.335605 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:08Z","lastTransitionTime":"2025-11-21T14:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:08 crc kubenswrapper[4897]: I1121 14:10:08.438201 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:08 crc kubenswrapper[4897]: I1121 14:10:08.438270 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:08 crc kubenswrapper[4897]: I1121 14:10:08.438292 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:08 crc kubenswrapper[4897]: I1121 14:10:08.438320 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:08 crc kubenswrapper[4897]: I1121 14:10:08.438339 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:08Z","lastTransitionTime":"2025-11-21T14:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:08 crc kubenswrapper[4897]: I1121 14:10:08.541065 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:08 crc kubenswrapper[4897]: I1121 14:10:08.541124 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:08 crc kubenswrapper[4897]: I1121 14:10:08.541149 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:08 crc kubenswrapper[4897]: I1121 14:10:08.541174 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:08 crc kubenswrapper[4897]: I1121 14:10:08.541191 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:08Z","lastTransitionTime":"2025-11-21T14:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:08 crc kubenswrapper[4897]: I1121 14:10:08.604134 4897 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tpfb6_222f65f2-0b82-4760-bcf7-779244720f01/ovnkube-controller/3.log" Nov 21 14:10:08 crc kubenswrapper[4897]: I1121 14:10:08.607932 4897 scope.go:117] "RemoveContainer" containerID="40b0bddce48b3626ad889ae2f43deb866acd521a8aab71253506364bfb47c723" Nov 21 14:10:08 crc kubenswrapper[4897]: E1121 14:10:08.608122 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-tpfb6_openshift-ovn-kubernetes(222f65f2-0b82-4760-bcf7-779244720f01)\"" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" podUID="222f65f2-0b82-4760-bcf7-779244720f01" Nov 21 14:10:08 crc kubenswrapper[4897]: I1121 14:10:08.623177 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48c6b1ea7497b33c00b956e400426c1d4f3c6c66e0225133434460800eef9314\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ea6068a6b6f8123b402de354e5374a0f02a7191a922d1718ab38f8898f47c52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:08Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:08 crc kubenswrapper[4897]: I1121 14:10:08.636354 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4qwqj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f46e2ba8-9ee0-406d-a668-2ba424618b69\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f43aa89d4d9defcb50d0dc6f90293b5f426b9cfc398b780c2e10f6a5c9f1e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbsgb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4qwqj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:08Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:08 crc kubenswrapper[4897]: I1121 14:10:08.644193 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:08 crc kubenswrapper[4897]: I1121 14:10:08.644245 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:08 crc kubenswrapper[4897]: I1121 14:10:08.644256 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:08 crc kubenswrapper[4897]: I1121 14:10:08.644272 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:08 crc kubenswrapper[4897]: I1121 14:10:08.644284 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:08Z","lastTransitionTime":"2025-11-21T14:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:08 crc kubenswrapper[4897]: I1121 14:10:08.649151 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-rs4c8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7xkdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7xkdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:23Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-rs4c8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:08Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:08 crc kubenswrapper[4897]: I1121 14:10:08.662233 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7qfmq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"223b7002-3e67-4848-b1fb-db2921deb64e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381a46c6290da2b36e572f46c16f1a50aa3e25359bcfbc02c7a5a18e1484860e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2m27\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7qfmq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:08Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:08 crc kubenswrapper[4897]: I1121 14:10:08.677975 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"402271c9-35e5-4bed-8410-c84c2e132878\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08651f8381818947818e0bb640c089be4a07022efc650befbb65ad58466bf8c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://419ab9403696491ef53a3d812d4c19d676f8fb1a1a4dfcdfcc0c6255377d2918\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://acc9629bb6b0dfbe5b4efe2f9f980e4e3865fdc6742782ebc49a5a7e0f775b9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0fc71bebc463919826d01260f38b63bfdb01199fd8dd6ae90d9ac9cd75db703\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0fc71bebc463919826d01260f38b63bfdb01199fd8dd6ae90d9ac9cd75db703\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:08Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:08 crc kubenswrapper[4897]: I1121 14:10:08.692155 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:08Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:08 crc kubenswrapper[4897]: I1121 14:10:08.703202 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7670227-d280-4847-b882-754429f56b0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://961961696b8750c248598f4cda032b18395e5c802275dcfe6d3e9f7ef3510f1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5854028821e6a360a259427b68e7bb4c2a7220e0c68f4cc91639b2ece0540a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-krv5b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:08Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:08 crc kubenswrapper[4897]: I1121 14:10:08.725158 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"222f65f2-0b82-4760-bcf7-779244720f01\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92f1d932a83d65ba6a9a906cc489ff51e76547d07d42656630c59421f833c4db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccd51aa0faf3b1327f207b355cddf86a7353c0051f7e5ce0396f3c4bc67e3b16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8971c38daad45bda3d61545f52120dad40992dbf2e1d6e1e776500ed58d73f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfee19f82e35869a641318a0c45bc7a2ae71d80039f5790e60e277be15c9d3ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec279f2f558a8ac653006321b18cbaf13185876fe303c5cc1c65988aef09edbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de095f3c760b13d5197f624eac3806a1f4186577f2b95eb522532d453eb266e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40b0bddce48b3626ad889ae2f43deb866acd521a8aab71253506364bfb47c723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://40b0bddce48b3626ad889ae2f43deb866acd521a8aab71253506364bfb47c723\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-21T14:10:06Z\\\",\\\"message\\\":\\\"25947 6970 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: failed to add event handler: handler {0x1e60340 0x1e60020 0x1e5ffc0} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:06Z is after 2025-08-24T17:21:41Z]\\\\nI1121 14:10:06.825948 6970 services_controller.go:452] Built service openshift-kube-scheduler/scheduler per-node LB for network=default: []services.LB{}\\\\nI1121 14:10:06.825939 6970 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-ingress-operator/metrics\\\\\\\"}\\\\nI1121 14:10:06.825960 6970 services_controller.go:453] Built service openshift-kube-scheduler/scheduler template LB for network=default: []services.LB{}\\\\nI1121 14:10:06.825946 6970 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-21T14:10:06Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-tpfb6_openshift-ovn-kubernetes(222f65f2-0b82-4760-bcf7-779244720f01)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b486a86e0e97ab409e590f30a1bf5cb9f584a7c67f96ff1236e42ca4f8675b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7877db6f5c2bf0eb31218c64d67bca48f921ff7ac62b4e495f29024c086bf738\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7877db6f5c2bf0eb31218c64d67bca48f921ff7ac62b4e495f29024c086bf738\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tpfb6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:08Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:08 crc kubenswrapper[4897]: I1121 14:10:08.741948 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:08Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:08 crc kubenswrapper[4897]: I1121 14:10:08.746249 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:08 crc kubenswrapper[4897]: I1121 14:10:08.746305 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:08 crc kubenswrapper[4897]: I1121 14:10:08.746316 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:08 crc kubenswrapper[4897]: I1121 14:10:08.746333 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:08 crc kubenswrapper[4897]: I1121 14:10:08.746345 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:08Z","lastTransitionTime":"2025-11-21T14:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:08 crc kubenswrapper[4897]: I1121 14:10:08.755959 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6ljgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f3b70b5-9294-4f69-9528-500d28f34c89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:10:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:10:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf92ab70c5a3196bf83f427ccf03acf188d1995766a9154bcd6267e3926c3ac8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://527cf28abdeef1b42d9c19fb88e3171dc05c4516c3a542a461b5b1daa85c6d25\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-21T14:09:58Z\\\",\\\"message\\\":\\\"2025-11-21T14:09:13+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_a63cd511-2e05-4faf-b47b-8037110976d1\\\\n2025-11-21T14:09:13+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_a63cd511-2e05-4faf-b47b-8037110976d1 to /host/opt/cni/bin/\\\\n2025-11-21T14:09:13Z [verbose] multus-daemon started\\\\n2025-11-21T14:09:13Z [verbose] Readiness Indicator file check\\\\n2025-11-21T14:09:58Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78jhb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6ljgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:08Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:08 crc kubenswrapper[4897]: I1121 14:10:08.766729 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x5pws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc9e717e-af03-48bf-b005-62b3d084c6fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f968eee1d8a1888229e23abda59fd58fa5e86f02ca96ff80fbe093904b27e8df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9c4f5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0f99b9213902c003e6ca754134d2482404ace39c184fb3b431d2b600481d5b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9c4f5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-x5pws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:08Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:08 crc kubenswrapper[4897]: I1121 14:10:08.789150 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6cd330fd-e8a6-49d5-bfc9-5f6122ccd49e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1bd64b9b938bc80284bcee54141e90cad613d339c4f27ef2934b29bc2a318916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://936f59996d26e42099dcd055f87fff9e3ac4283b57e1956fe97463f2334152d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca74b28a069dc2e609b0779164c3aa470a15ed9848af0dd5fc3479aaf750157b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://270c53b9a342f26be077a01cd4e87e9ccd0eb5d2c84208ae38cbf720463245e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66c521c26d541bbddb064dd9710f1a7b41090d2f47343977922272dfc6dab7af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6eb2ac410b49098e5b3603de5e6bf5ed4484fab0b300021d0bed2e5ed8a42527\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6eb2ac410b49098e5b3603de5e6bf5ed4484fab0b300021d0bed2e5ed8a42527\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03385e0c78e4806f60c160227ef4905a11003ed3d70f0e5c0dd0a477979d1d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03385e0c78e4806f60c160227ef4905a11003ed3d70f0e5c0dd0a477979d1d4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://50aea2fa8b3b141b4330ddce2c8e0715230c6f06dab64b3abfd98a3067224bec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50aea2fa8b3b141b4330ddce2c8e0715230c6f06dab64b3abfd98a3067224bec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:08Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:08 crc kubenswrapper[4897]: I1121 14:10:08.801667 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a6c0a99-565b-41f9-a8b5-c8990a0c8eb3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://737553dd1d16e2019ea89a376e3a868f9792ed5239aa7d6cc40a05d3d4f286ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6682a8886d2b9bce069cf9261d4701ef6907f19574fd035bd89bd05626023e5c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de079956e3d8de0101f3f87dcc120ef23876c4972382639e419b3f34a38cdfca\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://469b5be7467d4a58d3776ec0a967ed911be31b770b49965da4f5d2fb264d8a3d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1907bfbad7eb34da642e30d6aa81820aab457f92347333a364ac12396b0e9dae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"le observer\\\\nW1121 14:09:07.994401 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1121 14:09:07.994601 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1121 14:09:07.995290 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1461162674/tls.crt::/tmp/serving-cert-1461162674/tls.key\\\\\\\"\\\\nI1121 14:09:08.162382 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1121 14:09:08.165301 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1121 14:09:08.165318 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1121 14:09:08.165343 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1121 14:09:08.165348 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1121 14:09:08.172015 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1121 14:09:08.172083 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1121 14:09:08.172089 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1121 14:09:08.172094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1121 14:09:08.172098 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1121 14:09:08.172102 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1121 14:09:08.172106 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1121 14:09:08.172023 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1121 14:09:08.174214 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d78c87c440c7d8e30075f6143111b3f504fb99686484b8f19a9c8d61a55eccc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://101fb592fef069af009c613a1e9c9825024ed205498ab72334211ba90d6f7397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://101fb592fef069af009c613a1e9c9825024ed205498ab72334211ba90d6f7397\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:08Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:08 crc kubenswrapper[4897]: I1121 14:10:08.813180 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b78c993a-9966-4e84-85a1-b4f73aaf17f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce3a2602338a1c1e9380ee55f78360f3d9cf6ccc62678886dc5ae574a72e3d9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c256823f2fb49ecc121f6ef564cd86196f9b4a473e9bdea670173596ef14a823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c65f71f8746044f5d41687c670a0308cd09dc41a6531f9b5720e852b3d240756\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1dbe88e7e6ad63899e107ebb884a16edb9a42b21ae7c86d7fbfaf89785b0d36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:08Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:08 crc kubenswrapper[4897]: I1121 14:10:08.826368 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09877d9780caddd5f7f73e95fdf818a427ed645ed5997984113151f4b4759c13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:08Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:08 crc kubenswrapper[4897]: I1121 14:10:08.839661 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:08Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:08 crc kubenswrapper[4897]: I1121 14:10:08.848671 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:08 crc kubenswrapper[4897]: I1121 14:10:08.848742 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:08 crc kubenswrapper[4897]: I1121 14:10:08.848757 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:08 crc kubenswrapper[4897]: I1121 14:10:08.848779 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:08 crc kubenswrapper[4897]: I1121 14:10:08.848795 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:08Z","lastTransitionTime":"2025-11-21T14:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:08 crc kubenswrapper[4897]: I1121 14:10:08.856047 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c976d804a7ebeba2c879e920cbc389f96f7f5def4b4439e8c8980d5460ecdb65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:08Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:08 crc kubenswrapper[4897]: I1121 14:10:08.868222 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nl76g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a7d6101-d0d2-4828-b3ce-d01040c04640\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac58b6ef1bd08e5f950260c18120c72e50b90a0de182c6ab0a5dbbbe45cf7328\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e56402f62821baf335f378084f4ed7c21b60ccbafe148419a0a9a403f106ad19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e56402f62821baf335f378084f4ed7c21b60ccbafe148419a0a9a403f106ad19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6022ff28e83dafcd846a7f03b090262aa3229a0da89f1984e4b6a054a04ac24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6022ff28e83dafcd846a7f03b090262aa3229a0da89f1984e4b6a054a04ac24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89de56e32b796da11bf89dd598d04fbf2884745ddf0f5d87cccfcaf1c001d8d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89de56e32b796da11bf89dd598d04fbf2884745ddf0f5d87cccfcaf1c001d8d6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2dff3cacaba7f9238f482e67b670ea775e0f96621434b99f4e76c2dfdde0321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2dff3cacaba7f9238f482e67b670ea775e0f96621434b99f4e76c2dfdde0321\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a153389a4137ee580e5118f3248400460f9d014c992a956f508815755f583afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a153389a4137ee580e5118f3248400460f9d014c992a956f508815755f583afb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a8c051ad1c1f2373f79eb0781fa6db489aeea3d52ed6c0bb43eec11520ec2b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a8c051ad1c1f2373f79eb0781fa6db489aeea3d52ed6c0bb43eec11520ec2b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nl76g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:08Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:08 crc kubenswrapper[4897]: I1121 14:10:08.951623 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:08 crc kubenswrapper[4897]: I1121 14:10:08.951659 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:08 crc kubenswrapper[4897]: I1121 14:10:08.951668 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:08 crc kubenswrapper[4897]: I1121 14:10:08.951683 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:08 crc kubenswrapper[4897]: I1121 14:10:08.951694 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:08Z","lastTransitionTime":"2025-11-21T14:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:09 crc kubenswrapper[4897]: I1121 14:10:09.055179 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:09 crc kubenswrapper[4897]: I1121 14:10:09.055228 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:09 crc kubenswrapper[4897]: I1121 14:10:09.055240 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:09 crc kubenswrapper[4897]: I1121 14:10:09.055259 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:09 crc kubenswrapper[4897]: I1121 14:10:09.055272 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:09Z","lastTransitionTime":"2025-11-21T14:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:09 crc kubenswrapper[4897]: I1121 14:10:09.088811 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 21 14:10:09 crc kubenswrapper[4897]: I1121 14:10:09.088860 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 21 14:10:09 crc kubenswrapper[4897]: I1121 14:10:09.088901 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs4c8" Nov 21 14:10:09 crc kubenswrapper[4897]: I1121 14:10:09.088811 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 21 14:10:09 crc kubenswrapper[4897]: E1121 14:10:09.088981 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 21 14:10:09 crc kubenswrapper[4897]: E1121 14:10:09.089022 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs4c8" podUID="01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd" Nov 21 14:10:09 crc kubenswrapper[4897]: E1121 14:10:09.089076 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 21 14:10:09 crc kubenswrapper[4897]: E1121 14:10:09.089191 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 21 14:10:09 crc kubenswrapper[4897]: I1121 14:10:09.157284 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:09 crc kubenswrapper[4897]: I1121 14:10:09.157341 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:09 crc kubenswrapper[4897]: I1121 14:10:09.157354 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:09 crc kubenswrapper[4897]: I1121 14:10:09.157374 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:09 crc kubenswrapper[4897]: I1121 14:10:09.157390 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:09Z","lastTransitionTime":"2025-11-21T14:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:09 crc kubenswrapper[4897]: I1121 14:10:09.259557 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:09 crc kubenswrapper[4897]: I1121 14:10:09.259607 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:09 crc kubenswrapper[4897]: I1121 14:10:09.259630 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:09 crc kubenswrapper[4897]: I1121 14:10:09.259659 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:09 crc kubenswrapper[4897]: I1121 14:10:09.259674 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:09Z","lastTransitionTime":"2025-11-21T14:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:09 crc kubenswrapper[4897]: I1121 14:10:09.362085 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:09 crc kubenswrapper[4897]: I1121 14:10:09.362135 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:09 crc kubenswrapper[4897]: I1121 14:10:09.362148 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:09 crc kubenswrapper[4897]: I1121 14:10:09.362169 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:09 crc kubenswrapper[4897]: I1121 14:10:09.362184 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:09Z","lastTransitionTime":"2025-11-21T14:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:09 crc kubenswrapper[4897]: I1121 14:10:09.464784 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:09 crc kubenswrapper[4897]: I1121 14:10:09.464854 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:09 crc kubenswrapper[4897]: I1121 14:10:09.464877 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:09 crc kubenswrapper[4897]: I1121 14:10:09.464910 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:09 crc kubenswrapper[4897]: I1121 14:10:09.464936 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:09Z","lastTransitionTime":"2025-11-21T14:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:09 crc kubenswrapper[4897]: I1121 14:10:09.567940 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:09 crc kubenswrapper[4897]: I1121 14:10:09.568000 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:09 crc kubenswrapper[4897]: I1121 14:10:09.568016 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:09 crc kubenswrapper[4897]: I1121 14:10:09.568037 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:09 crc kubenswrapper[4897]: I1121 14:10:09.568051 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:09Z","lastTransitionTime":"2025-11-21T14:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:09 crc kubenswrapper[4897]: I1121 14:10:09.671150 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:09 crc kubenswrapper[4897]: I1121 14:10:09.671195 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:09 crc kubenswrapper[4897]: I1121 14:10:09.671206 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:09 crc kubenswrapper[4897]: I1121 14:10:09.671223 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:09 crc kubenswrapper[4897]: I1121 14:10:09.671235 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:09Z","lastTransitionTime":"2025-11-21T14:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:09 crc kubenswrapper[4897]: I1121 14:10:09.774310 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:09 crc kubenswrapper[4897]: I1121 14:10:09.774358 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:09 crc kubenswrapper[4897]: I1121 14:10:09.774366 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:09 crc kubenswrapper[4897]: I1121 14:10:09.774381 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:09 crc kubenswrapper[4897]: I1121 14:10:09.774392 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:09Z","lastTransitionTime":"2025-11-21T14:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:09 crc kubenswrapper[4897]: I1121 14:10:09.877327 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:09 crc kubenswrapper[4897]: I1121 14:10:09.877381 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:09 crc kubenswrapper[4897]: I1121 14:10:09.877392 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:09 crc kubenswrapper[4897]: I1121 14:10:09.877408 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:09 crc kubenswrapper[4897]: I1121 14:10:09.877422 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:09Z","lastTransitionTime":"2025-11-21T14:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:09 crc kubenswrapper[4897]: I1121 14:10:09.980693 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:09 crc kubenswrapper[4897]: I1121 14:10:09.980740 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:09 crc kubenswrapper[4897]: I1121 14:10:09.980754 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:09 crc kubenswrapper[4897]: I1121 14:10:09.980774 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:09 crc kubenswrapper[4897]: I1121 14:10:09.980787 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:09Z","lastTransitionTime":"2025-11-21T14:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:10 crc kubenswrapper[4897]: I1121 14:10:10.083207 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:10 crc kubenswrapper[4897]: I1121 14:10:10.083270 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:10 crc kubenswrapper[4897]: I1121 14:10:10.083289 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:10 crc kubenswrapper[4897]: I1121 14:10:10.083312 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:10 crc kubenswrapper[4897]: I1121 14:10:10.083331 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:10Z","lastTransitionTime":"2025-11-21T14:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:10 crc kubenswrapper[4897]: I1121 14:10:10.186304 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:10 crc kubenswrapper[4897]: I1121 14:10:10.186354 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:10 crc kubenswrapper[4897]: I1121 14:10:10.186366 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:10 crc kubenswrapper[4897]: I1121 14:10:10.186385 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:10 crc kubenswrapper[4897]: I1121 14:10:10.186401 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:10Z","lastTransitionTime":"2025-11-21T14:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:10 crc kubenswrapper[4897]: I1121 14:10:10.289110 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:10 crc kubenswrapper[4897]: I1121 14:10:10.289153 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:10 crc kubenswrapper[4897]: I1121 14:10:10.289167 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:10 crc kubenswrapper[4897]: I1121 14:10:10.289185 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:10 crc kubenswrapper[4897]: I1121 14:10:10.289197 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:10Z","lastTransitionTime":"2025-11-21T14:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:10 crc kubenswrapper[4897]: I1121 14:10:10.391838 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:10 crc kubenswrapper[4897]: I1121 14:10:10.391882 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:10 crc kubenswrapper[4897]: I1121 14:10:10.391896 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:10 crc kubenswrapper[4897]: I1121 14:10:10.391914 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:10 crc kubenswrapper[4897]: I1121 14:10:10.391927 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:10Z","lastTransitionTime":"2025-11-21T14:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:10 crc kubenswrapper[4897]: I1121 14:10:10.494412 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:10 crc kubenswrapper[4897]: I1121 14:10:10.494456 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:10 crc kubenswrapper[4897]: I1121 14:10:10.494468 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:10 crc kubenswrapper[4897]: I1121 14:10:10.494485 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:10 crc kubenswrapper[4897]: I1121 14:10:10.494496 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:10Z","lastTransitionTime":"2025-11-21T14:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:10 crc kubenswrapper[4897]: I1121 14:10:10.596097 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:10 crc kubenswrapper[4897]: I1121 14:10:10.596144 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:10 crc kubenswrapper[4897]: I1121 14:10:10.596156 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:10 crc kubenswrapper[4897]: I1121 14:10:10.596171 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:10 crc kubenswrapper[4897]: I1121 14:10:10.596182 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:10Z","lastTransitionTime":"2025-11-21T14:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:10 crc kubenswrapper[4897]: I1121 14:10:10.699425 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:10 crc kubenswrapper[4897]: I1121 14:10:10.699529 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:10 crc kubenswrapper[4897]: I1121 14:10:10.699549 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:10 crc kubenswrapper[4897]: I1121 14:10:10.699584 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:10 crc kubenswrapper[4897]: I1121 14:10:10.699606 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:10Z","lastTransitionTime":"2025-11-21T14:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:10 crc kubenswrapper[4897]: I1121 14:10:10.802809 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:10 crc kubenswrapper[4897]: I1121 14:10:10.802878 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:10 crc kubenswrapper[4897]: I1121 14:10:10.802896 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:10 crc kubenswrapper[4897]: I1121 14:10:10.802921 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:10 crc kubenswrapper[4897]: I1121 14:10:10.802938 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:10Z","lastTransitionTime":"2025-11-21T14:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:10 crc kubenswrapper[4897]: I1121 14:10:10.905934 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:10 crc kubenswrapper[4897]: I1121 14:10:10.905981 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:10 crc kubenswrapper[4897]: I1121 14:10:10.905990 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:10 crc kubenswrapper[4897]: I1121 14:10:10.906013 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:10 crc kubenswrapper[4897]: I1121 14:10:10.906025 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:10Z","lastTransitionTime":"2025-11-21T14:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:11 crc kubenswrapper[4897]: I1121 14:10:11.010004 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:11 crc kubenswrapper[4897]: I1121 14:10:11.010059 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:11 crc kubenswrapper[4897]: I1121 14:10:11.010070 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:11 crc kubenswrapper[4897]: I1121 14:10:11.010090 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:11 crc kubenswrapper[4897]: I1121 14:10:11.010104 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:11Z","lastTransitionTime":"2025-11-21T14:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:11 crc kubenswrapper[4897]: I1121 14:10:11.088949 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 21 14:10:11 crc kubenswrapper[4897]: I1121 14:10:11.088949 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 21 14:10:11 crc kubenswrapper[4897]: E1121 14:10:11.089169 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 21 14:10:11 crc kubenswrapper[4897]: I1121 14:10:11.088995 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 21 14:10:11 crc kubenswrapper[4897]: I1121 14:10:11.088985 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs4c8" Nov 21 14:10:11 crc kubenswrapper[4897]: E1121 14:10:11.089338 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 21 14:10:11 crc kubenswrapper[4897]: E1121 14:10:11.089432 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs4c8" podUID="01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd" Nov 21 14:10:11 crc kubenswrapper[4897]: E1121 14:10:11.089589 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 21 14:10:11 crc kubenswrapper[4897]: I1121 14:10:11.113575 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:11 crc kubenswrapper[4897]: I1121 14:10:11.113635 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:11 crc kubenswrapper[4897]: I1121 14:10:11.113649 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:11 crc kubenswrapper[4897]: I1121 14:10:11.113674 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:11 crc kubenswrapper[4897]: I1121 14:10:11.113694 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:11Z","lastTransitionTime":"2025-11-21T14:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:11 crc kubenswrapper[4897]: I1121 14:10:11.216332 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:11 crc kubenswrapper[4897]: I1121 14:10:11.216376 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:11 crc kubenswrapper[4897]: I1121 14:10:11.216389 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:11 crc kubenswrapper[4897]: I1121 14:10:11.216405 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:11 crc kubenswrapper[4897]: I1121 14:10:11.216415 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:11Z","lastTransitionTime":"2025-11-21T14:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:11 crc kubenswrapper[4897]: I1121 14:10:11.318756 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:11 crc kubenswrapper[4897]: I1121 14:10:11.318798 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:11 crc kubenswrapper[4897]: I1121 14:10:11.318811 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:11 crc kubenswrapper[4897]: I1121 14:10:11.318829 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:11 crc kubenswrapper[4897]: I1121 14:10:11.318848 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:11Z","lastTransitionTime":"2025-11-21T14:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:11 crc kubenswrapper[4897]: I1121 14:10:11.422385 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:11 crc kubenswrapper[4897]: I1121 14:10:11.422456 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:11 crc kubenswrapper[4897]: I1121 14:10:11.422474 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:11 crc kubenswrapper[4897]: I1121 14:10:11.422532 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:11 crc kubenswrapper[4897]: I1121 14:10:11.422551 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:11Z","lastTransitionTime":"2025-11-21T14:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:11 crc kubenswrapper[4897]: I1121 14:10:11.525469 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:11 crc kubenswrapper[4897]: I1121 14:10:11.525527 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:11 crc kubenswrapper[4897]: I1121 14:10:11.525536 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:11 crc kubenswrapper[4897]: I1121 14:10:11.525555 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:11 crc kubenswrapper[4897]: I1121 14:10:11.525568 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:11Z","lastTransitionTime":"2025-11-21T14:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:11 crc kubenswrapper[4897]: I1121 14:10:11.628777 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:11 crc kubenswrapper[4897]: I1121 14:10:11.628820 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:11 crc kubenswrapper[4897]: I1121 14:10:11.628830 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:11 crc kubenswrapper[4897]: I1121 14:10:11.628847 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:11 crc kubenswrapper[4897]: I1121 14:10:11.628861 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:11Z","lastTransitionTime":"2025-11-21T14:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:11 crc kubenswrapper[4897]: I1121 14:10:11.733190 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:11 crc kubenswrapper[4897]: I1121 14:10:11.733266 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:11 crc kubenswrapper[4897]: I1121 14:10:11.733287 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:11 crc kubenswrapper[4897]: I1121 14:10:11.733313 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:11 crc kubenswrapper[4897]: I1121 14:10:11.733337 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:11Z","lastTransitionTime":"2025-11-21T14:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:11 crc kubenswrapper[4897]: I1121 14:10:11.836075 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:11 crc kubenswrapper[4897]: I1121 14:10:11.836475 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:11 crc kubenswrapper[4897]: I1121 14:10:11.836487 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:11 crc kubenswrapper[4897]: I1121 14:10:11.836530 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:11 crc kubenswrapper[4897]: I1121 14:10:11.836550 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:11Z","lastTransitionTime":"2025-11-21T14:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:11 crc kubenswrapper[4897]: I1121 14:10:11.939274 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:11 crc kubenswrapper[4897]: I1121 14:10:11.939315 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:11 crc kubenswrapper[4897]: I1121 14:10:11.939325 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:11 crc kubenswrapper[4897]: I1121 14:10:11.939342 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:11 crc kubenswrapper[4897]: I1121 14:10:11.939354 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:11Z","lastTransitionTime":"2025-11-21T14:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:12 crc kubenswrapper[4897]: I1121 14:10:12.041920 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:12 crc kubenswrapper[4897]: I1121 14:10:12.041979 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:12 crc kubenswrapper[4897]: I1121 14:10:12.041988 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:12 crc kubenswrapper[4897]: I1121 14:10:12.042002 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:12 crc kubenswrapper[4897]: I1121 14:10:12.042012 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:12Z","lastTransitionTime":"2025-11-21T14:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:12 crc kubenswrapper[4897]: I1121 14:10:12.144387 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:12 crc kubenswrapper[4897]: I1121 14:10:12.144435 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:12 crc kubenswrapper[4897]: I1121 14:10:12.144445 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:12 crc kubenswrapper[4897]: I1121 14:10:12.144464 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:12 crc kubenswrapper[4897]: I1121 14:10:12.144478 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:12Z","lastTransitionTime":"2025-11-21T14:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:12 crc kubenswrapper[4897]: I1121 14:10:12.247415 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:12 crc kubenswrapper[4897]: I1121 14:10:12.247477 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:12 crc kubenswrapper[4897]: I1121 14:10:12.247491 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:12 crc kubenswrapper[4897]: I1121 14:10:12.247548 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:12 crc kubenswrapper[4897]: I1121 14:10:12.247564 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:12Z","lastTransitionTime":"2025-11-21T14:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:12 crc kubenswrapper[4897]: I1121 14:10:12.351554 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:12 crc kubenswrapper[4897]: I1121 14:10:12.351622 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:12 crc kubenswrapper[4897]: I1121 14:10:12.351647 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:12 crc kubenswrapper[4897]: I1121 14:10:12.351674 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:12 crc kubenswrapper[4897]: I1121 14:10:12.351694 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:12Z","lastTransitionTime":"2025-11-21T14:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:12 crc kubenswrapper[4897]: I1121 14:10:12.454886 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:12 crc kubenswrapper[4897]: I1121 14:10:12.454939 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:12 crc kubenswrapper[4897]: I1121 14:10:12.454958 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:12 crc kubenswrapper[4897]: I1121 14:10:12.454978 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:12 crc kubenswrapper[4897]: I1121 14:10:12.454991 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:12Z","lastTransitionTime":"2025-11-21T14:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:12 crc kubenswrapper[4897]: I1121 14:10:12.558057 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:12 crc kubenswrapper[4897]: I1121 14:10:12.558104 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:12 crc kubenswrapper[4897]: I1121 14:10:12.558113 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:12 crc kubenswrapper[4897]: I1121 14:10:12.558128 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:12 crc kubenswrapper[4897]: I1121 14:10:12.558139 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:12Z","lastTransitionTime":"2025-11-21T14:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:12 crc kubenswrapper[4897]: I1121 14:10:12.661632 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:12 crc kubenswrapper[4897]: I1121 14:10:12.661662 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:12 crc kubenswrapper[4897]: I1121 14:10:12.661669 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:12 crc kubenswrapper[4897]: I1121 14:10:12.661682 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:12 crc kubenswrapper[4897]: I1121 14:10:12.661691 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:12Z","lastTransitionTime":"2025-11-21T14:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:12 crc kubenswrapper[4897]: I1121 14:10:12.765438 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:12 crc kubenswrapper[4897]: I1121 14:10:12.765499 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:12 crc kubenswrapper[4897]: I1121 14:10:12.765542 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:12 crc kubenswrapper[4897]: I1121 14:10:12.765560 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:12 crc kubenswrapper[4897]: I1121 14:10:12.765572 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:12Z","lastTransitionTime":"2025-11-21T14:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:12 crc kubenswrapper[4897]: I1121 14:10:12.868615 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:12 crc kubenswrapper[4897]: I1121 14:10:12.868693 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:12 crc kubenswrapper[4897]: I1121 14:10:12.868714 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:12 crc kubenswrapper[4897]: I1121 14:10:12.868742 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:12 crc kubenswrapper[4897]: I1121 14:10:12.868762 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:12Z","lastTransitionTime":"2025-11-21T14:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:12 crc kubenswrapper[4897]: I1121 14:10:12.972007 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 21 14:10:12 crc kubenswrapper[4897]: I1121 14:10:12.972189 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 21 14:10:12 crc kubenswrapper[4897]: E1121 14:10:12.972252 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-21 14:11:16.972214862 +0000 UTC m=+154.256808357 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:10:12 crc kubenswrapper[4897]: E1121 14:10:12.972340 4897 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 21 14:10:12 crc kubenswrapper[4897]: I1121 14:10:12.972317 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 21 14:10:12 crc kubenswrapper[4897]: E1121 14:10:12.972412 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-21 14:11:16.972391547 +0000 UTC m=+154.256985012 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 21 14:10:12 crc kubenswrapper[4897]: I1121 14:10:12.972557 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:12 crc kubenswrapper[4897]: I1121 14:10:12.972617 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:12 crc kubenswrapper[4897]: I1121 14:10:12.972631 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:12 crc kubenswrapper[4897]: I1121 14:10:12.972652 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:12 crc kubenswrapper[4897]: E1121 14:10:12.972645 4897 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 21 14:10:12 crc kubenswrapper[4897]: I1121 14:10:12.972665 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:12Z","lastTransitionTime":"2025-11-21T14:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:12 crc kubenswrapper[4897]: E1121 14:10:12.972781 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-21 14:11:16.972730287 +0000 UTC m=+154.257323882 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 21 14:10:13 crc kubenswrapper[4897]: I1121 14:10:13.073700 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 21 14:10:13 crc kubenswrapper[4897]: I1121 14:10:13.073807 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 21 14:10:13 crc kubenswrapper[4897]: E1121 14:10:13.073971 4897 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 21 14:10:13 crc kubenswrapper[4897]: E1121 14:10:13.073993 4897 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 21 14:10:13 crc kubenswrapper[4897]: E1121 14:10:13.074008 4897 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 21 14:10:13 crc kubenswrapper[4897]: E1121 14:10:13.074082 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-21 14:11:17.074052667 +0000 UTC m=+154.358646152 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 21 14:10:13 crc kubenswrapper[4897]: E1121 14:10:13.074208 4897 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 21 14:10:13 crc kubenswrapper[4897]: E1121 14:10:13.074251 4897 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 21 14:10:13 crc kubenswrapper[4897]: E1121 14:10:13.074269 4897 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 21 14:10:13 crc kubenswrapper[4897]: E1121 14:10:13.074358 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-21 14:11:17.074335335 +0000 UTC m=+154.358928810 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 21 14:10:13 crc kubenswrapper[4897]: I1121 14:10:13.075611 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:13 crc kubenswrapper[4897]: I1121 14:10:13.075678 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:13 crc kubenswrapper[4897]: I1121 14:10:13.075696 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:13 crc kubenswrapper[4897]: I1121 14:10:13.075724 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:13 crc kubenswrapper[4897]: I1121 14:10:13.075743 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:13Z","lastTransitionTime":"2025-11-21T14:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:13 crc kubenswrapper[4897]: I1121 14:10:13.088579 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 21 14:10:13 crc kubenswrapper[4897]: I1121 14:10:13.088606 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs4c8" Nov 21 14:10:13 crc kubenswrapper[4897]: I1121 14:10:13.088597 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 21 14:10:13 crc kubenswrapper[4897]: E1121 14:10:13.088793 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 21 14:10:13 crc kubenswrapper[4897]: I1121 14:10:13.088823 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 21 14:10:13 crc kubenswrapper[4897]: E1121 14:10:13.088995 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 21 14:10:13 crc kubenswrapper[4897]: E1121 14:10:13.089145 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 21 14:10:13 crc kubenswrapper[4897]: E1121 14:10:13.089261 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs4c8" podUID="01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd" Nov 21 14:10:13 crc kubenswrapper[4897]: I1121 14:10:13.179642 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:13 crc kubenswrapper[4897]: I1121 14:10:13.179720 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:13 crc kubenswrapper[4897]: I1121 14:10:13.179738 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:13 crc kubenswrapper[4897]: I1121 14:10:13.179769 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:13 crc kubenswrapper[4897]: I1121 14:10:13.179788 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:13Z","lastTransitionTime":"2025-11-21T14:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:13 crc kubenswrapper[4897]: I1121 14:10:13.283027 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:13 crc kubenswrapper[4897]: I1121 14:10:13.283078 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:13 crc kubenswrapper[4897]: I1121 14:10:13.283091 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:13 crc kubenswrapper[4897]: I1121 14:10:13.283111 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:13 crc kubenswrapper[4897]: I1121 14:10:13.283124 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:13Z","lastTransitionTime":"2025-11-21T14:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:13 crc kubenswrapper[4897]: I1121 14:10:13.385723 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:13 crc kubenswrapper[4897]: I1121 14:10:13.385798 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:13 crc kubenswrapper[4897]: I1121 14:10:13.385811 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:13 crc kubenswrapper[4897]: I1121 14:10:13.385870 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:13 crc kubenswrapper[4897]: I1121 14:10:13.385885 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:13Z","lastTransitionTime":"2025-11-21T14:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:13 crc kubenswrapper[4897]: I1121 14:10:13.489450 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:13 crc kubenswrapper[4897]: I1121 14:10:13.489547 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:13 crc kubenswrapper[4897]: I1121 14:10:13.489564 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:13 crc kubenswrapper[4897]: I1121 14:10:13.489589 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:13 crc kubenswrapper[4897]: I1121 14:10:13.489606 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:13Z","lastTransitionTime":"2025-11-21T14:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:13 crc kubenswrapper[4897]: I1121 14:10:13.593116 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:13 crc kubenswrapper[4897]: I1121 14:10:13.593210 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:13 crc kubenswrapper[4897]: I1121 14:10:13.593231 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:13 crc kubenswrapper[4897]: I1121 14:10:13.593253 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:13 crc kubenswrapper[4897]: I1121 14:10:13.593265 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:13Z","lastTransitionTime":"2025-11-21T14:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:13 crc kubenswrapper[4897]: I1121 14:10:13.696012 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:13 crc kubenswrapper[4897]: I1121 14:10:13.696086 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:13 crc kubenswrapper[4897]: I1121 14:10:13.696103 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:13 crc kubenswrapper[4897]: I1121 14:10:13.696131 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:13 crc kubenswrapper[4897]: I1121 14:10:13.696149 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:13Z","lastTransitionTime":"2025-11-21T14:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:13 crc kubenswrapper[4897]: I1121 14:10:13.798927 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:13 crc kubenswrapper[4897]: I1121 14:10:13.799003 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:13 crc kubenswrapper[4897]: I1121 14:10:13.799021 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:13 crc kubenswrapper[4897]: I1121 14:10:13.799050 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:13 crc kubenswrapper[4897]: I1121 14:10:13.799069 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:13Z","lastTransitionTime":"2025-11-21T14:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:13 crc kubenswrapper[4897]: I1121 14:10:13.902329 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:13 crc kubenswrapper[4897]: I1121 14:10:13.902367 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:13 crc kubenswrapper[4897]: I1121 14:10:13.902376 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:13 crc kubenswrapper[4897]: I1121 14:10:13.902390 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:13 crc kubenswrapper[4897]: I1121 14:10:13.902400 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:13Z","lastTransitionTime":"2025-11-21T14:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:14 crc kubenswrapper[4897]: I1121 14:10:14.004940 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:14 crc kubenswrapper[4897]: I1121 14:10:14.004987 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:14 crc kubenswrapper[4897]: I1121 14:10:14.004997 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:14 crc kubenswrapper[4897]: I1121 14:10:14.005048 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:14 crc kubenswrapper[4897]: I1121 14:10:14.005059 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:14Z","lastTransitionTime":"2025-11-21T14:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:14 crc kubenswrapper[4897]: I1121 14:10:14.107194 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4qwqj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f46e2ba8-9ee0-406d-a668-2ba424618b69\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f43aa89d4d9defcb50d0dc6f90293b5f426b9cfc398b780c2e10f6a5c9f1e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbsgb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4qwqj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:14Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:14 crc kubenswrapper[4897]: I1121 14:10:14.107973 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:14 crc kubenswrapper[4897]: I1121 14:10:14.108020 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:14 crc kubenswrapper[4897]: I1121 14:10:14.108036 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:14 crc kubenswrapper[4897]: I1121 14:10:14.108060 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:14 crc kubenswrapper[4897]: I1121 14:10:14.108077 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:14Z","lastTransitionTime":"2025-11-21T14:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:14 crc kubenswrapper[4897]: I1121 14:10:14.123984 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-rs4c8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7xkdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7xkdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:23Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-rs4c8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:14Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:14 crc kubenswrapper[4897]: I1121 14:10:14.147946 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48c6b1ea7497b33c00b956e400426c1d4f3c6c66e0225133434460800eef9314\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ea6068a6b6f8123b402de354e5374a0f02a7191a922d1718ab38f8898f47c52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:14Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:14 crc kubenswrapper[4897]: I1121 14:10:14.169169 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7670227-d280-4847-b882-754429f56b0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://961961696b8750c248598f4cda032b18395e5c802275dcfe6d3e9f7ef3510f1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5854028821e6a360a259427b68e7bb4c2a7220e0c68f4cc91639b2ece0540a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-krv5b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:14Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:14 crc kubenswrapper[4897]: I1121 14:10:14.206791 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"222f65f2-0b82-4760-bcf7-779244720f01\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92f1d932a83d65ba6a9a906cc489ff51e76547d07d42656630c59421f833c4db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccd51aa0faf3b1327f207b355cddf86a7353c0051f7e5ce0396f3c4bc67e3b16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8971c38daad45bda3d61545f52120dad40992dbf2e1d6e1e776500ed58d73f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cfee19f82e35869a641318a0c45bc7a2ae71d80039f5790e60e277be15c9d3ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec279f2f558a8ac653006321b18cbaf13185876fe303c5cc1c65988aef09edbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de095f3c760b13d5197f624eac3806a1f4186577f2b95eb522532d453eb266e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40b0bddce48b3626ad889ae2f43deb866acd521a8aab71253506364bfb47c723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://40b0bddce48b3626ad889ae2f43deb866acd521a8aab71253506364bfb47c723\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-21T14:10:06Z\\\",\\\"message\\\":\\\"25947 6970 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: failed to add event handler: handler {0x1e60340 0x1e60020 0x1e5ffc0} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:06Z is after 2025-08-24T17:21:41Z]\\\\nI1121 14:10:06.825948 6970 services_controller.go:452] Built service openshift-kube-scheduler/scheduler per-node LB for network=default: []services.LB{}\\\\nI1121 14:10:06.825939 6970 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-ingress-operator/metrics\\\\\\\"}\\\\nI1121 14:10:06.825960 6970 services_controller.go:453] Built service openshift-kube-scheduler/scheduler template LB for network=default: []services.LB{}\\\\nI1121 14:10:06.825946 6970 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-21T14:10:06Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-tpfb6_openshift-ovn-kubernetes(222f65f2-0b82-4760-bcf7-779244720f01)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b486a86e0e97ab409e590f30a1bf5cb9f584a7c67f96ff1236e42ca4f8675b56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7877db6f5c2bf0eb31218c64d67bca48f921ff7ac62b4e495f29024c086bf738\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7877db6f5c2bf0eb31218c64d67bca48f921ff7ac62b4e495f29024c086bf738\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mmvwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tpfb6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:14Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:14 crc kubenswrapper[4897]: I1121 14:10:14.210091 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:14 crc kubenswrapper[4897]: I1121 14:10:14.210118 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:14 crc kubenswrapper[4897]: I1121 14:10:14.210128 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:14 crc kubenswrapper[4897]: I1121 14:10:14.210144 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:14 crc kubenswrapper[4897]: I1121 14:10:14.210153 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:14Z","lastTransitionTime":"2025-11-21T14:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:14 crc kubenswrapper[4897]: I1121 14:10:14.217923 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7qfmq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"223b7002-3e67-4848-b1fb-db2921deb64e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://381a46c6290da2b36e572f46c16f1a50aa3e25359bcfbc02c7a5a18e1484860e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f2m27\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7qfmq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:14Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:14 crc kubenswrapper[4897]: I1121 14:10:14.235465 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"402271c9-35e5-4bed-8410-c84c2e132878\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08651f8381818947818e0bb640c089be4a07022efc650befbb65ad58466bf8c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://419ab9403696491ef53a3d812d4c19d676f8fb1a1a4dfcdfcc0c6255377d2918\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://acc9629bb6b0dfbe5b4efe2f9f980e4e3865fdc6742782ebc49a5a7e0f775b9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0fc71bebc463919826d01260f38b63bfdb01199fd8dd6ae90d9ac9cd75db703\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0fc71bebc463919826d01260f38b63bfdb01199fd8dd6ae90d9ac9cd75db703\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:14Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:14 crc kubenswrapper[4897]: I1121 14:10:14.257860 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:14Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:14 crc kubenswrapper[4897]: I1121 14:10:14.271109 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b78c993a-9966-4e84-85a1-b4f73aaf17f4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce3a2602338a1c1e9380ee55f78360f3d9cf6ccc62678886dc5ae574a72e3d9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c256823f2fb49ecc121f6ef564cd86196f9b4a473e9bdea670173596ef14a823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c65f71f8746044f5d41687c670a0308cd09dc41a6531f9b5720e852b3d240756\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1dbe88e7e6ad63899e107ebb884a16edb9a42b21ae7c86d7fbfaf89785b0d36\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:14Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:14 crc kubenswrapper[4897]: I1121 14:10:14.288730 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09877d9780caddd5f7f73e95fdf818a427ed645ed5997984113151f4b4759c13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:14Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:14 crc kubenswrapper[4897]: I1121 14:10:14.304208 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:14Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:14 crc kubenswrapper[4897]: I1121 14:10:14.317238 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:14 crc kubenswrapper[4897]: I1121 14:10:14.317315 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:14 crc kubenswrapper[4897]: I1121 14:10:14.317336 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:14 crc kubenswrapper[4897]: I1121 14:10:14.317368 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:14 crc kubenswrapper[4897]: I1121 14:10:14.317383 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:14Z","lastTransitionTime":"2025-11-21T14:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:14 crc kubenswrapper[4897]: I1121 14:10:14.320890 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6ljgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f3b70b5-9294-4f69-9528-500d28f34c89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:10:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:10:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf92ab70c5a3196bf83f427ccf03acf188d1995766a9154bcd6267e3926c3ac8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://527cf28abdeef1b42d9c19fb88e3171dc05c4516c3a542a461b5b1daa85c6d25\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-21T14:09:58Z\\\",\\\"message\\\":\\\"2025-11-21T14:09:13+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_a63cd511-2e05-4faf-b47b-8037110976d1\\\\n2025-11-21T14:09:13+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_a63cd511-2e05-4faf-b47b-8037110976d1 to /host/opt/cni/bin/\\\\n2025-11-21T14:09:13Z [verbose] multus-daemon started\\\\n2025-11-21T14:09:13Z [verbose] Readiness Indicator file check\\\\n2025-11-21T14:09:58Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78jhb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6ljgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:14Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:14 crc kubenswrapper[4897]: I1121 14:10:14.335859 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x5pws" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc9e717e-af03-48bf-b005-62b3d084c6fc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f968eee1d8a1888229e23abda59fd58fa5e86f02ca96ff80fbe093904b27e8df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9c4f5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0f99b9213902c003e6ca754134d2482404ace39c184fb3b431d2b600481d5b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9c4f5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-x5pws\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:14Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:14 crc kubenswrapper[4897]: I1121 14:10:14.355625 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6cd330fd-e8a6-49d5-bfc9-5f6122ccd49e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1bd64b9b938bc80284bcee54141e90cad613d339c4f27ef2934b29bc2a318916\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://936f59996d26e42099dcd055f87fff9e3ac4283b57e1956fe97463f2334152d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca74b28a069dc2e609b0779164c3aa470a15ed9848af0dd5fc3479aaf750157b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://270c53b9a342f26be077a01cd4e87e9ccd0eb5d2c84208ae38cbf720463245e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66c521c26d541bbddb064dd9710f1a7b41090d2f47343977922272dfc6dab7af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6eb2ac410b49098e5b3603de5e6bf5ed4484fab0b300021d0bed2e5ed8a42527\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6eb2ac410b49098e5b3603de5e6bf5ed4484fab0b300021d0bed2e5ed8a42527\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03385e0c78e4806f60c160227ef4905a11003ed3d70f0e5c0dd0a477979d1d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03385e0c78e4806f60c160227ef4905a11003ed3d70f0e5c0dd0a477979d1d4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://50aea2fa8b3b141b4330ddce2c8e0715230c6f06dab64b3abfd98a3067224bec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50aea2fa8b3b141b4330ddce2c8e0715230c6f06dab64b3abfd98a3067224bec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:14Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:14 crc kubenswrapper[4897]: I1121 14:10:14.367781 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1a6c0a99-565b-41f9-a8b5-c8990a0c8eb3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://737553dd1d16e2019ea89a376e3a868f9792ed5239aa7d6cc40a05d3d4f286ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6682a8886d2b9bce069cf9261d4701ef6907f19574fd035bd89bd05626023e5c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://de079956e3d8de0101f3f87dcc120ef23876c4972382639e419b3f34a38cdfca\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://469b5be7467d4a58d3776ec0a967ed911be31b770b49965da4f5d2fb264d8a3d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1907bfbad7eb34da642e30d6aa81820aab457f92347333a364ac12396b0e9dae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"le observer\\\\nW1121 14:09:07.994401 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1121 14:09:07.994601 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1121 14:09:07.995290 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1461162674/tls.crt::/tmp/serving-cert-1461162674/tls.key\\\\\\\"\\\\nI1121 14:09:08.162382 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1121 14:09:08.165301 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1121 14:09:08.165318 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1121 14:09:08.165343 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1121 14:09:08.165348 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1121 14:09:08.172015 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1121 14:09:08.172083 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1121 14:09:08.172089 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1121 14:09:08.172094 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1121 14:09:08.172098 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1121 14:09:08.172102 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1121 14:09:08.172106 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1121 14:09:08.172023 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1121 14:09:08.174214 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d78c87c440c7d8e30075f6143111b3f504fb99686484b8f19a9c8d61a55eccc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://101fb592fef069af009c613a1e9c9825024ed205498ab72334211ba90d6f7397\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://101fb592fef069af009c613a1e9c9825024ed205498ab72334211ba90d6f7397\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:14Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:14 crc kubenswrapper[4897]: I1121 14:10:14.382061 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-nl76g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a7d6101-d0d2-4828-b3ce-d01040c04640\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac58b6ef1bd08e5f950260c18120c72e50b90a0de182c6ab0a5dbbbe45cf7328\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e56402f62821baf335f378084f4ed7c21b60ccbafe148419a0a9a403f106ad19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e56402f62821baf335f378084f4ed7c21b60ccbafe148419a0a9a403f106ad19\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6022ff28e83dafcd846a7f03b090262aa3229a0da89f1984e4b6a054a04ac24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6022ff28e83dafcd846a7f03b090262aa3229a0da89f1984e4b6a054a04ac24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89de56e32b796da11bf89dd598d04fbf2884745ddf0f5d87cccfcaf1c001d8d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89de56e32b796da11bf89dd598d04fbf2884745ddf0f5d87cccfcaf1c001d8d6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2dff3cacaba7f9238f482e67b670ea775e0f96621434b99f4e76c2dfdde0321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2dff3cacaba7f9238f482e67b670ea775e0f96621434b99f4e76c2dfdde0321\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a153389a4137ee580e5118f3248400460f9d014c992a956f508815755f583afb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a153389a4137ee580e5118f3248400460f9d014c992a956f508815755f583afb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a8c051ad1c1f2373f79eb0781fa6db489aeea3d52ed6c0bb43eec11520ec2b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a8c051ad1c1f2373f79eb0781fa6db489aeea3d52ed6c0bb43eec11520ec2b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:09:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:09:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4592l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-nl76g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:14Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:14 crc kubenswrapper[4897]: I1121 14:10:14.393496 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:14Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:14 crc kubenswrapper[4897]: I1121 14:10:14.404591 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c976d804a7ebeba2c879e920cbc389f96f7f5def4b4439e8c8980d5460ecdb65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:14Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:14 crc kubenswrapper[4897]: I1121 14:10:14.420436 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:14 crc kubenswrapper[4897]: I1121 14:10:14.420477 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:14 crc kubenswrapper[4897]: I1121 14:10:14.420488 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:14 crc kubenswrapper[4897]: I1121 14:10:14.420526 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:14 crc kubenswrapper[4897]: I1121 14:10:14.420542 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:14Z","lastTransitionTime":"2025-11-21T14:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:14 crc kubenswrapper[4897]: I1121 14:10:14.522827 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:14 crc kubenswrapper[4897]: I1121 14:10:14.522882 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:14 crc kubenswrapper[4897]: I1121 14:10:14.522893 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:14 crc kubenswrapper[4897]: I1121 14:10:14.522910 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:14 crc kubenswrapper[4897]: I1121 14:10:14.522923 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:14Z","lastTransitionTime":"2025-11-21T14:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:14 crc kubenswrapper[4897]: I1121 14:10:14.625297 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:14 crc kubenswrapper[4897]: I1121 14:10:14.625357 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:14 crc kubenswrapper[4897]: I1121 14:10:14.625371 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:14 crc kubenswrapper[4897]: I1121 14:10:14.625394 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:14 crc kubenswrapper[4897]: I1121 14:10:14.625444 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:14Z","lastTransitionTime":"2025-11-21T14:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:14 crc kubenswrapper[4897]: I1121 14:10:14.728134 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:14 crc kubenswrapper[4897]: I1121 14:10:14.728179 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:14 crc kubenswrapper[4897]: I1121 14:10:14.728191 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:14 crc kubenswrapper[4897]: I1121 14:10:14.728207 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:14 crc kubenswrapper[4897]: I1121 14:10:14.728219 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:14Z","lastTransitionTime":"2025-11-21T14:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:14 crc kubenswrapper[4897]: I1121 14:10:14.831010 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:14 crc kubenswrapper[4897]: I1121 14:10:14.831068 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:14 crc kubenswrapper[4897]: I1121 14:10:14.831096 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:14 crc kubenswrapper[4897]: I1121 14:10:14.831140 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:14 crc kubenswrapper[4897]: I1121 14:10:14.831165 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:14Z","lastTransitionTime":"2025-11-21T14:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:14 crc kubenswrapper[4897]: I1121 14:10:14.933709 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:14 crc kubenswrapper[4897]: I1121 14:10:14.933758 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:14 crc kubenswrapper[4897]: I1121 14:10:14.933771 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:14 crc kubenswrapper[4897]: I1121 14:10:14.933788 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:14 crc kubenswrapper[4897]: I1121 14:10:14.933799 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:14Z","lastTransitionTime":"2025-11-21T14:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:15 crc kubenswrapper[4897]: I1121 14:10:15.036474 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:15 crc kubenswrapper[4897]: I1121 14:10:15.036543 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:15 crc kubenswrapper[4897]: I1121 14:10:15.036555 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:15 crc kubenswrapper[4897]: I1121 14:10:15.036572 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:15 crc kubenswrapper[4897]: I1121 14:10:15.036583 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:15Z","lastTransitionTime":"2025-11-21T14:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:15 crc kubenswrapper[4897]: I1121 14:10:15.088782 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 21 14:10:15 crc kubenswrapper[4897]: I1121 14:10:15.088845 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 21 14:10:15 crc kubenswrapper[4897]: E1121 14:10:15.088930 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 21 14:10:15 crc kubenswrapper[4897]: E1121 14:10:15.089062 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 21 14:10:15 crc kubenswrapper[4897]: I1121 14:10:15.088792 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs4c8" Nov 21 14:10:15 crc kubenswrapper[4897]: E1121 14:10:15.089353 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs4c8" podUID="01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd" Nov 21 14:10:15 crc kubenswrapper[4897]: I1121 14:10:15.088811 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 21 14:10:15 crc kubenswrapper[4897]: E1121 14:10:15.089948 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 21 14:10:15 crc kubenswrapper[4897]: I1121 14:10:15.138824 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:15 crc kubenswrapper[4897]: I1121 14:10:15.138882 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:15 crc kubenswrapper[4897]: I1121 14:10:15.138901 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:15 crc kubenswrapper[4897]: I1121 14:10:15.138927 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:15 crc kubenswrapper[4897]: I1121 14:10:15.138946 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:15Z","lastTransitionTime":"2025-11-21T14:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:15 crc kubenswrapper[4897]: I1121 14:10:15.241951 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:15 crc kubenswrapper[4897]: I1121 14:10:15.242007 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:15 crc kubenswrapper[4897]: I1121 14:10:15.242017 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:15 crc kubenswrapper[4897]: I1121 14:10:15.242035 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:15 crc kubenswrapper[4897]: I1121 14:10:15.242046 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:15Z","lastTransitionTime":"2025-11-21T14:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:15 crc kubenswrapper[4897]: I1121 14:10:15.344180 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:15 crc kubenswrapper[4897]: I1121 14:10:15.344219 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:15 crc kubenswrapper[4897]: I1121 14:10:15.344229 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:15 crc kubenswrapper[4897]: I1121 14:10:15.344246 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:15 crc kubenswrapper[4897]: I1121 14:10:15.344260 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:15Z","lastTransitionTime":"2025-11-21T14:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:15 crc kubenswrapper[4897]: I1121 14:10:15.447013 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:15 crc kubenswrapper[4897]: I1121 14:10:15.447068 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:15 crc kubenswrapper[4897]: I1121 14:10:15.447083 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:15 crc kubenswrapper[4897]: I1121 14:10:15.447101 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:15 crc kubenswrapper[4897]: I1121 14:10:15.447113 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:15Z","lastTransitionTime":"2025-11-21T14:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:15 crc kubenswrapper[4897]: I1121 14:10:15.549689 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:15 crc kubenswrapper[4897]: I1121 14:10:15.549725 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:15 crc kubenswrapper[4897]: I1121 14:10:15.549734 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:15 crc kubenswrapper[4897]: I1121 14:10:15.549747 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:15 crc kubenswrapper[4897]: I1121 14:10:15.549758 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:15Z","lastTransitionTime":"2025-11-21T14:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:15 crc kubenswrapper[4897]: I1121 14:10:15.652723 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:15 crc kubenswrapper[4897]: I1121 14:10:15.652772 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:15 crc kubenswrapper[4897]: I1121 14:10:15.652868 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:15 crc kubenswrapper[4897]: I1121 14:10:15.652909 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:15 crc kubenswrapper[4897]: I1121 14:10:15.652924 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:15Z","lastTransitionTime":"2025-11-21T14:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:15 crc kubenswrapper[4897]: I1121 14:10:15.755422 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:15 crc kubenswrapper[4897]: I1121 14:10:15.755486 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:15 crc kubenswrapper[4897]: I1121 14:10:15.755522 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:15 crc kubenswrapper[4897]: I1121 14:10:15.755543 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:15 crc kubenswrapper[4897]: I1121 14:10:15.755558 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:15Z","lastTransitionTime":"2025-11-21T14:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:15 crc kubenswrapper[4897]: I1121 14:10:15.858266 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:15 crc kubenswrapper[4897]: I1121 14:10:15.858295 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:15 crc kubenswrapper[4897]: I1121 14:10:15.858302 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:15 crc kubenswrapper[4897]: I1121 14:10:15.858315 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:15 crc kubenswrapper[4897]: I1121 14:10:15.858324 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:15Z","lastTransitionTime":"2025-11-21T14:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:15 crc kubenswrapper[4897]: I1121 14:10:15.960858 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:15 crc kubenswrapper[4897]: I1121 14:10:15.960905 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:15 crc kubenswrapper[4897]: I1121 14:10:15.960915 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:15 crc kubenswrapper[4897]: I1121 14:10:15.960929 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:15 crc kubenswrapper[4897]: I1121 14:10:15.960939 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:15Z","lastTransitionTime":"2025-11-21T14:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:16 crc kubenswrapper[4897]: I1121 14:10:16.063627 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:16 crc kubenswrapper[4897]: I1121 14:10:16.063690 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:16 crc kubenswrapper[4897]: I1121 14:10:16.063703 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:16 crc kubenswrapper[4897]: I1121 14:10:16.063720 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:16 crc kubenswrapper[4897]: I1121 14:10:16.063731 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:16Z","lastTransitionTime":"2025-11-21T14:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:16 crc kubenswrapper[4897]: I1121 14:10:16.166690 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:16 crc kubenswrapper[4897]: I1121 14:10:16.166738 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:16 crc kubenswrapper[4897]: I1121 14:10:16.166751 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:16 crc kubenswrapper[4897]: I1121 14:10:16.166767 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:16 crc kubenswrapper[4897]: I1121 14:10:16.166778 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:16Z","lastTransitionTime":"2025-11-21T14:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:16 crc kubenswrapper[4897]: I1121 14:10:16.269234 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:16 crc kubenswrapper[4897]: I1121 14:10:16.269306 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:16 crc kubenswrapper[4897]: I1121 14:10:16.269317 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:16 crc kubenswrapper[4897]: I1121 14:10:16.269342 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:16 crc kubenswrapper[4897]: I1121 14:10:16.269354 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:16Z","lastTransitionTime":"2025-11-21T14:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:16 crc kubenswrapper[4897]: I1121 14:10:16.372750 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:16 crc kubenswrapper[4897]: I1121 14:10:16.372800 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:16 crc kubenswrapper[4897]: I1121 14:10:16.372810 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:16 crc kubenswrapper[4897]: I1121 14:10:16.372828 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:16 crc kubenswrapper[4897]: I1121 14:10:16.372844 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:16Z","lastTransitionTime":"2025-11-21T14:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:16 crc kubenswrapper[4897]: I1121 14:10:16.475436 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:16 crc kubenswrapper[4897]: I1121 14:10:16.475567 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:16 crc kubenswrapper[4897]: I1121 14:10:16.475584 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:16 crc kubenswrapper[4897]: I1121 14:10:16.475605 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:16 crc kubenswrapper[4897]: I1121 14:10:16.475620 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:16Z","lastTransitionTime":"2025-11-21T14:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:16 crc kubenswrapper[4897]: I1121 14:10:16.578216 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:16 crc kubenswrapper[4897]: I1121 14:10:16.578253 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:16 crc kubenswrapper[4897]: I1121 14:10:16.578263 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:16 crc kubenswrapper[4897]: I1121 14:10:16.578277 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:16 crc kubenswrapper[4897]: I1121 14:10:16.578286 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:16Z","lastTransitionTime":"2025-11-21T14:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:16 crc kubenswrapper[4897]: I1121 14:10:16.681211 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:16 crc kubenswrapper[4897]: I1121 14:10:16.681258 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:16 crc kubenswrapper[4897]: I1121 14:10:16.681272 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:16 crc kubenswrapper[4897]: I1121 14:10:16.681288 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:16 crc kubenswrapper[4897]: I1121 14:10:16.681299 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:16Z","lastTransitionTime":"2025-11-21T14:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:16 crc kubenswrapper[4897]: I1121 14:10:16.782974 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:16 crc kubenswrapper[4897]: I1121 14:10:16.783030 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:16 crc kubenswrapper[4897]: I1121 14:10:16.783047 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:16 crc kubenswrapper[4897]: I1121 14:10:16.783070 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:16 crc kubenswrapper[4897]: I1121 14:10:16.783090 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:16Z","lastTransitionTime":"2025-11-21T14:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:16 crc kubenswrapper[4897]: I1121 14:10:16.886130 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:16 crc kubenswrapper[4897]: I1121 14:10:16.886187 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:16 crc kubenswrapper[4897]: I1121 14:10:16.886196 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:16 crc kubenswrapper[4897]: I1121 14:10:16.886210 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:16 crc kubenswrapper[4897]: I1121 14:10:16.886220 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:16Z","lastTransitionTime":"2025-11-21T14:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:16 crc kubenswrapper[4897]: I1121 14:10:16.989156 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:16 crc kubenswrapper[4897]: I1121 14:10:16.989226 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:16 crc kubenswrapper[4897]: I1121 14:10:16.989243 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:16 crc kubenswrapper[4897]: I1121 14:10:16.989264 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:16 crc kubenswrapper[4897]: I1121 14:10:16.989279 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:16Z","lastTransitionTime":"2025-11-21T14:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:17 crc kubenswrapper[4897]: I1121 14:10:17.062721 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:17 crc kubenswrapper[4897]: I1121 14:10:17.062768 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:17 crc kubenswrapper[4897]: I1121 14:10:17.062780 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:17 crc kubenswrapper[4897]: I1121 14:10:17.062798 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:17 crc kubenswrapper[4897]: I1121 14:10:17.062813 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:17Z","lastTransitionTime":"2025-11-21T14:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:17 crc kubenswrapper[4897]: E1121 14:10:17.075040 4897 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:10:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:10:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:10:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:10:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:10:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:10:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:10:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:10:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8b8bdf9e-0fef-41ac-98ad-710dc2137183\\\",\\\"systemUUID\\\":\\\"8a210af6-a71f-4ea9-90a3-aac32fb868b7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:17Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:17 crc kubenswrapper[4897]: I1121 14:10:17.078370 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:17 crc kubenswrapper[4897]: I1121 14:10:17.078427 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:17 crc kubenswrapper[4897]: I1121 14:10:17.078437 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:17 crc kubenswrapper[4897]: I1121 14:10:17.078454 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:17 crc kubenswrapper[4897]: I1121 14:10:17.078464 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:17Z","lastTransitionTime":"2025-11-21T14:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:17 crc kubenswrapper[4897]: I1121 14:10:17.087940 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 21 14:10:17 crc kubenswrapper[4897]: I1121 14:10:17.087980 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 21 14:10:17 crc kubenswrapper[4897]: I1121 14:10:17.087995 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 21 14:10:17 crc kubenswrapper[4897]: E1121 14:10:17.088063 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 21 14:10:17 crc kubenswrapper[4897]: I1121 14:10:17.087959 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs4c8" Nov 21 14:10:17 crc kubenswrapper[4897]: E1121 14:10:17.088164 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 21 14:10:17 crc kubenswrapper[4897]: E1121 14:10:17.088422 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs4c8" podUID="01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd" Nov 21 14:10:17 crc kubenswrapper[4897]: E1121 14:10:17.088563 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 21 14:10:17 crc kubenswrapper[4897]: E1121 14:10:17.089964 4897 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:10:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:10:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:10:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:10:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:10:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:10:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:10:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:10:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8b8bdf9e-0fef-41ac-98ad-710dc2137183\\\",\\\"systemUUID\\\":\\\"8a210af6-a71f-4ea9-90a3-aac32fb868b7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:17Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:17 crc kubenswrapper[4897]: I1121 14:10:17.093714 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:17 crc kubenswrapper[4897]: I1121 14:10:17.093753 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:17 crc kubenswrapper[4897]: I1121 14:10:17.093763 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:17 crc kubenswrapper[4897]: I1121 14:10:17.093778 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:17 crc kubenswrapper[4897]: I1121 14:10:17.093790 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:17Z","lastTransitionTime":"2025-11-21T14:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:17 crc kubenswrapper[4897]: I1121 14:10:17.101401 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Nov 21 14:10:17 crc kubenswrapper[4897]: E1121 14:10:17.105950 4897 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:10:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:10:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:10:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:10:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:10:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:10:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:10:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:10:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8b8bdf9e-0fef-41ac-98ad-710dc2137183\\\",\\\"systemUUID\\\":\\\"8a210af6-a71f-4ea9-90a3-aac32fb868b7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:17Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:17 crc kubenswrapper[4897]: I1121 14:10:17.109370 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:17 crc kubenswrapper[4897]: I1121 14:10:17.109406 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:17 crc kubenswrapper[4897]: I1121 14:10:17.109415 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:17 crc kubenswrapper[4897]: I1121 14:10:17.109432 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:17 crc kubenswrapper[4897]: I1121 14:10:17.109443 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:17Z","lastTransitionTime":"2025-11-21T14:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:17 crc kubenswrapper[4897]: E1121 14:10:17.122740 4897 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:10:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:10:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:10:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:10:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:10:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:10:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:10:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:10:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8b8bdf9e-0fef-41ac-98ad-710dc2137183\\\",\\\"systemUUID\\\":\\\"8a210af6-a71f-4ea9-90a3-aac32fb868b7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:17Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:17 crc kubenswrapper[4897]: I1121 14:10:17.125872 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:17 crc kubenswrapper[4897]: I1121 14:10:17.125909 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:17 crc kubenswrapper[4897]: I1121 14:10:17.125920 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:17 crc kubenswrapper[4897]: I1121 14:10:17.125935 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:17 crc kubenswrapper[4897]: I1121 14:10:17.125946 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:17Z","lastTransitionTime":"2025-11-21T14:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:17 crc kubenswrapper[4897]: E1121 14:10:17.138400 4897 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:10:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:10:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:10:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:10:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:10:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:10:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-21T14:10:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-21T14:10:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"8b8bdf9e-0fef-41ac-98ad-710dc2137183\\\",\\\"systemUUID\\\":\\\"8a210af6-a71f-4ea9-90a3-aac32fb868b7\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:17Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:17 crc kubenswrapper[4897]: E1121 14:10:17.138536 4897 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 21 14:10:17 crc kubenswrapper[4897]: I1121 14:10:17.140338 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:17 crc kubenswrapper[4897]: I1121 14:10:17.140458 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:17 crc kubenswrapper[4897]: I1121 14:10:17.140561 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:17 crc kubenswrapper[4897]: I1121 14:10:17.140583 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:17 crc kubenswrapper[4897]: I1121 14:10:17.140594 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:17Z","lastTransitionTime":"2025-11-21T14:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:17 crc kubenswrapper[4897]: I1121 14:10:17.243432 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:17 crc kubenswrapper[4897]: I1121 14:10:17.243462 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:17 crc kubenswrapper[4897]: I1121 14:10:17.243471 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:17 crc kubenswrapper[4897]: I1121 14:10:17.243485 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:17 crc kubenswrapper[4897]: I1121 14:10:17.243494 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:17Z","lastTransitionTime":"2025-11-21T14:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:17 crc kubenswrapper[4897]: I1121 14:10:17.346343 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:17 crc kubenswrapper[4897]: I1121 14:10:17.346385 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:17 crc kubenswrapper[4897]: I1121 14:10:17.346396 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:17 crc kubenswrapper[4897]: I1121 14:10:17.346412 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:17 crc kubenswrapper[4897]: I1121 14:10:17.346422 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:17Z","lastTransitionTime":"2025-11-21T14:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:17 crc kubenswrapper[4897]: I1121 14:10:17.449844 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:17 crc kubenswrapper[4897]: I1121 14:10:17.449899 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:17 crc kubenswrapper[4897]: I1121 14:10:17.449916 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:17 crc kubenswrapper[4897]: I1121 14:10:17.449937 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:17 crc kubenswrapper[4897]: I1121 14:10:17.449953 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:17Z","lastTransitionTime":"2025-11-21T14:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:17 crc kubenswrapper[4897]: I1121 14:10:17.552746 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:17 crc kubenswrapper[4897]: I1121 14:10:17.552777 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:17 crc kubenswrapper[4897]: I1121 14:10:17.552792 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:17 crc kubenswrapper[4897]: I1121 14:10:17.552816 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:17 crc kubenswrapper[4897]: I1121 14:10:17.552829 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:17Z","lastTransitionTime":"2025-11-21T14:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:17 crc kubenswrapper[4897]: I1121 14:10:17.655694 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:17 crc kubenswrapper[4897]: I1121 14:10:17.655748 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:17 crc kubenswrapper[4897]: I1121 14:10:17.655759 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:17 crc kubenswrapper[4897]: I1121 14:10:17.655776 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:17 crc kubenswrapper[4897]: I1121 14:10:17.655790 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:17Z","lastTransitionTime":"2025-11-21T14:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:17 crc kubenswrapper[4897]: I1121 14:10:17.758617 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:17 crc kubenswrapper[4897]: I1121 14:10:17.758692 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:17 crc kubenswrapper[4897]: I1121 14:10:17.758704 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:17 crc kubenswrapper[4897]: I1121 14:10:17.758724 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:17 crc kubenswrapper[4897]: I1121 14:10:17.758738 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:17Z","lastTransitionTime":"2025-11-21T14:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:17 crc kubenswrapper[4897]: I1121 14:10:17.861176 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:17 crc kubenswrapper[4897]: I1121 14:10:17.861220 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:17 crc kubenswrapper[4897]: I1121 14:10:17.861228 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:17 crc kubenswrapper[4897]: I1121 14:10:17.861242 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:17 crc kubenswrapper[4897]: I1121 14:10:17.861251 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:17Z","lastTransitionTime":"2025-11-21T14:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:17 crc kubenswrapper[4897]: I1121 14:10:17.963587 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:17 crc kubenswrapper[4897]: I1121 14:10:17.963665 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:17 crc kubenswrapper[4897]: I1121 14:10:17.963693 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:17 crc kubenswrapper[4897]: I1121 14:10:17.963724 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:17 crc kubenswrapper[4897]: I1121 14:10:17.963746 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:17Z","lastTransitionTime":"2025-11-21T14:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:18 crc kubenswrapper[4897]: I1121 14:10:18.067097 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:18 crc kubenswrapper[4897]: I1121 14:10:18.067142 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:18 crc kubenswrapper[4897]: I1121 14:10:18.067153 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:18 crc kubenswrapper[4897]: I1121 14:10:18.067171 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:18 crc kubenswrapper[4897]: I1121 14:10:18.067215 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:18Z","lastTransitionTime":"2025-11-21T14:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:18 crc kubenswrapper[4897]: I1121 14:10:18.170327 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:18 crc kubenswrapper[4897]: I1121 14:10:18.170385 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:18 crc kubenswrapper[4897]: I1121 14:10:18.170402 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:18 crc kubenswrapper[4897]: I1121 14:10:18.170424 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:18 crc kubenswrapper[4897]: I1121 14:10:18.170445 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:18Z","lastTransitionTime":"2025-11-21T14:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:18 crc kubenswrapper[4897]: I1121 14:10:18.273950 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:18 crc kubenswrapper[4897]: I1121 14:10:18.274026 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:18 crc kubenswrapper[4897]: I1121 14:10:18.274045 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:18 crc kubenswrapper[4897]: I1121 14:10:18.274068 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:18 crc kubenswrapper[4897]: I1121 14:10:18.274085 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:18Z","lastTransitionTime":"2025-11-21T14:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:18 crc kubenswrapper[4897]: I1121 14:10:18.378168 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:18 crc kubenswrapper[4897]: I1121 14:10:18.378235 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:18 crc kubenswrapper[4897]: I1121 14:10:18.378255 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:18 crc kubenswrapper[4897]: I1121 14:10:18.378280 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:18 crc kubenswrapper[4897]: I1121 14:10:18.378301 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:18Z","lastTransitionTime":"2025-11-21T14:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:18 crc kubenswrapper[4897]: I1121 14:10:18.482387 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:18 crc kubenswrapper[4897]: I1121 14:10:18.482445 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:18 crc kubenswrapper[4897]: I1121 14:10:18.482458 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:18 crc kubenswrapper[4897]: I1121 14:10:18.482478 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:18 crc kubenswrapper[4897]: I1121 14:10:18.482492 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:18Z","lastTransitionTime":"2025-11-21T14:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:18 crc kubenswrapper[4897]: I1121 14:10:18.585573 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:18 crc kubenswrapper[4897]: I1121 14:10:18.585639 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:18 crc kubenswrapper[4897]: I1121 14:10:18.585658 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:18 crc kubenswrapper[4897]: I1121 14:10:18.585684 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:18 crc kubenswrapper[4897]: I1121 14:10:18.585704 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:18Z","lastTransitionTime":"2025-11-21T14:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:18 crc kubenswrapper[4897]: I1121 14:10:18.688415 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:18 crc kubenswrapper[4897]: I1121 14:10:18.688479 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:18 crc kubenswrapper[4897]: I1121 14:10:18.688492 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:18 crc kubenswrapper[4897]: I1121 14:10:18.688551 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:18 crc kubenswrapper[4897]: I1121 14:10:18.688567 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:18Z","lastTransitionTime":"2025-11-21T14:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:18 crc kubenswrapper[4897]: I1121 14:10:18.791139 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:18 crc kubenswrapper[4897]: I1121 14:10:18.791195 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:18 crc kubenswrapper[4897]: I1121 14:10:18.791209 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:18 crc kubenswrapper[4897]: I1121 14:10:18.791231 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:18 crc kubenswrapper[4897]: I1121 14:10:18.791251 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:18Z","lastTransitionTime":"2025-11-21T14:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:18 crc kubenswrapper[4897]: I1121 14:10:18.894458 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:18 crc kubenswrapper[4897]: I1121 14:10:18.894523 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:18 crc kubenswrapper[4897]: I1121 14:10:18.894540 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:18 crc kubenswrapper[4897]: I1121 14:10:18.894561 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:18 crc kubenswrapper[4897]: I1121 14:10:18.894574 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:18Z","lastTransitionTime":"2025-11-21T14:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:18 crc kubenswrapper[4897]: I1121 14:10:18.997267 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:18 crc kubenswrapper[4897]: I1121 14:10:18.997313 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:18 crc kubenswrapper[4897]: I1121 14:10:18.997324 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:18 crc kubenswrapper[4897]: I1121 14:10:18.997342 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:18 crc kubenswrapper[4897]: I1121 14:10:18.997353 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:18Z","lastTransitionTime":"2025-11-21T14:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:19 crc kubenswrapper[4897]: I1121 14:10:19.088259 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs4c8" Nov 21 14:10:19 crc kubenswrapper[4897]: I1121 14:10:19.088309 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 21 14:10:19 crc kubenswrapper[4897]: I1121 14:10:19.088339 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 21 14:10:19 crc kubenswrapper[4897]: I1121 14:10:19.088317 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 21 14:10:19 crc kubenswrapper[4897]: E1121 14:10:19.088436 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs4c8" podUID="01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd" Nov 21 14:10:19 crc kubenswrapper[4897]: E1121 14:10:19.088546 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 21 14:10:19 crc kubenswrapper[4897]: E1121 14:10:19.088653 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 21 14:10:19 crc kubenswrapper[4897]: E1121 14:10:19.088762 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 21 14:10:19 crc kubenswrapper[4897]: I1121 14:10:19.099735 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:19 crc kubenswrapper[4897]: I1121 14:10:19.099764 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:19 crc kubenswrapper[4897]: I1121 14:10:19.099775 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:19 crc kubenswrapper[4897]: I1121 14:10:19.099788 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:19 crc kubenswrapper[4897]: I1121 14:10:19.099797 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:19Z","lastTransitionTime":"2025-11-21T14:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:19 crc kubenswrapper[4897]: I1121 14:10:19.203272 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:19 crc kubenswrapper[4897]: I1121 14:10:19.203323 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:19 crc kubenswrapper[4897]: I1121 14:10:19.203333 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:19 crc kubenswrapper[4897]: I1121 14:10:19.203353 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:19 crc kubenswrapper[4897]: I1121 14:10:19.203365 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:19Z","lastTransitionTime":"2025-11-21T14:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:19 crc kubenswrapper[4897]: I1121 14:10:19.305731 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:19 crc kubenswrapper[4897]: I1121 14:10:19.305765 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:19 crc kubenswrapper[4897]: I1121 14:10:19.305774 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:19 crc kubenswrapper[4897]: I1121 14:10:19.305788 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:19 crc kubenswrapper[4897]: I1121 14:10:19.305798 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:19Z","lastTransitionTime":"2025-11-21T14:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:19 crc kubenswrapper[4897]: I1121 14:10:19.408025 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:19 crc kubenswrapper[4897]: I1121 14:10:19.408081 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:19 crc kubenswrapper[4897]: I1121 14:10:19.408099 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:19 crc kubenswrapper[4897]: I1121 14:10:19.408120 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:19 crc kubenswrapper[4897]: I1121 14:10:19.408132 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:19Z","lastTransitionTime":"2025-11-21T14:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:19 crc kubenswrapper[4897]: I1121 14:10:19.511952 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:19 crc kubenswrapper[4897]: I1121 14:10:19.512002 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:19 crc kubenswrapper[4897]: I1121 14:10:19.512016 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:19 crc kubenswrapper[4897]: I1121 14:10:19.512036 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:19 crc kubenswrapper[4897]: I1121 14:10:19.512051 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:19Z","lastTransitionTime":"2025-11-21T14:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:19 crc kubenswrapper[4897]: I1121 14:10:19.615785 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:19 crc kubenswrapper[4897]: I1121 14:10:19.615842 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:19 crc kubenswrapper[4897]: I1121 14:10:19.615853 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:19 crc kubenswrapper[4897]: I1121 14:10:19.615871 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:19 crc kubenswrapper[4897]: I1121 14:10:19.615885 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:19Z","lastTransitionTime":"2025-11-21T14:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:19 crc kubenswrapper[4897]: I1121 14:10:19.719200 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:19 crc kubenswrapper[4897]: I1121 14:10:19.719253 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:19 crc kubenswrapper[4897]: I1121 14:10:19.719279 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:19 crc kubenswrapper[4897]: I1121 14:10:19.719304 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:19 crc kubenswrapper[4897]: I1121 14:10:19.719321 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:19Z","lastTransitionTime":"2025-11-21T14:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:19 crc kubenswrapper[4897]: I1121 14:10:19.822087 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:19 crc kubenswrapper[4897]: I1121 14:10:19.822122 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:19 crc kubenswrapper[4897]: I1121 14:10:19.822131 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:19 crc kubenswrapper[4897]: I1121 14:10:19.822148 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:19 crc kubenswrapper[4897]: I1121 14:10:19.822157 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:19Z","lastTransitionTime":"2025-11-21T14:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:19 crc kubenswrapper[4897]: I1121 14:10:19.924580 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:19 crc kubenswrapper[4897]: I1121 14:10:19.924625 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:19 crc kubenswrapper[4897]: I1121 14:10:19.924641 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:19 crc kubenswrapper[4897]: I1121 14:10:19.924662 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:19 crc kubenswrapper[4897]: I1121 14:10:19.924676 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:19Z","lastTransitionTime":"2025-11-21T14:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:20 crc kubenswrapper[4897]: I1121 14:10:20.026636 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:20 crc kubenswrapper[4897]: I1121 14:10:20.026680 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:20 crc kubenswrapper[4897]: I1121 14:10:20.026690 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:20 crc kubenswrapper[4897]: I1121 14:10:20.026705 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:20 crc kubenswrapper[4897]: I1121 14:10:20.026715 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:20Z","lastTransitionTime":"2025-11-21T14:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:20 crc kubenswrapper[4897]: I1121 14:10:20.129363 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:20 crc kubenswrapper[4897]: I1121 14:10:20.129425 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:20 crc kubenswrapper[4897]: I1121 14:10:20.129435 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:20 crc kubenswrapper[4897]: I1121 14:10:20.129451 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:20 crc kubenswrapper[4897]: I1121 14:10:20.129461 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:20Z","lastTransitionTime":"2025-11-21T14:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:20 crc kubenswrapper[4897]: I1121 14:10:20.231994 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:20 crc kubenswrapper[4897]: I1121 14:10:20.232033 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:20 crc kubenswrapper[4897]: I1121 14:10:20.232046 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:20 crc kubenswrapper[4897]: I1121 14:10:20.232063 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:20 crc kubenswrapper[4897]: I1121 14:10:20.232075 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:20Z","lastTransitionTime":"2025-11-21T14:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:20 crc kubenswrapper[4897]: I1121 14:10:20.335149 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:20 crc kubenswrapper[4897]: I1121 14:10:20.335187 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:20 crc kubenswrapper[4897]: I1121 14:10:20.335197 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:20 crc kubenswrapper[4897]: I1121 14:10:20.335268 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:20 crc kubenswrapper[4897]: I1121 14:10:20.335283 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:20Z","lastTransitionTime":"2025-11-21T14:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:20 crc kubenswrapper[4897]: I1121 14:10:20.437740 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:20 crc kubenswrapper[4897]: I1121 14:10:20.437810 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:20 crc kubenswrapper[4897]: I1121 14:10:20.437827 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:20 crc kubenswrapper[4897]: I1121 14:10:20.437856 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:20 crc kubenswrapper[4897]: I1121 14:10:20.437877 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:20Z","lastTransitionTime":"2025-11-21T14:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:20 crc kubenswrapper[4897]: I1121 14:10:20.541448 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:20 crc kubenswrapper[4897]: I1121 14:10:20.541522 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:20 crc kubenswrapper[4897]: I1121 14:10:20.541542 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:20 crc kubenswrapper[4897]: I1121 14:10:20.541559 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:20 crc kubenswrapper[4897]: I1121 14:10:20.541571 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:20Z","lastTransitionTime":"2025-11-21T14:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:20 crc kubenswrapper[4897]: I1121 14:10:20.644116 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:20 crc kubenswrapper[4897]: I1121 14:10:20.644170 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:20 crc kubenswrapper[4897]: I1121 14:10:20.644187 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:20 crc kubenswrapper[4897]: I1121 14:10:20.644209 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:20 crc kubenswrapper[4897]: I1121 14:10:20.644222 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:20Z","lastTransitionTime":"2025-11-21T14:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:20 crc kubenswrapper[4897]: I1121 14:10:20.748247 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:20 crc kubenswrapper[4897]: I1121 14:10:20.748312 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:20 crc kubenswrapper[4897]: I1121 14:10:20.748328 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:20 crc kubenswrapper[4897]: I1121 14:10:20.748349 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:20 crc kubenswrapper[4897]: I1121 14:10:20.748363 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:20Z","lastTransitionTime":"2025-11-21T14:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:20 crc kubenswrapper[4897]: I1121 14:10:20.851582 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:20 crc kubenswrapper[4897]: I1121 14:10:20.851658 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:20 crc kubenswrapper[4897]: I1121 14:10:20.851676 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:20 crc kubenswrapper[4897]: I1121 14:10:20.851705 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:20 crc kubenswrapper[4897]: I1121 14:10:20.851724 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:20Z","lastTransitionTime":"2025-11-21T14:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:20 crc kubenswrapper[4897]: I1121 14:10:20.954673 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:20 crc kubenswrapper[4897]: I1121 14:10:20.954765 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:20 crc kubenswrapper[4897]: I1121 14:10:20.954788 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:20 crc kubenswrapper[4897]: I1121 14:10:20.954818 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:20 crc kubenswrapper[4897]: I1121 14:10:20.954844 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:20Z","lastTransitionTime":"2025-11-21T14:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:21 crc kubenswrapper[4897]: I1121 14:10:21.057920 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:21 crc kubenswrapper[4897]: I1121 14:10:21.057997 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:21 crc kubenswrapper[4897]: I1121 14:10:21.058016 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:21 crc kubenswrapper[4897]: I1121 14:10:21.058043 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:21 crc kubenswrapper[4897]: I1121 14:10:21.058061 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:21Z","lastTransitionTime":"2025-11-21T14:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:21 crc kubenswrapper[4897]: I1121 14:10:21.088315 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 21 14:10:21 crc kubenswrapper[4897]: I1121 14:10:21.088387 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs4c8" Nov 21 14:10:21 crc kubenswrapper[4897]: I1121 14:10:21.088411 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 21 14:10:21 crc kubenswrapper[4897]: I1121 14:10:21.088555 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 21 14:10:21 crc kubenswrapper[4897]: E1121 14:10:21.088564 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 21 14:10:21 crc kubenswrapper[4897]: E1121 14:10:21.088922 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 21 14:10:21 crc kubenswrapper[4897]: E1121 14:10:21.089003 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs4c8" podUID="01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd" Nov 21 14:10:21 crc kubenswrapper[4897]: E1121 14:10:21.089123 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 21 14:10:21 crc kubenswrapper[4897]: I1121 14:10:21.089388 4897 scope.go:117] "RemoveContainer" containerID="40b0bddce48b3626ad889ae2f43deb866acd521a8aab71253506364bfb47c723" Nov 21 14:10:21 crc kubenswrapper[4897]: E1121 14:10:21.089770 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-tpfb6_openshift-ovn-kubernetes(222f65f2-0b82-4760-bcf7-779244720f01)\"" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" podUID="222f65f2-0b82-4760-bcf7-779244720f01" Nov 21 14:10:21 crc kubenswrapper[4897]: I1121 14:10:21.160737 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:21 crc kubenswrapper[4897]: I1121 14:10:21.160780 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:21 crc kubenswrapper[4897]: I1121 14:10:21.160790 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:21 crc kubenswrapper[4897]: I1121 14:10:21.160806 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:21 crc kubenswrapper[4897]: I1121 14:10:21.160816 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:21Z","lastTransitionTime":"2025-11-21T14:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:21 crc kubenswrapper[4897]: I1121 14:10:21.263314 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:21 crc kubenswrapper[4897]: I1121 14:10:21.263376 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:21 crc kubenswrapper[4897]: I1121 14:10:21.263389 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:21 crc kubenswrapper[4897]: I1121 14:10:21.263409 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:21 crc kubenswrapper[4897]: I1121 14:10:21.263425 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:21Z","lastTransitionTime":"2025-11-21T14:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:21 crc kubenswrapper[4897]: I1121 14:10:21.366236 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:21 crc kubenswrapper[4897]: I1121 14:10:21.366294 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:21 crc kubenswrapper[4897]: I1121 14:10:21.366310 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:21 crc kubenswrapper[4897]: I1121 14:10:21.366332 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:21 crc kubenswrapper[4897]: I1121 14:10:21.366347 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:21Z","lastTransitionTime":"2025-11-21T14:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:21 crc kubenswrapper[4897]: I1121 14:10:21.470184 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:21 crc kubenswrapper[4897]: I1121 14:10:21.470229 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:21 crc kubenswrapper[4897]: I1121 14:10:21.470239 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:21 crc kubenswrapper[4897]: I1121 14:10:21.470254 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:21 crc kubenswrapper[4897]: I1121 14:10:21.470265 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:21Z","lastTransitionTime":"2025-11-21T14:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:21 crc kubenswrapper[4897]: I1121 14:10:21.573029 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:21 crc kubenswrapper[4897]: I1121 14:10:21.573065 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:21 crc kubenswrapper[4897]: I1121 14:10:21.573073 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:21 crc kubenswrapper[4897]: I1121 14:10:21.573086 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:21 crc kubenswrapper[4897]: I1121 14:10:21.573095 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:21Z","lastTransitionTime":"2025-11-21T14:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:21 crc kubenswrapper[4897]: I1121 14:10:21.675483 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:21 crc kubenswrapper[4897]: I1121 14:10:21.675677 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:21 crc kubenswrapper[4897]: I1121 14:10:21.675694 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:21 crc kubenswrapper[4897]: I1121 14:10:21.675713 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:21 crc kubenswrapper[4897]: I1121 14:10:21.675725 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:21Z","lastTransitionTime":"2025-11-21T14:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:21 crc kubenswrapper[4897]: I1121 14:10:21.777840 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:21 crc kubenswrapper[4897]: I1121 14:10:21.777870 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:21 crc kubenswrapper[4897]: I1121 14:10:21.777879 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:21 crc kubenswrapper[4897]: I1121 14:10:21.777892 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:21 crc kubenswrapper[4897]: I1121 14:10:21.777902 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:21Z","lastTransitionTime":"2025-11-21T14:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:21 crc kubenswrapper[4897]: I1121 14:10:21.880414 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:21 crc kubenswrapper[4897]: I1121 14:10:21.880486 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:21 crc kubenswrapper[4897]: I1121 14:10:21.880497 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:21 crc kubenswrapper[4897]: I1121 14:10:21.880525 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:21 crc kubenswrapper[4897]: I1121 14:10:21.880535 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:21Z","lastTransitionTime":"2025-11-21T14:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:21 crc kubenswrapper[4897]: I1121 14:10:21.983393 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:21 crc kubenswrapper[4897]: I1121 14:10:21.983434 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:21 crc kubenswrapper[4897]: I1121 14:10:21.983443 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:21 crc kubenswrapper[4897]: I1121 14:10:21.983461 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:21 crc kubenswrapper[4897]: I1121 14:10:21.983471 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:21Z","lastTransitionTime":"2025-11-21T14:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:22 crc kubenswrapper[4897]: I1121 14:10:22.086141 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:22 crc kubenswrapper[4897]: I1121 14:10:22.086189 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:22 crc kubenswrapper[4897]: I1121 14:10:22.086201 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:22 crc kubenswrapper[4897]: I1121 14:10:22.086222 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:22 crc kubenswrapper[4897]: I1121 14:10:22.086239 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:22Z","lastTransitionTime":"2025-11-21T14:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:22 crc kubenswrapper[4897]: I1121 14:10:22.189126 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:22 crc kubenswrapper[4897]: I1121 14:10:22.189240 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:22 crc kubenswrapper[4897]: I1121 14:10:22.189258 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:22 crc kubenswrapper[4897]: I1121 14:10:22.189277 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:22 crc kubenswrapper[4897]: I1121 14:10:22.189289 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:22Z","lastTransitionTime":"2025-11-21T14:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:22 crc kubenswrapper[4897]: I1121 14:10:22.292243 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:22 crc kubenswrapper[4897]: I1121 14:10:22.292322 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:22 crc kubenswrapper[4897]: I1121 14:10:22.292343 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:22 crc kubenswrapper[4897]: I1121 14:10:22.292369 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:22 crc kubenswrapper[4897]: I1121 14:10:22.292387 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:22Z","lastTransitionTime":"2025-11-21T14:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:22 crc kubenswrapper[4897]: I1121 14:10:22.395492 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:22 crc kubenswrapper[4897]: I1121 14:10:22.395626 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:22 crc kubenswrapper[4897]: I1121 14:10:22.395654 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:22 crc kubenswrapper[4897]: I1121 14:10:22.395688 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:22 crc kubenswrapper[4897]: I1121 14:10:22.395709 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:22Z","lastTransitionTime":"2025-11-21T14:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:22 crc kubenswrapper[4897]: I1121 14:10:22.499301 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:22 crc kubenswrapper[4897]: I1121 14:10:22.499384 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:22 crc kubenswrapper[4897]: I1121 14:10:22.499406 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:22 crc kubenswrapper[4897]: I1121 14:10:22.499433 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:22 crc kubenswrapper[4897]: I1121 14:10:22.499452 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:22Z","lastTransitionTime":"2025-11-21T14:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:22 crc kubenswrapper[4897]: I1121 14:10:22.602216 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:22 crc kubenswrapper[4897]: I1121 14:10:22.602305 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:22 crc kubenswrapper[4897]: I1121 14:10:22.602317 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:22 crc kubenswrapper[4897]: I1121 14:10:22.602333 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:22 crc kubenswrapper[4897]: I1121 14:10:22.602342 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:22Z","lastTransitionTime":"2025-11-21T14:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:22 crc kubenswrapper[4897]: I1121 14:10:22.704874 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:22 crc kubenswrapper[4897]: I1121 14:10:22.705191 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:22 crc kubenswrapper[4897]: I1121 14:10:22.705203 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:22 crc kubenswrapper[4897]: I1121 14:10:22.705218 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:22 crc kubenswrapper[4897]: I1121 14:10:22.705228 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:22Z","lastTransitionTime":"2025-11-21T14:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:22 crc kubenswrapper[4897]: I1121 14:10:22.809396 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:22 crc kubenswrapper[4897]: I1121 14:10:22.809460 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:22 crc kubenswrapper[4897]: I1121 14:10:22.809482 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:22 crc kubenswrapper[4897]: I1121 14:10:22.809546 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:22 crc kubenswrapper[4897]: I1121 14:10:22.809570 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:22Z","lastTransitionTime":"2025-11-21T14:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:22 crc kubenswrapper[4897]: I1121 14:10:22.913440 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:22 crc kubenswrapper[4897]: I1121 14:10:22.913492 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:22 crc kubenswrapper[4897]: I1121 14:10:22.913547 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:22 crc kubenswrapper[4897]: I1121 14:10:22.913573 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:22 crc kubenswrapper[4897]: I1121 14:10:22.913592 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:22Z","lastTransitionTime":"2025-11-21T14:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:23 crc kubenswrapper[4897]: I1121 14:10:23.015824 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:23 crc kubenswrapper[4897]: I1121 14:10:23.015866 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:23 crc kubenswrapper[4897]: I1121 14:10:23.015877 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:23 crc kubenswrapper[4897]: I1121 14:10:23.015911 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:23 crc kubenswrapper[4897]: I1121 14:10:23.015922 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:23Z","lastTransitionTime":"2025-11-21T14:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:23 crc kubenswrapper[4897]: I1121 14:10:23.088581 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 21 14:10:23 crc kubenswrapper[4897]: I1121 14:10:23.088583 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 21 14:10:23 crc kubenswrapper[4897]: I1121 14:10:23.088717 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs4c8" Nov 21 14:10:23 crc kubenswrapper[4897]: E1121 14:10:23.088816 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 21 14:10:23 crc kubenswrapper[4897]: I1121 14:10:23.088910 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 21 14:10:23 crc kubenswrapper[4897]: E1121 14:10:23.089129 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs4c8" podUID="01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd" Nov 21 14:10:23 crc kubenswrapper[4897]: E1121 14:10:23.089166 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 21 14:10:23 crc kubenswrapper[4897]: E1121 14:10:23.089256 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 21 14:10:23 crc kubenswrapper[4897]: I1121 14:10:23.118265 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:23 crc kubenswrapper[4897]: I1121 14:10:23.118322 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:23 crc kubenswrapper[4897]: I1121 14:10:23.118337 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:23 crc kubenswrapper[4897]: I1121 14:10:23.118358 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:23 crc kubenswrapper[4897]: I1121 14:10:23.118377 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:23Z","lastTransitionTime":"2025-11-21T14:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:23 crc kubenswrapper[4897]: I1121 14:10:23.221998 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:23 crc kubenswrapper[4897]: I1121 14:10:23.222060 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:23 crc kubenswrapper[4897]: I1121 14:10:23.222084 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:23 crc kubenswrapper[4897]: I1121 14:10:23.222111 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:23 crc kubenswrapper[4897]: I1121 14:10:23.222130 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:23Z","lastTransitionTime":"2025-11-21T14:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:23 crc kubenswrapper[4897]: I1121 14:10:23.325298 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:23 crc kubenswrapper[4897]: I1121 14:10:23.325373 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:23 crc kubenswrapper[4897]: I1121 14:10:23.325402 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:23 crc kubenswrapper[4897]: I1121 14:10:23.325426 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:23 crc kubenswrapper[4897]: I1121 14:10:23.325444 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:23Z","lastTransitionTime":"2025-11-21T14:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:23 crc kubenswrapper[4897]: I1121 14:10:23.428054 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:23 crc kubenswrapper[4897]: I1121 14:10:23.428139 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:23 crc kubenswrapper[4897]: I1121 14:10:23.428150 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:23 crc kubenswrapper[4897]: I1121 14:10:23.428166 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:23 crc kubenswrapper[4897]: I1121 14:10:23.428176 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:23Z","lastTransitionTime":"2025-11-21T14:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:23 crc kubenswrapper[4897]: I1121 14:10:23.530340 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:23 crc kubenswrapper[4897]: I1121 14:10:23.530393 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:23 crc kubenswrapper[4897]: I1121 14:10:23.530407 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:23 crc kubenswrapper[4897]: I1121 14:10:23.530428 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:23 crc kubenswrapper[4897]: I1121 14:10:23.530441 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:23Z","lastTransitionTime":"2025-11-21T14:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:23 crc kubenswrapper[4897]: I1121 14:10:23.633152 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:23 crc kubenswrapper[4897]: I1121 14:10:23.633232 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:23 crc kubenswrapper[4897]: I1121 14:10:23.633255 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:23 crc kubenswrapper[4897]: I1121 14:10:23.633285 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:23 crc kubenswrapper[4897]: I1121 14:10:23.633323 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:23Z","lastTransitionTime":"2025-11-21T14:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:23 crc kubenswrapper[4897]: I1121 14:10:23.736635 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:23 crc kubenswrapper[4897]: I1121 14:10:23.736703 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:23 crc kubenswrapper[4897]: I1121 14:10:23.736722 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:23 crc kubenswrapper[4897]: I1121 14:10:23.736750 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:23 crc kubenswrapper[4897]: I1121 14:10:23.736770 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:23Z","lastTransitionTime":"2025-11-21T14:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:23 crc kubenswrapper[4897]: I1121 14:10:23.839614 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:23 crc kubenswrapper[4897]: I1121 14:10:23.839696 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:23 crc kubenswrapper[4897]: I1121 14:10:23.839715 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:23 crc kubenswrapper[4897]: I1121 14:10:23.839743 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:23 crc kubenswrapper[4897]: I1121 14:10:23.839763 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:23Z","lastTransitionTime":"2025-11-21T14:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:23 crc kubenswrapper[4897]: I1121 14:10:23.943556 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:23 crc kubenswrapper[4897]: I1121 14:10:23.943639 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:23 crc kubenswrapper[4897]: I1121 14:10:23.943658 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:23 crc kubenswrapper[4897]: I1121 14:10:23.943683 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:23 crc kubenswrapper[4897]: I1121 14:10:23.943701 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:23Z","lastTransitionTime":"2025-11-21T14:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:24 crc kubenswrapper[4897]: I1121 14:10:24.047145 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:24 crc kubenswrapper[4897]: I1121 14:10:24.047192 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:24 crc kubenswrapper[4897]: I1121 14:10:24.047201 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:24 crc kubenswrapper[4897]: I1121 14:10:24.047215 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:24 crc kubenswrapper[4897]: I1121 14:10:24.047226 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:24Z","lastTransitionTime":"2025-11-21T14:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:24 crc kubenswrapper[4897]: I1121 14:10:24.102297 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4054ce14-987b-4b9c-a686-9209f961d052\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1091bce69830e7805c81aea0fcabd87cb16fd9feef5becdd44d9d2baba613437\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ca61ced18b5737bdd0f0f711d97c0b06ab34d5fc2412db2f11a3375e0baf5740\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ca61ced18b5737bdd0f0f711d97c0b06ab34d5fc2412db2f11a3375e0baf5740\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:24Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:24 crc kubenswrapper[4897]: I1121 14:10:24.120739 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48c6b1ea7497b33c00b956e400426c1d4f3c6c66e0225133434460800eef9314\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ea6068a6b6f8123b402de354e5374a0f02a7191a922d1718ab38f8898f47c52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:24Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:24 crc kubenswrapper[4897]: I1121 14:10:24.132812 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4qwqj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f46e2ba8-9ee0-406d-a668-2ba424618b69\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f43aa89d4d9defcb50d0dc6f90293b5f426b9cfc398b780c2e10f6a5c9f1e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbsgb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4qwqj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:24Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:24 crc kubenswrapper[4897]: I1121 14:10:24.144135 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-rs4c8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7xkdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7xkdf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:23Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-rs4c8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:24Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:24 crc kubenswrapper[4897]: I1121 14:10:24.149861 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:24 crc kubenswrapper[4897]: I1121 14:10:24.149898 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:24 crc kubenswrapper[4897]: I1121 14:10:24.149912 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:24 crc kubenswrapper[4897]: I1121 14:10:24.149931 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:24 crc kubenswrapper[4897]: I1121 14:10:24.149944 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:24Z","lastTransitionTime":"2025-11-21T14:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:24 crc kubenswrapper[4897]: I1121 14:10:24.159002 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"402271c9-35e5-4bed-8410-c84c2e132878\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08651f8381818947818e0bb640c089be4a07022efc650befbb65ad58466bf8c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://419ab9403696491ef53a3d812d4c19d676f8fb1a1a4dfcdfcc0c6255377d2918\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://acc9629bb6b0dfbe5b4efe2f9f980e4e3865fdc6742782ebc49a5a7e0f775b9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:08:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e0fc71bebc463919826d01260f38b63bfdb01199fd8dd6ae90d9ac9cd75db703\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e0fc71bebc463919826d01260f38b63bfdb01199fd8dd6ae90d9ac9cd75db703\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-21T14:08:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-21T14:08:46Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:08:44Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:24Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:24 crc kubenswrapper[4897]: I1121 14:10:24.178652 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:24Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:24 crc kubenswrapper[4897]: I1121 14:10:24.194977 4897 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7670227-d280-4847-b882-754429f56b0f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-21T14:09:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://961961696b8750c248598f4cda032b18395e5c802275dcfe6d3e9f7ef3510f1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5854028821e6a360a259427b68e7bb4c2a7220e0c68f4cc91639b2ece0540a02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-21T14:09:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jfbnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-21T14:09:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-krv5b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-21T14:10:24Z is after 2025-08-24T17:21:41Z" Nov 21 14:10:24 crc kubenswrapper[4897]: I1121 14:10:24.252554 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:24 crc kubenswrapper[4897]: I1121 14:10:24.252599 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:24 crc kubenswrapper[4897]: I1121 14:10:24.252614 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:24 crc kubenswrapper[4897]: I1121 14:10:24.252634 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:24 crc kubenswrapper[4897]: I1121 14:10:24.252650 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:24Z","lastTransitionTime":"2025-11-21T14:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:24 crc kubenswrapper[4897]: I1121 14:10:24.275796 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-7qfmq" podStartSLOduration=75.275775116 podStartE2EDuration="1m15.275775116s" podCreationTimestamp="2025-11-21 14:09:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:10:24.248153223 +0000 UTC m=+101.532746708" watchObservedRunningTime="2025-11-21 14:10:24.275775116 +0000 UTC m=+101.560368581" Nov 21 14:10:24 crc kubenswrapper[4897]: I1121 14:10:24.308419 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=75.308401918 podStartE2EDuration="1m15.308401918s" podCreationTimestamp="2025-11-21 14:09:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:10:24.2806116 +0000 UTC m=+101.565205115" watchObservedRunningTime="2025-11-21 14:10:24.308401918 +0000 UTC m=+101.592995393" Nov 21 14:10:24 crc kubenswrapper[4897]: I1121 14:10:24.308690 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=74.308685916 podStartE2EDuration="1m14.308685916s" podCreationTimestamp="2025-11-21 14:09:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:10:24.307963575 +0000 UTC m=+101.592557050" watchObservedRunningTime="2025-11-21 14:10:24.308685916 +0000 UTC m=+101.593279391" Nov 21 14:10:24 crc kubenswrapper[4897]: I1121 14:10:24.325060 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=71.324997757 podStartE2EDuration="1m11.324997757s" podCreationTimestamp="2025-11-21 14:09:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:10:24.324912905 +0000 UTC m=+101.609506380" watchObservedRunningTime="2025-11-21 14:10:24.324997757 +0000 UTC m=+101.609591252" Nov 21 14:10:24 crc kubenswrapper[4897]: I1121 14:10:24.355084 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:24 crc kubenswrapper[4897]: I1121 14:10:24.355534 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:24 crc kubenswrapper[4897]: I1121 14:10:24.355622 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:24 crc kubenswrapper[4897]: I1121 14:10:24.355727 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:24 crc kubenswrapper[4897]: I1121 14:10:24.355795 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:24Z","lastTransitionTime":"2025-11-21T14:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:24 crc kubenswrapper[4897]: I1121 14:10:24.392529 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-6ljgx" podStartSLOduration=75.392482072 podStartE2EDuration="1m15.392482072s" podCreationTimestamp="2025-11-21 14:09:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:10:24.377341663 +0000 UTC m=+101.661935138" watchObservedRunningTime="2025-11-21 14:10:24.392482072 +0000 UTC m=+101.677075547" Nov 21 14:10:24 crc kubenswrapper[4897]: I1121 14:10:24.397081 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-x5pws" podStartSLOduration=75.397060428 podStartE2EDuration="1m15.397060428s" podCreationTimestamp="2025-11-21 14:09:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:10:24.392607885 +0000 UTC m=+101.677201370" watchObservedRunningTime="2025-11-21 14:10:24.397060428 +0000 UTC m=+101.681653903" Nov 21 14:10:24 crc kubenswrapper[4897]: I1121 14:10:24.442280 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-nl76g" podStartSLOduration=75.442256857 podStartE2EDuration="1m15.442256857s" podCreationTimestamp="2025-11-21 14:09:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:10:24.441492367 +0000 UTC m=+101.726085852" watchObservedRunningTime="2025-11-21 14:10:24.442256857 +0000 UTC m=+101.726850332" Nov 21 14:10:24 crc kubenswrapper[4897]: I1121 14:10:24.458720 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:24 crc kubenswrapper[4897]: I1121 14:10:24.459168 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:24 crc kubenswrapper[4897]: I1121 14:10:24.459298 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:24 crc kubenswrapper[4897]: I1121 14:10:24.459402 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:24 crc kubenswrapper[4897]: I1121 14:10:24.459491 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:24Z","lastTransitionTime":"2025-11-21T14:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:24 crc kubenswrapper[4897]: I1121 14:10:24.561909 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:24 crc kubenswrapper[4897]: I1121 14:10:24.561970 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:24 crc kubenswrapper[4897]: I1121 14:10:24.561982 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:24 crc kubenswrapper[4897]: I1121 14:10:24.562025 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:24 crc kubenswrapper[4897]: I1121 14:10:24.562036 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:24Z","lastTransitionTime":"2025-11-21T14:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:24 crc kubenswrapper[4897]: I1121 14:10:24.663608 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:24 crc kubenswrapper[4897]: I1121 14:10:24.663655 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:24 crc kubenswrapper[4897]: I1121 14:10:24.663668 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:24 crc kubenswrapper[4897]: I1121 14:10:24.663687 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:24 crc kubenswrapper[4897]: I1121 14:10:24.663700 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:24Z","lastTransitionTime":"2025-11-21T14:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:24 crc kubenswrapper[4897]: I1121 14:10:24.766950 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:24 crc kubenswrapper[4897]: I1121 14:10:24.767303 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:24 crc kubenswrapper[4897]: I1121 14:10:24.767405 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:24 crc kubenswrapper[4897]: I1121 14:10:24.767486 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:24 crc kubenswrapper[4897]: I1121 14:10:24.767581 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:24Z","lastTransitionTime":"2025-11-21T14:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:24 crc kubenswrapper[4897]: I1121 14:10:24.869944 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:24 crc kubenswrapper[4897]: I1121 14:10:24.869977 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:24 crc kubenswrapper[4897]: I1121 14:10:24.869986 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:24 crc kubenswrapper[4897]: I1121 14:10:24.870001 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:24 crc kubenswrapper[4897]: I1121 14:10:24.870013 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:24Z","lastTransitionTime":"2025-11-21T14:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:24 crc kubenswrapper[4897]: I1121 14:10:24.972912 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:24 crc kubenswrapper[4897]: I1121 14:10:24.973159 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:24 crc kubenswrapper[4897]: I1121 14:10:24.973272 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:24 crc kubenswrapper[4897]: I1121 14:10:24.973348 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:24 crc kubenswrapper[4897]: I1121 14:10:24.973422 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:24Z","lastTransitionTime":"2025-11-21T14:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:25 crc kubenswrapper[4897]: I1121 14:10:25.076016 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:25 crc kubenswrapper[4897]: I1121 14:10:25.076050 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:25 crc kubenswrapper[4897]: I1121 14:10:25.076061 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:25 crc kubenswrapper[4897]: I1121 14:10:25.076077 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:25 crc kubenswrapper[4897]: I1121 14:10:25.076089 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:25Z","lastTransitionTime":"2025-11-21T14:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:25 crc kubenswrapper[4897]: I1121 14:10:25.088631 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 21 14:10:25 crc kubenswrapper[4897]: I1121 14:10:25.088659 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 21 14:10:25 crc kubenswrapper[4897]: I1121 14:10:25.088661 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 21 14:10:25 crc kubenswrapper[4897]: I1121 14:10:25.088764 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs4c8" Nov 21 14:10:25 crc kubenswrapper[4897]: E1121 14:10:25.088858 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 21 14:10:25 crc kubenswrapper[4897]: E1121 14:10:25.089008 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 21 14:10:25 crc kubenswrapper[4897]: E1121 14:10:25.089195 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 21 14:10:25 crc kubenswrapper[4897]: E1121 14:10:25.089254 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs4c8" podUID="01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd" Nov 21 14:10:25 crc kubenswrapper[4897]: I1121 14:10:25.178331 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:25 crc kubenswrapper[4897]: I1121 14:10:25.178378 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:25 crc kubenswrapper[4897]: I1121 14:10:25.178390 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:25 crc kubenswrapper[4897]: I1121 14:10:25.178406 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:25 crc kubenswrapper[4897]: I1121 14:10:25.178417 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:25Z","lastTransitionTime":"2025-11-21T14:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:25 crc kubenswrapper[4897]: I1121 14:10:25.280193 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:25 crc kubenswrapper[4897]: I1121 14:10:25.280231 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:25 crc kubenswrapper[4897]: I1121 14:10:25.280240 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:25 crc kubenswrapper[4897]: I1121 14:10:25.280254 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:25 crc kubenswrapper[4897]: I1121 14:10:25.280264 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:25Z","lastTransitionTime":"2025-11-21T14:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:25 crc kubenswrapper[4897]: I1121 14:10:25.381911 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:25 crc kubenswrapper[4897]: I1121 14:10:25.381945 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:25 crc kubenswrapper[4897]: I1121 14:10:25.381953 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:25 crc kubenswrapper[4897]: I1121 14:10:25.381965 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:25 crc kubenswrapper[4897]: I1121 14:10:25.381975 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:25Z","lastTransitionTime":"2025-11-21T14:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:25 crc kubenswrapper[4897]: I1121 14:10:25.484488 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:25 crc kubenswrapper[4897]: I1121 14:10:25.484555 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:25 crc kubenswrapper[4897]: I1121 14:10:25.484568 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:25 crc kubenswrapper[4897]: I1121 14:10:25.484584 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:25 crc kubenswrapper[4897]: I1121 14:10:25.484597 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:25Z","lastTransitionTime":"2025-11-21T14:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:25 crc kubenswrapper[4897]: I1121 14:10:25.587165 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:25 crc kubenswrapper[4897]: I1121 14:10:25.587207 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:25 crc kubenswrapper[4897]: I1121 14:10:25.587217 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:25 crc kubenswrapper[4897]: I1121 14:10:25.587230 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:25 crc kubenswrapper[4897]: I1121 14:10:25.587240 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:25Z","lastTransitionTime":"2025-11-21T14:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:25 crc kubenswrapper[4897]: I1121 14:10:25.689482 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:25 crc kubenswrapper[4897]: I1121 14:10:25.689547 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:25 crc kubenswrapper[4897]: I1121 14:10:25.689579 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:25 crc kubenswrapper[4897]: I1121 14:10:25.689596 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:25 crc kubenswrapper[4897]: I1121 14:10:25.689609 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:25Z","lastTransitionTime":"2025-11-21T14:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:25 crc kubenswrapper[4897]: I1121 14:10:25.792611 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:25 crc kubenswrapper[4897]: I1121 14:10:25.792660 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:25 crc kubenswrapper[4897]: I1121 14:10:25.792674 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:25 crc kubenswrapper[4897]: I1121 14:10:25.792692 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:25 crc kubenswrapper[4897]: I1121 14:10:25.792704 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:25Z","lastTransitionTime":"2025-11-21T14:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:25 crc kubenswrapper[4897]: I1121 14:10:25.895495 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:25 crc kubenswrapper[4897]: I1121 14:10:25.895607 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:25 crc kubenswrapper[4897]: I1121 14:10:25.895624 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:25 crc kubenswrapper[4897]: I1121 14:10:25.895650 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:25 crc kubenswrapper[4897]: I1121 14:10:25.895671 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:25Z","lastTransitionTime":"2025-11-21T14:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:25 crc kubenswrapper[4897]: I1121 14:10:25.997935 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:25 crc kubenswrapper[4897]: I1121 14:10:25.997995 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:25 crc kubenswrapper[4897]: I1121 14:10:25.998005 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:25 crc kubenswrapper[4897]: I1121 14:10:25.998022 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:25 crc kubenswrapper[4897]: I1121 14:10:25.998036 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:25Z","lastTransitionTime":"2025-11-21T14:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:26 crc kubenswrapper[4897]: I1121 14:10:26.099933 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:26 crc kubenswrapper[4897]: I1121 14:10:26.099981 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:26 crc kubenswrapper[4897]: I1121 14:10:26.099994 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:26 crc kubenswrapper[4897]: I1121 14:10:26.100015 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:26 crc kubenswrapper[4897]: I1121 14:10:26.100029 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:26Z","lastTransitionTime":"2025-11-21T14:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:26 crc kubenswrapper[4897]: I1121 14:10:26.202857 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:26 crc kubenswrapper[4897]: I1121 14:10:26.202909 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:26 crc kubenswrapper[4897]: I1121 14:10:26.202920 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:26 crc kubenswrapper[4897]: I1121 14:10:26.202939 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:26 crc kubenswrapper[4897]: I1121 14:10:26.202952 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:26Z","lastTransitionTime":"2025-11-21T14:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:26 crc kubenswrapper[4897]: I1121 14:10:26.305357 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:26 crc kubenswrapper[4897]: I1121 14:10:26.305398 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:26 crc kubenswrapper[4897]: I1121 14:10:26.305406 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:26 crc kubenswrapper[4897]: I1121 14:10:26.305435 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:26 crc kubenswrapper[4897]: I1121 14:10:26.305445 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:26Z","lastTransitionTime":"2025-11-21T14:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:26 crc kubenswrapper[4897]: I1121 14:10:26.407811 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:26 crc kubenswrapper[4897]: I1121 14:10:26.407857 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:26 crc kubenswrapper[4897]: I1121 14:10:26.407869 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:26 crc kubenswrapper[4897]: I1121 14:10:26.407884 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:26 crc kubenswrapper[4897]: I1121 14:10:26.407896 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:26Z","lastTransitionTime":"2025-11-21T14:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:26 crc kubenswrapper[4897]: I1121 14:10:26.510055 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:26 crc kubenswrapper[4897]: I1121 14:10:26.510098 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:26 crc kubenswrapper[4897]: I1121 14:10:26.510111 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:26 crc kubenswrapper[4897]: I1121 14:10:26.510127 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:26 crc kubenswrapper[4897]: I1121 14:10:26.510137 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:26Z","lastTransitionTime":"2025-11-21T14:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:26 crc kubenswrapper[4897]: I1121 14:10:26.612161 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:26 crc kubenswrapper[4897]: I1121 14:10:26.612215 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:26 crc kubenswrapper[4897]: I1121 14:10:26.612225 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:26 crc kubenswrapper[4897]: I1121 14:10:26.612242 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:26 crc kubenswrapper[4897]: I1121 14:10:26.612256 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:26Z","lastTransitionTime":"2025-11-21T14:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:26 crc kubenswrapper[4897]: I1121 14:10:26.715228 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:26 crc kubenswrapper[4897]: I1121 14:10:26.715287 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:26 crc kubenswrapper[4897]: I1121 14:10:26.715300 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:26 crc kubenswrapper[4897]: I1121 14:10:26.715320 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:26 crc kubenswrapper[4897]: I1121 14:10:26.715334 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:26Z","lastTransitionTime":"2025-11-21T14:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:26 crc kubenswrapper[4897]: I1121 14:10:26.817876 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:26 crc kubenswrapper[4897]: I1121 14:10:26.817923 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:26 crc kubenswrapper[4897]: I1121 14:10:26.817931 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:26 crc kubenswrapper[4897]: I1121 14:10:26.817947 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:26 crc kubenswrapper[4897]: I1121 14:10:26.817957 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:26Z","lastTransitionTime":"2025-11-21T14:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:26 crc kubenswrapper[4897]: I1121 14:10:26.920622 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:26 crc kubenswrapper[4897]: I1121 14:10:26.920658 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:26 crc kubenswrapper[4897]: I1121 14:10:26.920667 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:26 crc kubenswrapper[4897]: I1121 14:10:26.920682 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:26 crc kubenswrapper[4897]: I1121 14:10:26.920716 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:26Z","lastTransitionTime":"2025-11-21T14:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:27 crc kubenswrapper[4897]: I1121 14:10:27.023687 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:27 crc kubenswrapper[4897]: I1121 14:10:27.023731 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:27 crc kubenswrapper[4897]: I1121 14:10:27.023743 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:27 crc kubenswrapper[4897]: I1121 14:10:27.023759 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:27 crc kubenswrapper[4897]: I1121 14:10:27.023770 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:27Z","lastTransitionTime":"2025-11-21T14:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:27 crc kubenswrapper[4897]: I1121 14:10:27.088421 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 21 14:10:27 crc kubenswrapper[4897]: I1121 14:10:27.088421 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 21 14:10:27 crc kubenswrapper[4897]: E1121 14:10:27.088623 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 21 14:10:27 crc kubenswrapper[4897]: I1121 14:10:27.088560 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs4c8" Nov 21 14:10:27 crc kubenswrapper[4897]: I1121 14:10:27.088537 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 21 14:10:27 crc kubenswrapper[4897]: E1121 14:10:27.088712 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 21 14:10:27 crc kubenswrapper[4897]: E1121 14:10:27.088814 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 21 14:10:27 crc kubenswrapper[4897]: E1121 14:10:27.089024 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs4c8" podUID="01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd" Nov 21 14:10:27 crc kubenswrapper[4897]: I1121 14:10:27.125967 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:27 crc kubenswrapper[4897]: I1121 14:10:27.126003 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:27 crc kubenswrapper[4897]: I1121 14:10:27.126015 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:27 crc kubenswrapper[4897]: I1121 14:10:27.126033 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:27 crc kubenswrapper[4897]: I1121 14:10:27.126046 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:27Z","lastTransitionTime":"2025-11-21T14:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:27 crc kubenswrapper[4897]: I1121 14:10:27.227959 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:27 crc kubenswrapper[4897]: I1121 14:10:27.228005 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:27 crc kubenswrapper[4897]: I1121 14:10:27.228014 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:27 crc kubenswrapper[4897]: I1121 14:10:27.228028 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:27 crc kubenswrapper[4897]: I1121 14:10:27.228038 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:27Z","lastTransitionTime":"2025-11-21T14:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:27 crc kubenswrapper[4897]: I1121 14:10:27.324697 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 21 14:10:27 crc kubenswrapper[4897]: I1121 14:10:27.324801 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 21 14:10:27 crc kubenswrapper[4897]: I1121 14:10:27.324817 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 21 14:10:27 crc kubenswrapper[4897]: I1121 14:10:27.324844 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 21 14:10:27 crc kubenswrapper[4897]: I1121 14:10:27.324861 4897 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-21T14:10:27Z","lastTransitionTime":"2025-11-21T14:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 21 14:10:27 crc kubenswrapper[4897]: I1121 14:10:27.377968 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-f5jzp"] Nov 21 14:10:27 crc kubenswrapper[4897]: I1121 14:10:27.379020 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-f5jzp" Nov 21 14:10:27 crc kubenswrapper[4897]: I1121 14:10:27.382016 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Nov 21 14:10:27 crc kubenswrapper[4897]: I1121 14:10:27.382068 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Nov 21 14:10:27 crc kubenswrapper[4897]: I1121 14:10:27.382109 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Nov 21 14:10:27 crc kubenswrapper[4897]: I1121 14:10:27.382477 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Nov 21 14:10:27 crc kubenswrapper[4897]: I1121 14:10:27.411587 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=10.411553434 podStartE2EDuration="10.411553434s" podCreationTimestamp="2025-11-21 14:10:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:10:27.410481994 +0000 UTC m=+104.695075479" watchObservedRunningTime="2025-11-21 14:10:27.411553434 +0000 UTC m=+104.696146929" Nov 21 14:10:27 crc kubenswrapper[4897]: I1121 14:10:27.441868 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/c1a017f2-8450-4d42-bd81-4ee1b1692259-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-f5jzp\" (UID: \"c1a017f2-8450-4d42-bd81-4ee1b1692259\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-f5jzp" Nov 21 14:10:27 crc kubenswrapper[4897]: I1121 14:10:27.441961 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/c1a017f2-8450-4d42-bd81-4ee1b1692259-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-f5jzp\" (UID: \"c1a017f2-8450-4d42-bd81-4ee1b1692259\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-f5jzp" Nov 21 14:10:27 crc kubenswrapper[4897]: I1121 14:10:27.442024 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c1a017f2-8450-4d42-bd81-4ee1b1692259-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-f5jzp\" (UID: \"c1a017f2-8450-4d42-bd81-4ee1b1692259\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-f5jzp" Nov 21 14:10:27 crc kubenswrapper[4897]: I1121 14:10:27.442061 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c1a017f2-8450-4d42-bd81-4ee1b1692259-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-f5jzp\" (UID: \"c1a017f2-8450-4d42-bd81-4ee1b1692259\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-f5jzp" Nov 21 14:10:27 crc kubenswrapper[4897]: I1121 14:10:27.442123 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c1a017f2-8450-4d42-bd81-4ee1b1692259-service-ca\") pod \"cluster-version-operator-5c965bbfc6-f5jzp\" (UID: \"c1a017f2-8450-4d42-bd81-4ee1b1692259\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-f5jzp" Nov 21 14:10:27 crc kubenswrapper[4897]: I1121 14:10:27.442384 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-4qwqj" podStartSLOduration=79.442368286 podStartE2EDuration="1m19.442368286s" podCreationTimestamp="2025-11-21 14:09:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:10:27.441297866 +0000 UTC m=+104.725891351" watchObservedRunningTime="2025-11-21 14:10:27.442368286 +0000 UTC m=+104.726961781" Nov 21 14:10:27 crc kubenswrapper[4897]: I1121 14:10:27.489191 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=48.489158158 podStartE2EDuration="48.489158158s" podCreationTimestamp="2025-11-21 14:09:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:10:27.471966123 +0000 UTC m=+104.756559618" watchObservedRunningTime="2025-11-21 14:10:27.489158158 +0000 UTC m=+104.773751643" Nov 21 14:10:27 crc kubenswrapper[4897]: I1121 14:10:27.499442 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podStartSLOduration=78.499413432 podStartE2EDuration="1m18.499413432s" podCreationTimestamp="2025-11-21 14:09:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:10:27.499034841 +0000 UTC m=+104.783628316" watchObservedRunningTime="2025-11-21 14:10:27.499413432 +0000 UTC m=+104.784006917" Nov 21 14:10:27 crc kubenswrapper[4897]: I1121 14:10:27.543973 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/c1a017f2-8450-4d42-bd81-4ee1b1692259-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-f5jzp\" (UID: \"c1a017f2-8450-4d42-bd81-4ee1b1692259\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-f5jzp" Nov 21 14:10:27 crc kubenswrapper[4897]: I1121 14:10:27.544037 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/c1a017f2-8450-4d42-bd81-4ee1b1692259-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-f5jzp\" (UID: \"c1a017f2-8450-4d42-bd81-4ee1b1692259\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-f5jzp" Nov 21 14:10:27 crc kubenswrapper[4897]: I1121 14:10:27.544094 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c1a017f2-8450-4d42-bd81-4ee1b1692259-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-f5jzp\" (UID: \"c1a017f2-8450-4d42-bd81-4ee1b1692259\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-f5jzp" Nov 21 14:10:27 crc kubenswrapper[4897]: I1121 14:10:27.544115 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c1a017f2-8450-4d42-bd81-4ee1b1692259-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-f5jzp\" (UID: \"c1a017f2-8450-4d42-bd81-4ee1b1692259\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-f5jzp" Nov 21 14:10:27 crc kubenswrapper[4897]: I1121 14:10:27.544138 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c1a017f2-8450-4d42-bd81-4ee1b1692259-service-ca\") pod \"cluster-version-operator-5c965bbfc6-f5jzp\" (UID: \"c1a017f2-8450-4d42-bd81-4ee1b1692259\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-f5jzp" Nov 21 14:10:27 crc kubenswrapper[4897]: I1121 14:10:27.544124 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/c1a017f2-8450-4d42-bd81-4ee1b1692259-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-f5jzp\" (UID: \"c1a017f2-8450-4d42-bd81-4ee1b1692259\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-f5jzp" Nov 21 14:10:27 crc kubenswrapper[4897]: I1121 14:10:27.544237 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/c1a017f2-8450-4d42-bd81-4ee1b1692259-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-f5jzp\" (UID: \"c1a017f2-8450-4d42-bd81-4ee1b1692259\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-f5jzp" Nov 21 14:10:27 crc kubenswrapper[4897]: I1121 14:10:27.546113 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c1a017f2-8450-4d42-bd81-4ee1b1692259-service-ca\") pod \"cluster-version-operator-5c965bbfc6-f5jzp\" (UID: \"c1a017f2-8450-4d42-bd81-4ee1b1692259\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-f5jzp" Nov 21 14:10:27 crc kubenswrapper[4897]: I1121 14:10:27.557157 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c1a017f2-8450-4d42-bd81-4ee1b1692259-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-f5jzp\" (UID: \"c1a017f2-8450-4d42-bd81-4ee1b1692259\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-f5jzp" Nov 21 14:10:27 crc kubenswrapper[4897]: I1121 14:10:27.566915 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c1a017f2-8450-4d42-bd81-4ee1b1692259-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-f5jzp\" (UID: \"c1a017f2-8450-4d42-bd81-4ee1b1692259\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-f5jzp" Nov 21 14:10:27 crc kubenswrapper[4897]: I1121 14:10:27.698715 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-f5jzp" Nov 21 14:10:27 crc kubenswrapper[4897]: W1121 14:10:27.713476 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc1a017f2_8450_4d42_bd81_4ee1b1692259.slice/crio-ee2feb491c62559015f13f7de3d6183803b47ab8b770419d1941dd6e19ac3485 WatchSource:0}: Error finding container ee2feb491c62559015f13f7de3d6183803b47ab8b770419d1941dd6e19ac3485: Status 404 returned error can't find the container with id ee2feb491c62559015f13f7de3d6183803b47ab8b770419d1941dd6e19ac3485 Nov 21 14:10:27 crc kubenswrapper[4897]: I1121 14:10:27.747080 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd-metrics-certs\") pod \"network-metrics-daemon-rs4c8\" (UID: \"01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd\") " pod="openshift-multus/network-metrics-daemon-rs4c8" Nov 21 14:10:27 crc kubenswrapper[4897]: E1121 14:10:27.747354 4897 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 21 14:10:27 crc kubenswrapper[4897]: E1121 14:10:27.747475 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd-metrics-certs podName:01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd nodeName:}" failed. No retries permitted until 2025-11-21 14:11:31.747451358 +0000 UTC m=+169.032044833 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd-metrics-certs") pod "network-metrics-daemon-rs4c8" (UID: "01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 21 14:10:28 crc kubenswrapper[4897]: I1121 14:10:28.676971 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-f5jzp" event={"ID":"c1a017f2-8450-4d42-bd81-4ee1b1692259","Type":"ContainerStarted","Data":"9bcc849077eb92c7761092db7f2956ef2d022acf5a8b1fcf049ccb7dfe0f6312"} Nov 21 14:10:28 crc kubenswrapper[4897]: I1121 14:10:28.677039 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-f5jzp" event={"ID":"c1a017f2-8450-4d42-bd81-4ee1b1692259","Type":"ContainerStarted","Data":"ee2feb491c62559015f13f7de3d6183803b47ab8b770419d1941dd6e19ac3485"} Nov 21 14:10:29 crc kubenswrapper[4897]: I1121 14:10:29.088491 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs4c8" Nov 21 14:10:29 crc kubenswrapper[4897]: I1121 14:10:29.088549 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 21 14:10:29 crc kubenswrapper[4897]: I1121 14:10:29.088562 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 21 14:10:29 crc kubenswrapper[4897]: E1121 14:10:29.088680 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs4c8" podUID="01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd" Nov 21 14:10:29 crc kubenswrapper[4897]: I1121 14:10:29.088801 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 21 14:10:29 crc kubenswrapper[4897]: E1121 14:10:29.088918 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 21 14:10:29 crc kubenswrapper[4897]: E1121 14:10:29.089063 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 21 14:10:29 crc kubenswrapper[4897]: E1121 14:10:29.089119 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 21 14:10:31 crc kubenswrapper[4897]: I1121 14:10:31.087951 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 21 14:10:31 crc kubenswrapper[4897]: I1121 14:10:31.087966 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 21 14:10:31 crc kubenswrapper[4897]: I1121 14:10:31.087986 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 21 14:10:31 crc kubenswrapper[4897]: I1121 14:10:31.088808 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs4c8" Nov 21 14:10:31 crc kubenswrapper[4897]: E1121 14:10:31.088890 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 21 14:10:31 crc kubenswrapper[4897]: E1121 14:10:31.088969 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 21 14:10:31 crc kubenswrapper[4897]: E1121 14:10:31.089051 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs4c8" podUID="01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd" Nov 21 14:10:31 crc kubenswrapper[4897]: E1121 14:10:31.089137 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 21 14:10:33 crc kubenswrapper[4897]: I1121 14:10:33.088209 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 21 14:10:33 crc kubenswrapper[4897]: I1121 14:10:33.088248 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 21 14:10:33 crc kubenswrapper[4897]: I1121 14:10:33.088274 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 21 14:10:33 crc kubenswrapper[4897]: E1121 14:10:33.088846 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 21 14:10:33 crc kubenswrapper[4897]: E1121 14:10:33.088680 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 21 14:10:33 crc kubenswrapper[4897]: I1121 14:10:33.088326 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs4c8" Nov 21 14:10:33 crc kubenswrapper[4897]: E1121 14:10:33.088927 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 21 14:10:33 crc kubenswrapper[4897]: E1121 14:10:33.088993 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs4c8" podUID="01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd" Nov 21 14:10:35 crc kubenswrapper[4897]: I1121 14:10:35.088384 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 21 14:10:35 crc kubenswrapper[4897]: I1121 14:10:35.088449 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 21 14:10:35 crc kubenswrapper[4897]: I1121 14:10:35.088582 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 21 14:10:35 crc kubenswrapper[4897]: I1121 14:10:35.088623 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs4c8" Nov 21 14:10:35 crc kubenswrapper[4897]: E1121 14:10:35.088747 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 21 14:10:35 crc kubenswrapper[4897]: E1121 14:10:35.088831 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs4c8" podUID="01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd" Nov 21 14:10:35 crc kubenswrapper[4897]: E1121 14:10:35.089066 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 21 14:10:35 crc kubenswrapper[4897]: E1121 14:10:35.088968 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 21 14:10:36 crc kubenswrapper[4897]: I1121 14:10:36.089047 4897 scope.go:117] "RemoveContainer" containerID="40b0bddce48b3626ad889ae2f43deb866acd521a8aab71253506364bfb47c723" Nov 21 14:10:36 crc kubenswrapper[4897]: E1121 14:10:36.089254 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-tpfb6_openshift-ovn-kubernetes(222f65f2-0b82-4760-bcf7-779244720f01)\"" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" podUID="222f65f2-0b82-4760-bcf7-779244720f01" Nov 21 14:10:37 crc kubenswrapper[4897]: I1121 14:10:37.088137 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs4c8" Nov 21 14:10:37 crc kubenswrapper[4897]: I1121 14:10:37.088187 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 21 14:10:37 crc kubenswrapper[4897]: I1121 14:10:37.088222 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 21 14:10:37 crc kubenswrapper[4897]: I1121 14:10:37.088138 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 21 14:10:37 crc kubenswrapper[4897]: E1121 14:10:37.088318 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs4c8" podUID="01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd" Nov 21 14:10:37 crc kubenswrapper[4897]: E1121 14:10:37.088370 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 21 14:10:37 crc kubenswrapper[4897]: E1121 14:10:37.088416 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 21 14:10:37 crc kubenswrapper[4897]: E1121 14:10:37.088451 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 21 14:10:39 crc kubenswrapper[4897]: I1121 14:10:39.088462 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 21 14:10:39 crc kubenswrapper[4897]: E1121 14:10:39.088714 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 21 14:10:39 crc kubenswrapper[4897]: I1121 14:10:39.088484 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs4c8" Nov 21 14:10:39 crc kubenswrapper[4897]: I1121 14:10:39.088478 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 21 14:10:39 crc kubenswrapper[4897]: E1121 14:10:39.088836 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs4c8" podUID="01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd" Nov 21 14:10:39 crc kubenswrapper[4897]: I1121 14:10:39.088388 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 21 14:10:39 crc kubenswrapper[4897]: E1121 14:10:39.089246 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 21 14:10:39 crc kubenswrapper[4897]: E1121 14:10:39.089394 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 21 14:10:41 crc kubenswrapper[4897]: I1121 14:10:41.088576 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 21 14:10:41 crc kubenswrapper[4897]: I1121 14:10:41.088579 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 21 14:10:41 crc kubenswrapper[4897]: I1121 14:10:41.088727 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs4c8" Nov 21 14:10:41 crc kubenswrapper[4897]: I1121 14:10:41.088737 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 21 14:10:41 crc kubenswrapper[4897]: E1121 14:10:41.088860 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 21 14:10:41 crc kubenswrapper[4897]: E1121 14:10:41.089045 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 21 14:10:41 crc kubenswrapper[4897]: E1121 14:10:41.089177 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs4c8" podUID="01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd" Nov 21 14:10:41 crc kubenswrapper[4897]: E1121 14:10:41.089298 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 21 14:10:43 crc kubenswrapper[4897]: I1121 14:10:43.088771 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 21 14:10:43 crc kubenswrapper[4897]: E1121 14:10:43.088941 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 21 14:10:43 crc kubenswrapper[4897]: I1121 14:10:43.088791 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 21 14:10:43 crc kubenswrapper[4897]: I1121 14:10:43.089082 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs4c8" Nov 21 14:10:43 crc kubenswrapper[4897]: E1121 14:10:43.089145 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 21 14:10:43 crc kubenswrapper[4897]: I1121 14:10:43.089122 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 21 14:10:43 crc kubenswrapper[4897]: E1121 14:10:43.089303 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs4c8" podUID="01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd" Nov 21 14:10:43 crc kubenswrapper[4897]: E1121 14:10:43.089551 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 21 14:10:44 crc kubenswrapper[4897]: E1121 14:10:44.042012 4897 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Nov 21 14:10:44 crc kubenswrapper[4897]: E1121 14:10:44.244043 4897 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 21 14:10:45 crc kubenswrapper[4897]: I1121 14:10:45.088432 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 21 14:10:45 crc kubenswrapper[4897]: I1121 14:10:45.088475 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 21 14:10:45 crc kubenswrapper[4897]: I1121 14:10:45.088432 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 21 14:10:45 crc kubenswrapper[4897]: E1121 14:10:45.088922 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 21 14:10:45 crc kubenswrapper[4897]: E1121 14:10:45.089072 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 21 14:10:45 crc kubenswrapper[4897]: E1121 14:10:45.089160 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 21 14:10:45 crc kubenswrapper[4897]: I1121 14:10:45.089040 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs4c8" Nov 21 14:10:45 crc kubenswrapper[4897]: E1121 14:10:45.089374 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs4c8" podUID="01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd" Nov 21 14:10:45 crc kubenswrapper[4897]: I1121 14:10:45.731976 4897 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-6ljgx_5f3b70b5-9294-4f69-9528-500d28f34c89/kube-multus/1.log" Nov 21 14:10:45 crc kubenswrapper[4897]: I1121 14:10:45.732937 4897 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-6ljgx_5f3b70b5-9294-4f69-9528-500d28f34c89/kube-multus/0.log" Nov 21 14:10:45 crc kubenswrapper[4897]: I1121 14:10:45.733055 4897 generic.go:334] "Generic (PLEG): container finished" podID="5f3b70b5-9294-4f69-9528-500d28f34c89" containerID="bf92ab70c5a3196bf83f427ccf03acf188d1995766a9154bcd6267e3926c3ac8" exitCode=1 Nov 21 14:10:45 crc kubenswrapper[4897]: I1121 14:10:45.733155 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-6ljgx" event={"ID":"5f3b70b5-9294-4f69-9528-500d28f34c89","Type":"ContainerDied","Data":"bf92ab70c5a3196bf83f427ccf03acf188d1995766a9154bcd6267e3926c3ac8"} Nov 21 14:10:45 crc kubenswrapper[4897]: I1121 14:10:45.733239 4897 scope.go:117] "RemoveContainer" containerID="527cf28abdeef1b42d9c19fb88e3171dc05c4516c3a542a461b5b1daa85c6d25" Nov 21 14:10:45 crc kubenswrapper[4897]: I1121 14:10:45.733634 4897 scope.go:117] "RemoveContainer" containerID="bf92ab70c5a3196bf83f427ccf03acf188d1995766a9154bcd6267e3926c3ac8" Nov 21 14:10:45 crc kubenswrapper[4897]: E1121 14:10:45.733824 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-6ljgx_openshift-multus(5f3b70b5-9294-4f69-9528-500d28f34c89)\"" pod="openshift-multus/multus-6ljgx" podUID="5f3b70b5-9294-4f69-9528-500d28f34c89" Nov 21 14:10:45 crc kubenswrapper[4897]: I1121 14:10:45.755728 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-f5jzp" podStartSLOduration=97.755709403 podStartE2EDuration="1m37.755709403s" podCreationTimestamp="2025-11-21 14:09:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:10:28.696906199 +0000 UTC m=+105.981499674" watchObservedRunningTime="2025-11-21 14:10:45.755709403 +0000 UTC m=+123.040302878" Nov 21 14:10:46 crc kubenswrapper[4897]: I1121 14:10:46.736589 4897 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-6ljgx_5f3b70b5-9294-4f69-9528-500d28f34c89/kube-multus/1.log" Nov 21 14:10:47 crc kubenswrapper[4897]: I1121 14:10:47.088661 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 21 14:10:47 crc kubenswrapper[4897]: I1121 14:10:47.088657 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs4c8" Nov 21 14:10:47 crc kubenswrapper[4897]: E1121 14:10:47.088800 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 21 14:10:47 crc kubenswrapper[4897]: I1121 14:10:47.088678 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 21 14:10:47 crc kubenswrapper[4897]: E1121 14:10:47.088878 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs4c8" podUID="01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd" Nov 21 14:10:47 crc kubenswrapper[4897]: I1121 14:10:47.088658 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 21 14:10:47 crc kubenswrapper[4897]: E1121 14:10:47.088946 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 21 14:10:47 crc kubenswrapper[4897]: E1121 14:10:47.089128 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 21 14:10:49 crc kubenswrapper[4897]: I1121 14:10:49.088389 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 21 14:10:49 crc kubenswrapper[4897]: E1121 14:10:49.089595 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 21 14:10:49 crc kubenswrapper[4897]: I1121 14:10:49.088377 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 21 14:10:49 crc kubenswrapper[4897]: E1121 14:10:49.089860 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 21 14:10:49 crc kubenswrapper[4897]: I1121 14:10:49.089263 4897 scope.go:117] "RemoveContainer" containerID="40b0bddce48b3626ad889ae2f43deb866acd521a8aab71253506364bfb47c723" Nov 21 14:10:49 crc kubenswrapper[4897]: I1121 14:10:49.088385 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 21 14:10:49 crc kubenswrapper[4897]: E1121 14:10:49.090167 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 21 14:10:49 crc kubenswrapper[4897]: I1121 14:10:49.088377 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs4c8" Nov 21 14:10:49 crc kubenswrapper[4897]: E1121 14:10:49.090312 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs4c8" podUID="01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd" Nov 21 14:10:49 crc kubenswrapper[4897]: E1121 14:10:49.245057 4897 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 21 14:10:49 crc kubenswrapper[4897]: I1121 14:10:49.747274 4897 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tpfb6_222f65f2-0b82-4760-bcf7-779244720f01/ovnkube-controller/3.log" Nov 21 14:10:49 crc kubenswrapper[4897]: I1121 14:10:49.749630 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" event={"ID":"222f65f2-0b82-4760-bcf7-779244720f01","Type":"ContainerStarted","Data":"eaaa7662c416148afbfc746cc7b87fac845ce54740988de2c6d0af4b1a21b9e7"} Nov 21 14:10:49 crc kubenswrapper[4897]: I1121 14:10:49.750077 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" Nov 21 14:10:49 crc kubenswrapper[4897]: I1121 14:10:49.777437 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" podStartSLOduration=100.777418668 podStartE2EDuration="1m40.777418668s" podCreationTimestamp="2025-11-21 14:09:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:10:49.776256426 +0000 UTC m=+127.060849901" watchObservedRunningTime="2025-11-21 14:10:49.777418668 +0000 UTC m=+127.062012143" Nov 21 14:10:49 crc kubenswrapper[4897]: I1121 14:10:49.950997 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-rs4c8"] Nov 21 14:10:49 crc kubenswrapper[4897]: I1121 14:10:49.951104 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs4c8" Nov 21 14:10:49 crc kubenswrapper[4897]: E1121 14:10:49.951192 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs4c8" podUID="01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd" Nov 21 14:10:51 crc kubenswrapper[4897]: I1121 14:10:51.088371 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 21 14:10:51 crc kubenswrapper[4897]: E1121 14:10:51.088815 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 21 14:10:51 crc kubenswrapper[4897]: I1121 14:10:51.088391 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs4c8" Nov 21 14:10:51 crc kubenswrapper[4897]: I1121 14:10:51.088371 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 21 14:10:51 crc kubenswrapper[4897]: E1121 14:10:51.088952 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs4c8" podUID="01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd" Nov 21 14:10:51 crc kubenswrapper[4897]: I1121 14:10:51.088444 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 21 14:10:51 crc kubenswrapper[4897]: E1121 14:10:51.089010 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 21 14:10:51 crc kubenswrapper[4897]: E1121 14:10:51.089051 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 21 14:10:53 crc kubenswrapper[4897]: I1121 14:10:53.088913 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 21 14:10:53 crc kubenswrapper[4897]: I1121 14:10:53.089058 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs4c8" Nov 21 14:10:53 crc kubenswrapper[4897]: I1121 14:10:53.089065 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 21 14:10:53 crc kubenswrapper[4897]: I1121 14:10:53.089083 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 21 14:10:53 crc kubenswrapper[4897]: E1121 14:10:53.089193 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 21 14:10:53 crc kubenswrapper[4897]: E1121 14:10:53.089312 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs4c8" podUID="01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd" Nov 21 14:10:53 crc kubenswrapper[4897]: E1121 14:10:53.089404 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 21 14:10:53 crc kubenswrapper[4897]: E1121 14:10:53.089478 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 21 14:10:54 crc kubenswrapper[4897]: E1121 14:10:54.245948 4897 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 21 14:10:55 crc kubenswrapper[4897]: I1121 14:10:55.088618 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 21 14:10:55 crc kubenswrapper[4897]: I1121 14:10:55.088682 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 21 14:10:55 crc kubenswrapper[4897]: I1121 14:10:55.088692 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 21 14:10:55 crc kubenswrapper[4897]: E1121 14:10:55.088769 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 21 14:10:55 crc kubenswrapper[4897]: E1121 14:10:55.088873 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 21 14:10:55 crc kubenswrapper[4897]: I1121 14:10:55.088938 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs4c8" Nov 21 14:10:55 crc kubenswrapper[4897]: E1121 14:10:55.088969 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 21 14:10:55 crc kubenswrapper[4897]: E1121 14:10:55.089059 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs4c8" podUID="01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd" Nov 21 14:10:57 crc kubenswrapper[4897]: I1121 14:10:57.088186 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 21 14:10:57 crc kubenswrapper[4897]: I1121 14:10:57.088273 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 21 14:10:57 crc kubenswrapper[4897]: I1121 14:10:57.088325 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 21 14:10:57 crc kubenswrapper[4897]: E1121 14:10:57.088423 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 21 14:10:57 crc kubenswrapper[4897]: E1121 14:10:57.088495 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 21 14:10:57 crc kubenswrapper[4897]: E1121 14:10:57.088601 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 21 14:10:57 crc kubenswrapper[4897]: I1121 14:10:57.088943 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs4c8" Nov 21 14:10:57 crc kubenswrapper[4897]: E1121 14:10:57.089049 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs4c8" podUID="01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd" Nov 21 14:10:57 crc kubenswrapper[4897]: I1121 14:10:57.089499 4897 scope.go:117] "RemoveContainer" containerID="bf92ab70c5a3196bf83f427ccf03acf188d1995766a9154bcd6267e3926c3ac8" Nov 21 14:10:57 crc kubenswrapper[4897]: I1121 14:10:57.777626 4897 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-6ljgx_5f3b70b5-9294-4f69-9528-500d28f34c89/kube-multus/1.log" Nov 21 14:10:57 crc kubenswrapper[4897]: I1121 14:10:57.777680 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-6ljgx" event={"ID":"5f3b70b5-9294-4f69-9528-500d28f34c89","Type":"ContainerStarted","Data":"382da75081f6e2afec6e1492c40398e4b099071e26f7162a289311aedd0d7777"} Nov 21 14:10:59 crc kubenswrapper[4897]: I1121 14:10:59.087895 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 21 14:10:59 crc kubenswrapper[4897]: I1121 14:10:59.087974 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 21 14:10:59 crc kubenswrapper[4897]: I1121 14:10:59.087922 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 21 14:10:59 crc kubenswrapper[4897]: E1121 14:10:59.088049 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 21 14:10:59 crc kubenswrapper[4897]: E1121 14:10:59.088204 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 21 14:10:59 crc kubenswrapper[4897]: E1121 14:10:59.088244 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 21 14:10:59 crc kubenswrapper[4897]: I1121 14:10:59.088605 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs4c8" Nov 21 14:10:59 crc kubenswrapper[4897]: E1121 14:10:59.088743 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-rs4c8" podUID="01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd" Nov 21 14:11:01 crc kubenswrapper[4897]: I1121 14:11:01.089002 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs4c8" Nov 21 14:11:01 crc kubenswrapper[4897]: I1121 14:11:01.089025 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 21 14:11:01 crc kubenswrapper[4897]: I1121 14:11:01.089062 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 21 14:11:01 crc kubenswrapper[4897]: I1121 14:11:01.089086 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 21 14:11:01 crc kubenswrapper[4897]: I1121 14:11:01.092984 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Nov 21 14:11:01 crc kubenswrapper[4897]: I1121 14:11:01.092998 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Nov 21 14:11:01 crc kubenswrapper[4897]: I1121 14:11:01.093115 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Nov 21 14:11:01 crc kubenswrapper[4897]: I1121 14:11:01.093165 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Nov 21 14:11:01 crc kubenswrapper[4897]: I1121 14:11:01.093649 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Nov 21 14:11:01 crc kubenswrapper[4897]: I1121 14:11:01.093761 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.093393 4897 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.123277 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-d8pkd"] Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.123892 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-d8pkd" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.125605 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-g5z94"] Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.126113 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-g5z94" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.126360 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gxw89"] Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.126879 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gxw89" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.128010 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.128745 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.128872 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.129412 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.129606 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.129631 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.129727 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.129863 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.131855 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-gspw5"] Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.132221 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p85cq"] Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.132422 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-pgf6s"] Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.132734 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-gspw5" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.132877 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p85cq" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.133020 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-pgf6s" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.133894 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.134217 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.134477 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.134571 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.134660 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.134742 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.134800 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.135844 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-hk7mf"] Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.136206 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.136305 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.136386 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.136438 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.136544 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.136610 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-hk7mf" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.137163 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-2trmp"] Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.143554 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.147200 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-dfltg"] Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.155664 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-rnnlh"] Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.156084 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-dfltg" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.159092 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-2trmp" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.160193 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.160352 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.160468 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.160646 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.160647 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.162034 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-rnnlh" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.162987 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.163386 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.163504 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.163682 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.163792 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.163885 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.164020 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.164108 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.163700 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.164198 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.163748 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.164299 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-lclf5"] Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.164927 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lclf5" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.166904 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.169234 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.169398 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.169608 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.171014 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.171150 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.171389 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.172213 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.172445 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.172723 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.172931 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.173046 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.173169 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.173586 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-mnn4f"] Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.173613 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.174349 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-mnn4f" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.174519 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.174927 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.175066 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.175196 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-vgwv2"] Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.175304 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.175610 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.175713 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.175736 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rv2nz"] Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.176009 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.176062 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rv2nz" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.176201 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.176383 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-vgwv2" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.176400 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.176481 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.176543 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.176487 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.177735 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.178415 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.179861 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-7fsr9"] Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.180218 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-8kddm"] Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.180556 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-8kddm" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.180558 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-7fsr9" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.184467 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-t5jsz"] Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.185206 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-t5jsz" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.188399 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dpwxx\" (UniqueName: \"kubernetes.io/projected/92733842-6b61-4327-b35c-e37cc833f7a3-kube-api-access-dpwxx\") pod \"cluster-image-registry-operator-dc59b4c8b-p85cq\" (UID: \"92733842-6b61-4327-b35c-e37cc833f7a3\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p85cq" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.188447 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/69303f47-f9a7-4c1c-b569-b1e4de56b2ff-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-hk7mf\" (UID: \"69303f47-f9a7-4c1c-b569-b1e4de56b2ff\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-hk7mf" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.188465 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n9k72\" (UniqueName: \"kubernetes.io/projected/4bf1a18e-ec09-4f71-b6a1-645698067dc1-kube-api-access-n9k72\") pod \"controller-manager-879f6c89f-rnnlh\" (UID: \"4bf1a18e-ec09-4f71-b6a1-645698067dc1\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rnnlh" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.188489 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/0e4b5f84-4954-4d85-8bae-db18ed352464-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-2trmp\" (UID: \"0e4b5f84-4954-4d85-8bae-db18ed352464\") " pod="openshift-authentication/oauth-openshift-558db77b4-2trmp" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.188521 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/18d8d58b-3dab-4949-b9d1-313870a96dd5-serving-cert\") pod \"console-operator-58897d9998-dfltg\" (UID: \"18d8d58b-3dab-4949-b9d1-313870a96dd5\") " pod="openshift-console-operator/console-operator-58897d9998-dfltg" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.188540 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d7952588-d49d-41c1-b5ec-60e6ea4a1efa-serving-cert\") pod \"etcd-operator-b45778765-mnn4f\" (UID: \"d7952588-d49d-41c1-b5ec-60e6ea4a1efa\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mnn4f" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.188559 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/911a0a7e-7a63-4b4a-8b74-6c21f8da7e31-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-d8pkd\" (UID: \"911a0a7e-7a63-4b4a-8b74-6c21f8da7e31\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-d8pkd" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.188582 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-snt8h\" (UniqueName: \"kubernetes.io/projected/cc2bd33e-fa09-4675-9d48-b546668b5b86-kube-api-access-snt8h\") pod \"openshift-controller-manager-operator-756b6f6bc6-gxw89\" (UID: \"cc2bd33e-fa09-4675-9d48-b546668b5b86\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gxw89" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.188598 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/107352c2-4357-43c2-9958-cb8a91714fa6-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-gspw5\" (UID: \"107352c2-4357-43c2-9958-cb8a91714fa6\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-gspw5" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.188616 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/911a0a7e-7a63-4b4a-8b74-6c21f8da7e31-etcd-client\") pod \"apiserver-7bbb656c7d-d8pkd\" (UID: \"911a0a7e-7a63-4b4a-8b74-6c21f8da7e31\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-d8pkd" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.188633 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0e4b5f84-4954-4d85-8bae-db18ed352464-audit-policies\") pod \"oauth-openshift-558db77b4-2trmp\" (UID: \"0e4b5f84-4954-4d85-8bae-db18ed352464\") " pod="openshift-authentication/oauth-openshift-558db77b4-2trmp" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.188660 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21295e15-1d49-4261-9970-fb61c1d4854c-config\") pod \"route-controller-manager-6576b87f9c-lclf5\" (UID: \"21295e15-1d49-4261-9970-fb61c1d4854c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lclf5" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.188677 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/69303f47-f9a7-4c1c-b569-b1e4de56b2ff-images\") pod \"machine-api-operator-5694c8668f-hk7mf\" (UID: \"69303f47-f9a7-4c1c-b569-b1e4de56b2ff\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-hk7mf" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.188691 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/0e4b5f84-4954-4d85-8bae-db18ed352464-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-2trmp\" (UID: \"0e4b5f84-4954-4d85-8bae-db18ed352464\") " pod="openshift-authentication/oauth-openshift-558db77b4-2trmp" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.188707 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/0e4b5f84-4954-4d85-8bae-db18ed352464-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-2trmp\" (UID: \"0e4b5f84-4954-4d85-8bae-db18ed352464\") " pod="openshift-authentication/oauth-openshift-558db77b4-2trmp" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.188732 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/0e4b5f84-4954-4d85-8bae-db18ed352464-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-2trmp\" (UID: \"0e4b5f84-4954-4d85-8bae-db18ed352464\") " pod="openshift-authentication/oauth-openshift-558db77b4-2trmp" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.188748 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mtpkb\" (UniqueName: \"kubernetes.io/projected/21295e15-1d49-4261-9970-fb61c1d4854c-kube-api-access-mtpkb\") pod \"route-controller-manager-6576b87f9c-lclf5\" (UID: \"21295e15-1d49-4261-9970-fb61c1d4854c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lclf5" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.188764 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/0e4b5f84-4954-4d85-8bae-db18ed352464-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-2trmp\" (UID: \"0e4b5f84-4954-4d85-8bae-db18ed352464\") " pod="openshift-authentication/oauth-openshift-558db77b4-2trmp" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.188782 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5qm7h\" (UniqueName: \"kubernetes.io/projected/18d8d58b-3dab-4949-b9d1-313870a96dd5-kube-api-access-5qm7h\") pod \"console-operator-58897d9998-dfltg\" (UID: \"18d8d58b-3dab-4949-b9d1-313870a96dd5\") " pod="openshift-console-operator/console-operator-58897d9998-dfltg" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.188796 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/21295e15-1d49-4261-9970-fb61c1d4854c-serving-cert\") pod \"route-controller-manager-6576b87f9c-lclf5\" (UID: \"21295e15-1d49-4261-9970-fb61c1d4854c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lclf5" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.188814 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/26d0d35d-8516-46dd-8094-d72c27cfe267-machine-approver-tls\") pod \"machine-approver-56656f9798-g5z94\" (UID: \"26d0d35d-8516-46dd-8094-d72c27cfe267\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-g5z94" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.188829 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/911a0a7e-7a63-4b4a-8b74-6c21f8da7e31-serving-cert\") pod \"apiserver-7bbb656c7d-d8pkd\" (UID: \"911a0a7e-7a63-4b4a-8b74-6c21f8da7e31\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-d8pkd" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.188848 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc2bd33e-fa09-4675-9d48-b546668b5b86-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-gxw89\" (UID: \"cc2bd33e-fa09-4675-9d48-b546668b5b86\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gxw89" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.188868 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/0e4b5f84-4954-4d85-8bae-db18ed352464-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-2trmp\" (UID: \"0e4b5f84-4954-4d85-8bae-db18ed352464\") " pod="openshift-authentication/oauth-openshift-558db77b4-2trmp" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.188905 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/69303f47-f9a7-4c1c-b569-b1e4de56b2ff-config\") pod \"machine-api-operator-5694c8668f-hk7mf\" (UID: \"69303f47-f9a7-4c1c-b569-b1e4de56b2ff\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-hk7mf" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.188920 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/911a0a7e-7a63-4b4a-8b74-6c21f8da7e31-audit-policies\") pod \"apiserver-7bbb656c7d-d8pkd\" (UID: \"911a0a7e-7a63-4b4a-8b74-6c21f8da7e31\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-d8pkd" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.188936 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/0e4b5f84-4954-4d85-8bae-db18ed352464-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-2trmp\" (UID: \"0e4b5f84-4954-4d85-8bae-db18ed352464\") " pod="openshift-authentication/oauth-openshift-558db77b4-2trmp" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.188953 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bw5l7\" (UniqueName: \"kubernetes.io/projected/26d0d35d-8516-46dd-8094-d72c27cfe267-kube-api-access-bw5l7\") pod \"machine-approver-56656f9798-g5z94\" (UID: \"26d0d35d-8516-46dd-8094-d72c27cfe267\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-g5z94" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.188972 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/911a0a7e-7a63-4b4a-8b74-6c21f8da7e31-audit-dir\") pod \"apiserver-7bbb656c7d-d8pkd\" (UID: \"911a0a7e-7a63-4b4a-8b74-6c21f8da7e31\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-d8pkd" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.188988 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jw64h\" (UniqueName: \"kubernetes.io/projected/911a0a7e-7a63-4b4a-8b74-6c21f8da7e31-kube-api-access-jw64h\") pod \"apiserver-7bbb656c7d-d8pkd\" (UID: \"911a0a7e-7a63-4b4a-8b74-6c21f8da7e31\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-d8pkd" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.189005 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/0e4b5f84-4954-4d85-8bae-db18ed352464-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-2trmp\" (UID: \"0e4b5f84-4954-4d85-8bae-db18ed352464\") " pod="openshift-authentication/oauth-openshift-558db77b4-2trmp" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.189021 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/d7952588-d49d-41c1-b5ec-60e6ea4a1efa-etcd-ca\") pod \"etcd-operator-b45778765-mnn4f\" (UID: \"d7952588-d49d-41c1-b5ec-60e6ea4a1efa\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mnn4f" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.189040 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6g94g\" (UniqueName: \"kubernetes.io/projected/d7952588-d49d-41c1-b5ec-60e6ea4a1efa-kube-api-access-6g94g\") pod \"etcd-operator-b45778765-mnn4f\" (UID: \"d7952588-d49d-41c1-b5ec-60e6ea4a1efa\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mnn4f" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.189058 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/0e4b5f84-4954-4d85-8bae-db18ed352464-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-2trmp\" (UID: \"0e4b5f84-4954-4d85-8bae-db18ed352464\") " pod="openshift-authentication/oauth-openshift-558db77b4-2trmp" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.189075 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26d0d35d-8516-46dd-8094-d72c27cfe267-config\") pod \"machine-approver-56656f9798-g5z94\" (UID: \"26d0d35d-8516-46dd-8094-d72c27cfe267\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-g5z94" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.189090 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/92733842-6b61-4327-b35c-e37cc833f7a3-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-p85cq\" (UID: \"92733842-6b61-4327-b35c-e37cc833f7a3\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p85cq" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.189108 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4bf1a18e-ec09-4f71-b6a1-645698067dc1-client-ca\") pod \"controller-manager-879f6c89f-rnnlh\" (UID: \"4bf1a18e-ec09-4f71-b6a1-645698067dc1\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rnnlh" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.189132 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qbtbc\" (UniqueName: \"kubernetes.io/projected/107352c2-4357-43c2-9958-cb8a91714fa6-kube-api-access-qbtbc\") pod \"cluster-samples-operator-665b6dd947-gspw5\" (UID: \"107352c2-4357-43c2-9958-cb8a91714fa6\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-gspw5" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.189149 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a309e9aa-bf13-404b-90dd-8ddc81467822-metrics-tls\") pod \"dns-operator-744455d44c-pgf6s\" (UID: \"a309e9aa-bf13-404b-90dd-8ddc81467822\") " pod="openshift-dns-operator/dns-operator-744455d44c-pgf6s" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.189166 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7952588-d49d-41c1-b5ec-60e6ea4a1efa-config\") pod \"etcd-operator-b45778765-mnn4f\" (UID: \"d7952588-d49d-41c1-b5ec-60e6ea4a1efa\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mnn4f" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.189184 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4bf1a18e-ec09-4f71-b6a1-645698067dc1-config\") pod \"controller-manager-879f6c89f-rnnlh\" (UID: \"4bf1a18e-ec09-4f71-b6a1-645698067dc1\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rnnlh" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.189202 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0e4b5f84-4954-4d85-8bae-db18ed352464-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-2trmp\" (UID: \"0e4b5f84-4954-4d85-8bae-db18ed352464\") " pod="openshift-authentication/oauth-openshift-558db77b4-2trmp" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.189217 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmhkr\" (UniqueName: \"kubernetes.io/projected/0e4b5f84-4954-4d85-8bae-db18ed352464-kube-api-access-vmhkr\") pod \"oauth-openshift-558db77b4-2trmp\" (UID: \"0e4b5f84-4954-4d85-8bae-db18ed352464\") " pod="openshift-authentication/oauth-openshift-558db77b4-2trmp" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.189232 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4bf1a18e-ec09-4f71-b6a1-645698067dc1-serving-cert\") pod \"controller-manager-879f6c89f-rnnlh\" (UID: \"4bf1a18e-ec09-4f71-b6a1-645698067dc1\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rnnlh" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.189250 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/92733842-6b61-4327-b35c-e37cc833f7a3-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-p85cq\" (UID: \"92733842-6b61-4327-b35c-e37cc833f7a3\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p85cq" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.189265 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/0e4b5f84-4954-4d85-8bae-db18ed352464-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-2trmp\" (UID: \"0e4b5f84-4954-4d85-8bae-db18ed352464\") " pod="openshift-authentication/oauth-openshift-558db77b4-2trmp" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.189289 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18d8d58b-3dab-4949-b9d1-313870a96dd5-config\") pod \"console-operator-58897d9998-dfltg\" (UID: \"18d8d58b-3dab-4949-b9d1-313870a96dd5\") " pod="openshift-console-operator/console-operator-58897d9998-dfltg" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.189304 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d7952588-d49d-41c1-b5ec-60e6ea4a1efa-etcd-client\") pod \"etcd-operator-b45778765-mnn4f\" (UID: \"d7952588-d49d-41c1-b5ec-60e6ea4a1efa\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mnn4f" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.189319 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/911a0a7e-7a63-4b4a-8b74-6c21f8da7e31-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-d8pkd\" (UID: \"911a0a7e-7a63-4b4a-8b74-6c21f8da7e31\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-d8pkd" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.189334 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4bf1a18e-ec09-4f71-b6a1-645698067dc1-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-rnnlh\" (UID: \"4bf1a18e-ec09-4f71-b6a1-645698067dc1\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rnnlh" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.189350 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cc2bd33e-fa09-4675-9d48-b546668b5b86-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-gxw89\" (UID: \"cc2bd33e-fa09-4675-9d48-b546668b5b86\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gxw89" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.189365 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/92733842-6b61-4327-b35c-e37cc833f7a3-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-p85cq\" (UID: \"92733842-6b61-4327-b35c-e37cc833f7a3\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p85cq" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.189378 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/911a0a7e-7a63-4b4a-8b74-6c21f8da7e31-encryption-config\") pod \"apiserver-7bbb656c7d-d8pkd\" (UID: \"911a0a7e-7a63-4b4a-8b74-6c21f8da7e31\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-d8pkd" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.189394 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0e4b5f84-4954-4d85-8bae-db18ed352464-audit-dir\") pod \"oauth-openshift-558db77b4-2trmp\" (UID: \"0e4b5f84-4954-4d85-8bae-db18ed352464\") " pod="openshift-authentication/oauth-openshift-558db77b4-2trmp" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.189409 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/18d8d58b-3dab-4949-b9d1-313870a96dd5-trusted-ca\") pod \"console-operator-58897d9998-dfltg\" (UID: \"18d8d58b-3dab-4949-b9d1-313870a96dd5\") " pod="openshift-console-operator/console-operator-58897d9998-dfltg" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.189422 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/26d0d35d-8516-46dd-8094-d72c27cfe267-auth-proxy-config\") pod \"machine-approver-56656f9798-g5z94\" (UID: \"26d0d35d-8516-46dd-8094-d72c27cfe267\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-g5z94" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.189437 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/d7952588-d49d-41c1-b5ec-60e6ea4a1efa-etcd-service-ca\") pod \"etcd-operator-b45778765-mnn4f\" (UID: \"d7952588-d49d-41c1-b5ec-60e6ea4a1efa\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mnn4f" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.189453 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xjzmb\" (UniqueName: \"kubernetes.io/projected/a309e9aa-bf13-404b-90dd-8ddc81467822-kube-api-access-xjzmb\") pod \"dns-operator-744455d44c-pgf6s\" (UID: \"a309e9aa-bf13-404b-90dd-8ddc81467822\") " pod="openshift-dns-operator/dns-operator-744455d44c-pgf6s" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.189468 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/21295e15-1d49-4261-9970-fb61c1d4854c-client-ca\") pod \"route-controller-manager-6576b87f9c-lclf5\" (UID: \"21295e15-1d49-4261-9970-fb61c1d4854c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lclf5" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.189481 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-68nxr\" (UniqueName: \"kubernetes.io/projected/69303f47-f9a7-4c1c-b569-b1e4de56b2ff-kube-api-access-68nxr\") pod \"machine-api-operator-5694c8668f-hk7mf\" (UID: \"69303f47-f9a7-4c1c-b569-b1e4de56b2ff\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-hk7mf" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.192539 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.194925 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5n549"] Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.196232 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.196583 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.196822 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.197011 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.197747 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.198189 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5n549" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.198640 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-5z687"] Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.199086 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.201586 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.201680 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.220294 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-szznp"] Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.223096 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-szznp" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.223551 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-lp8th"] Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.265995 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-lp8th" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.269342 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-2hmqv"] Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.269686 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.270099 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2hmqv" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.270691 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.272727 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-2kgg2"] Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.273122 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-8s4px"] Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.273578 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.274054 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.277053 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.281568 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4cx2q"] Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.281908 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-v8bk6"] Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.281949 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8s4px" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.282155 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5prl8"] Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.282474 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5prl8" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.283186 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.283396 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.283616 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.283723 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.283821 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.284429 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h59mj"] Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.288379 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-2kgg2" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.288855 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4cx2q" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.288917 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-v8bk6" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.290048 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-74phw"] Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.290569 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-cc48c"] Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.291119 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-b9s66"] Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.291591 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h59mj" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.291608 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-74phw" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.291708 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.291729 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.291965 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.292045 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.292123 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.292208 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.292234 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.292346 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.292411 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-b9s66" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.292597 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-cc48c" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.292886 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.292921 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.293032 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.293057 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.293247 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.293340 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.293375 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.293444 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.293493 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.293749 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.293947 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.294081 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.294725 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/92733842-6b61-4327-b35c-e37cc833f7a3-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-p85cq\" (UID: \"92733842-6b61-4327-b35c-e37cc833f7a3\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p85cq" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.294775 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/0e4b5f84-4954-4d85-8bae-db18ed352464-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-2trmp\" (UID: \"0e4b5f84-4954-4d85-8bae-db18ed352464\") " pod="openshift-authentication/oauth-openshift-558db77b4-2trmp" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.294798 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/74ff4ad0-9479-4fa2-8df9-501c332e386c-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-5n549\" (UID: \"74ff4ad0-9479-4fa2-8df9-501c332e386c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5n549" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.294843 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1fedc911-62a7-496f-93d1-f09769b89e86-config\") pod \"openshift-apiserver-operator-796bbdcf4f-rv2nz\" (UID: \"1fedc911-62a7-496f-93d1-f09769b89e86\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rv2nz" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.294863 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/911a0a7e-7a63-4b4a-8b74-6c21f8da7e31-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-d8pkd\" (UID: \"911a0a7e-7a63-4b4a-8b74-6c21f8da7e31\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-d8pkd" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.294880 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4bf1a18e-ec09-4f71-b6a1-645698067dc1-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-rnnlh\" (UID: \"4bf1a18e-ec09-4f71-b6a1-645698067dc1\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rnnlh" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.294896 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1fedc911-62a7-496f-93d1-f09769b89e86-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-rv2nz\" (UID: \"1fedc911-62a7-496f-93d1-f09769b89e86\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rv2nz" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.294934 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18d8d58b-3dab-4949-b9d1-313870a96dd5-config\") pod \"console-operator-58897d9998-dfltg\" (UID: \"18d8d58b-3dab-4949-b9d1-313870a96dd5\") " pod="openshift-console-operator/console-operator-58897d9998-dfltg" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.294952 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d7952588-d49d-41c1-b5ec-60e6ea4a1efa-etcd-client\") pod \"etcd-operator-b45778765-mnn4f\" (UID: \"d7952588-d49d-41c1-b5ec-60e6ea4a1efa\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mnn4f" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.294967 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cc2bd33e-fa09-4675-9d48-b546668b5b86-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-gxw89\" (UID: \"cc2bd33e-fa09-4675-9d48-b546668b5b86\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gxw89" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.295004 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/92733842-6b61-4327-b35c-e37cc833f7a3-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-p85cq\" (UID: \"92733842-6b61-4327-b35c-e37cc833f7a3\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p85cq" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.295021 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/911a0a7e-7a63-4b4a-8b74-6c21f8da7e31-encryption-config\") pod \"apiserver-7bbb656c7d-d8pkd\" (UID: \"911a0a7e-7a63-4b4a-8b74-6c21f8da7e31\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-d8pkd" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.295037 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0e4b5f84-4954-4d85-8bae-db18ed352464-audit-dir\") pod \"oauth-openshift-558db77b4-2trmp\" (UID: \"0e4b5f84-4954-4d85-8bae-db18ed352464\") " pod="openshift-authentication/oauth-openshift-558db77b4-2trmp" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.295052 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/18d8d58b-3dab-4949-b9d1-313870a96dd5-trusted-ca\") pod \"console-operator-58897d9998-dfltg\" (UID: \"18d8d58b-3dab-4949-b9d1-313870a96dd5\") " pod="openshift-console-operator/console-operator-58897d9998-dfltg" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.295066 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/26d0d35d-8516-46dd-8094-d72c27cfe267-auth-proxy-config\") pod \"machine-approver-56656f9798-g5z94\" (UID: \"26d0d35d-8516-46dd-8094-d72c27cfe267\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-g5z94" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.295080 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/d7952588-d49d-41c1-b5ec-60e6ea4a1efa-etcd-service-ca\") pod \"etcd-operator-b45778765-mnn4f\" (UID: \"d7952588-d49d-41c1-b5ec-60e6ea4a1efa\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mnn4f" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.295097 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xjzmb\" (UniqueName: \"kubernetes.io/projected/a309e9aa-bf13-404b-90dd-8ddc81467822-kube-api-access-xjzmb\") pod \"dns-operator-744455d44c-pgf6s\" (UID: \"a309e9aa-bf13-404b-90dd-8ddc81467822\") " pod="openshift-dns-operator/dns-operator-744455d44c-pgf6s" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.295112 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/21295e15-1d49-4261-9970-fb61c1d4854c-client-ca\") pod \"route-controller-manager-6576b87f9c-lclf5\" (UID: \"21295e15-1d49-4261-9970-fb61c1d4854c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lclf5" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.295128 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-68nxr\" (UniqueName: \"kubernetes.io/projected/69303f47-f9a7-4c1c-b569-b1e4de56b2ff-kube-api-access-68nxr\") pod \"machine-api-operator-5694c8668f-hk7mf\" (UID: \"69303f47-f9a7-4c1c-b569-b1e4de56b2ff\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-hk7mf" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.295143 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dpwxx\" (UniqueName: \"kubernetes.io/projected/92733842-6b61-4327-b35c-e37cc833f7a3-kube-api-access-dpwxx\") pod \"cluster-image-registry-operator-dc59b4c8b-p85cq\" (UID: \"92733842-6b61-4327-b35c-e37cc833f7a3\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p85cq" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.295164 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/69303f47-f9a7-4c1c-b569-b1e4de56b2ff-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-hk7mf\" (UID: \"69303f47-f9a7-4c1c-b569-b1e4de56b2ff\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-hk7mf" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.295180 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n9k72\" (UniqueName: \"kubernetes.io/projected/4bf1a18e-ec09-4f71-b6a1-645698067dc1-kube-api-access-n9k72\") pod \"controller-manager-879f6c89f-rnnlh\" (UID: \"4bf1a18e-ec09-4f71-b6a1-645698067dc1\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rnnlh" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.295203 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/18d8d58b-3dab-4949-b9d1-313870a96dd5-serving-cert\") pod \"console-operator-58897d9998-dfltg\" (UID: \"18d8d58b-3dab-4949-b9d1-313870a96dd5\") " pod="openshift-console-operator/console-operator-58897d9998-dfltg" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.295219 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d7952588-d49d-41c1-b5ec-60e6ea4a1efa-serving-cert\") pod \"etcd-operator-b45778765-mnn4f\" (UID: \"d7952588-d49d-41c1-b5ec-60e6ea4a1efa\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mnn4f" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.295233 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/911a0a7e-7a63-4b4a-8b74-6c21f8da7e31-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-d8pkd\" (UID: \"911a0a7e-7a63-4b4a-8b74-6c21f8da7e31\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-d8pkd" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.295248 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/0e4b5f84-4954-4d85-8bae-db18ed352464-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-2trmp\" (UID: \"0e4b5f84-4954-4d85-8bae-db18ed352464\") " pod="openshift-authentication/oauth-openshift-558db77b4-2trmp" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.295267 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-snt8h\" (UniqueName: \"kubernetes.io/projected/cc2bd33e-fa09-4675-9d48-b546668b5b86-kube-api-access-snt8h\") pod \"openshift-controller-manager-operator-756b6f6bc6-gxw89\" (UID: \"cc2bd33e-fa09-4675-9d48-b546668b5b86\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gxw89" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.295283 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/107352c2-4357-43c2-9958-cb8a91714fa6-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-gspw5\" (UID: \"107352c2-4357-43c2-9958-cb8a91714fa6\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-gspw5" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.295297 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/911a0a7e-7a63-4b4a-8b74-6c21f8da7e31-etcd-client\") pod \"apiserver-7bbb656c7d-d8pkd\" (UID: \"911a0a7e-7a63-4b4a-8b74-6c21f8da7e31\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-d8pkd" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.295313 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0e4b5f84-4954-4d85-8bae-db18ed352464-audit-policies\") pod \"oauth-openshift-558db77b4-2trmp\" (UID: \"0e4b5f84-4954-4d85-8bae-db18ed352464\") " pod="openshift-authentication/oauth-openshift-558db77b4-2trmp" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.295327 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21295e15-1d49-4261-9970-fb61c1d4854c-config\") pod \"route-controller-manager-6576b87f9c-lclf5\" (UID: \"21295e15-1d49-4261-9970-fb61c1d4854c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lclf5" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.295343 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/69303f47-f9a7-4c1c-b569-b1e4de56b2ff-images\") pod \"machine-api-operator-5694c8668f-hk7mf\" (UID: \"69303f47-f9a7-4c1c-b569-b1e4de56b2ff\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-hk7mf" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.295405 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a59737d6-7609-4077-8fb8-d925f25a65f8-serving-cert\") pod \"authentication-operator-69f744f599-8kddm\" (UID: \"a59737d6-7609-4077-8fb8-d925f25a65f8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8kddm" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.295466 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/0e4b5f84-4954-4d85-8bae-db18ed352464-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-2trmp\" (UID: \"0e4b5f84-4954-4d85-8bae-db18ed352464\") " pod="openshift-authentication/oauth-openshift-558db77b4-2trmp" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.295486 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/0e4b5f84-4954-4d85-8bae-db18ed352464-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-2trmp\" (UID: \"0e4b5f84-4954-4d85-8bae-db18ed352464\") " pod="openshift-authentication/oauth-openshift-558db77b4-2trmp" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.295533 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/0e4b5f84-4954-4d85-8bae-db18ed352464-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-2trmp\" (UID: \"0e4b5f84-4954-4d85-8bae-db18ed352464\") " pod="openshift-authentication/oauth-openshift-558db77b4-2trmp" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.295559 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mtpkb\" (UniqueName: \"kubernetes.io/projected/21295e15-1d49-4261-9970-fb61c1d4854c-kube-api-access-mtpkb\") pod \"route-controller-manager-6576b87f9c-lclf5\" (UID: \"21295e15-1d49-4261-9970-fb61c1d4854c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lclf5" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.295587 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/0e4b5f84-4954-4d85-8bae-db18ed352464-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-2trmp\" (UID: \"0e4b5f84-4954-4d85-8bae-db18ed352464\") " pod="openshift-authentication/oauth-openshift-558db77b4-2trmp" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.295610 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a59737d6-7609-4077-8fb8-d925f25a65f8-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-8kddm\" (UID: \"a59737d6-7609-4077-8fb8-d925f25a65f8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8kddm" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.295627 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5qm7h\" (UniqueName: \"kubernetes.io/projected/18d8d58b-3dab-4949-b9d1-313870a96dd5-kube-api-access-5qm7h\") pod \"console-operator-58897d9998-dfltg\" (UID: \"18d8d58b-3dab-4949-b9d1-313870a96dd5\") " pod="openshift-console-operator/console-operator-58897d9998-dfltg" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.295641 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/21295e15-1d49-4261-9970-fb61c1d4854c-serving-cert\") pod \"route-controller-manager-6576b87f9c-lclf5\" (UID: \"21295e15-1d49-4261-9970-fb61c1d4854c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lclf5" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.295656 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/911a0a7e-7a63-4b4a-8b74-6c21f8da7e31-serving-cert\") pod \"apiserver-7bbb656c7d-d8pkd\" (UID: \"911a0a7e-7a63-4b4a-8b74-6c21f8da7e31\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-d8pkd" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.295674 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-skszv\" (UniqueName: \"kubernetes.io/projected/cfacdafa-2513-4d81-9222-c4b4fd8f8c10-kube-api-access-skszv\") pod \"migrator-59844c95c7-8s4px\" (UID: \"cfacdafa-2513-4d81-9222-c4b4fd8f8c10\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8s4px" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.295691 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/26d0d35d-8516-46dd-8094-d72c27cfe267-machine-approver-tls\") pod \"machine-approver-56656f9798-g5z94\" (UID: \"26d0d35d-8516-46dd-8094-d72c27cfe267\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-g5z94" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.295710 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/69303f47-f9a7-4c1c-b569-b1e4de56b2ff-config\") pod \"machine-api-operator-5694c8668f-hk7mf\" (UID: \"69303f47-f9a7-4c1c-b569-b1e4de56b2ff\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-hk7mf" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.295729 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/911a0a7e-7a63-4b4a-8b74-6c21f8da7e31-audit-policies\") pod \"apiserver-7bbb656c7d-d8pkd\" (UID: \"911a0a7e-7a63-4b4a-8b74-6c21f8da7e31\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-d8pkd" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.295745 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc2bd33e-fa09-4675-9d48-b546668b5b86-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-gxw89\" (UID: \"cc2bd33e-fa09-4675-9d48-b546668b5b86\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gxw89" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.295762 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/0e4b5f84-4954-4d85-8bae-db18ed352464-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-2trmp\" (UID: \"0e4b5f84-4954-4d85-8bae-db18ed352464\") " pod="openshift-authentication/oauth-openshift-558db77b4-2trmp" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.295778 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bw5l7\" (UniqueName: \"kubernetes.io/projected/26d0d35d-8516-46dd-8094-d72c27cfe267-kube-api-access-bw5l7\") pod \"machine-approver-56656f9798-g5z94\" (UID: \"26d0d35d-8516-46dd-8094-d72c27cfe267\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-g5z94" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.295793 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/911a0a7e-7a63-4b4a-8b74-6c21f8da7e31-audit-dir\") pod \"apiserver-7bbb656c7d-d8pkd\" (UID: \"911a0a7e-7a63-4b4a-8b74-6c21f8da7e31\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-d8pkd" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.295810 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/0e4b5f84-4954-4d85-8bae-db18ed352464-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-2trmp\" (UID: \"0e4b5f84-4954-4d85-8bae-db18ed352464\") " pod="openshift-authentication/oauth-openshift-558db77b4-2trmp" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.295828 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jw64h\" (UniqueName: \"kubernetes.io/projected/911a0a7e-7a63-4b4a-8b74-6c21f8da7e31-kube-api-access-jw64h\") pod \"apiserver-7bbb656c7d-d8pkd\" (UID: \"911a0a7e-7a63-4b4a-8b74-6c21f8da7e31\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-d8pkd" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.295844 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/0e4b5f84-4954-4d85-8bae-db18ed352464-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-2trmp\" (UID: \"0e4b5f84-4954-4d85-8bae-db18ed352464\") " pod="openshift-authentication/oauth-openshift-558db77b4-2trmp" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.295863 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/d7952588-d49d-41c1-b5ec-60e6ea4a1efa-etcd-ca\") pod \"etcd-operator-b45778765-mnn4f\" (UID: \"d7952588-d49d-41c1-b5ec-60e6ea4a1efa\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mnn4f" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.295879 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6g94g\" (UniqueName: \"kubernetes.io/projected/d7952588-d49d-41c1-b5ec-60e6ea4a1efa-kube-api-access-6g94g\") pod \"etcd-operator-b45778765-mnn4f\" (UID: \"d7952588-d49d-41c1-b5ec-60e6ea4a1efa\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mnn4f" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.295893 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/0e4b5f84-4954-4d85-8bae-db18ed352464-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-2trmp\" (UID: \"0e4b5f84-4954-4d85-8bae-db18ed352464\") " pod="openshift-authentication/oauth-openshift-558db77b4-2trmp" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.295909 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26d0d35d-8516-46dd-8094-d72c27cfe267-config\") pod \"machine-approver-56656f9798-g5z94\" (UID: \"26d0d35d-8516-46dd-8094-d72c27cfe267\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-g5z94" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.295924 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/74ff4ad0-9479-4fa2-8df9-501c332e386c-config\") pod \"kube-apiserver-operator-766d6c64bb-5n549\" (UID: \"74ff4ad0-9479-4fa2-8df9-501c332e386c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5n549" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.295944 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a59737d6-7609-4077-8fb8-d925f25a65f8-config\") pod \"authentication-operator-69f744f599-8kddm\" (UID: \"a59737d6-7609-4077-8fb8-d925f25a65f8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8kddm" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.295962 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rtj8k\" (UniqueName: \"kubernetes.io/projected/a59737d6-7609-4077-8fb8-d925f25a65f8-kube-api-access-rtj8k\") pod \"authentication-operator-69f744f599-8kddm\" (UID: \"a59737d6-7609-4077-8fb8-d925f25a65f8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8kddm" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.295979 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2ldbq\" (UniqueName: \"kubernetes.io/projected/1fedc911-62a7-496f-93d1-f09769b89e86-kube-api-access-2ldbq\") pod \"openshift-apiserver-operator-796bbdcf4f-rv2nz\" (UID: \"1fedc911-62a7-496f-93d1-f09769b89e86\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rv2nz" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.295995 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/92733842-6b61-4327-b35c-e37cc833f7a3-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-p85cq\" (UID: \"92733842-6b61-4327-b35c-e37cc833f7a3\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p85cq" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.296011 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/74ff4ad0-9479-4fa2-8df9-501c332e386c-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-5n549\" (UID: \"74ff4ad0-9479-4fa2-8df9-501c332e386c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5n549" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.296028 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4bf1a18e-ec09-4f71-b6a1-645698067dc1-client-ca\") pod \"controller-manager-879f6c89f-rnnlh\" (UID: \"4bf1a18e-ec09-4f71-b6a1-645698067dc1\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rnnlh" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.296057 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qbtbc\" (UniqueName: \"kubernetes.io/projected/107352c2-4357-43c2-9958-cb8a91714fa6-kube-api-access-qbtbc\") pod \"cluster-samples-operator-665b6dd947-gspw5\" (UID: \"107352c2-4357-43c2-9958-cb8a91714fa6\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-gspw5" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.296074 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a309e9aa-bf13-404b-90dd-8ddc81467822-metrics-tls\") pod \"dns-operator-744455d44c-pgf6s\" (UID: \"a309e9aa-bf13-404b-90dd-8ddc81467822\") " pod="openshift-dns-operator/dns-operator-744455d44c-pgf6s" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.296092 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7952588-d49d-41c1-b5ec-60e6ea4a1efa-config\") pod \"etcd-operator-b45778765-mnn4f\" (UID: \"d7952588-d49d-41c1-b5ec-60e6ea4a1efa\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mnn4f" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.296109 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4bf1a18e-ec09-4f71-b6a1-645698067dc1-config\") pod \"controller-manager-879f6c89f-rnnlh\" (UID: \"4bf1a18e-ec09-4f71-b6a1-645698067dc1\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rnnlh" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.296125 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0e4b5f84-4954-4d85-8bae-db18ed352464-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-2trmp\" (UID: \"0e4b5f84-4954-4d85-8bae-db18ed352464\") " pod="openshift-authentication/oauth-openshift-558db77b4-2trmp" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.296176 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmhkr\" (UniqueName: \"kubernetes.io/projected/0e4b5f84-4954-4d85-8bae-db18ed352464-kube-api-access-vmhkr\") pod \"oauth-openshift-558db77b4-2trmp\" (UID: \"0e4b5f84-4954-4d85-8bae-db18ed352464\") " pod="openshift-authentication/oauth-openshift-558db77b4-2trmp" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.296196 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4bf1a18e-ec09-4f71-b6a1-645698067dc1-serving-cert\") pod \"controller-manager-879f6c89f-rnnlh\" (UID: \"4bf1a18e-ec09-4f71-b6a1-645698067dc1\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rnnlh" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.296242 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a59737d6-7609-4077-8fb8-d925f25a65f8-service-ca-bundle\") pod \"authentication-operator-69f744f599-8kddm\" (UID: \"a59737d6-7609-4077-8fb8-d925f25a65f8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8kddm" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.298217 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.298317 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/92733842-6b61-4327-b35c-e37cc833f7a3-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-p85cq\" (UID: \"92733842-6b61-4327-b35c-e37cc833f7a3\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p85cq" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.298810 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-vf46x"] Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.299421 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-vf46x" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.301749 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0e4b5f84-4954-4d85-8bae-db18ed352464-audit-dir\") pod \"oauth-openshift-558db77b4-2trmp\" (UID: \"0e4b5f84-4954-4d85-8bae-db18ed352464\") " pod="openshift-authentication/oauth-openshift-558db77b4-2trmp" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.302741 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/26d0d35d-8516-46dd-8094-d72c27cfe267-auth-proxy-config\") pod \"machine-approver-56656f9798-g5z94\" (UID: \"26d0d35d-8516-46dd-8094-d72c27cfe267\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-g5z94" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.302982 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/911a0a7e-7a63-4b4a-8b74-6c21f8da7e31-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-d8pkd\" (UID: \"911a0a7e-7a63-4b4a-8b74-6c21f8da7e31\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-d8pkd" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.303573 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.304811 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6sfv4"] Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.304837 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/911a0a7e-7a63-4b4a-8b74-6c21f8da7e31-audit-policies\") pod \"apiserver-7bbb656c7d-d8pkd\" (UID: \"911a0a7e-7a63-4b4a-8b74-6c21f8da7e31\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-d8pkd" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.305369 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6sfv4" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.308716 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/911a0a7e-7a63-4b4a-8b74-6c21f8da7e31-audit-dir\") pod \"apiserver-7bbb656c7d-d8pkd\" (UID: \"911a0a7e-7a63-4b4a-8b74-6c21f8da7e31\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-d8pkd" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.305493 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/69303f47-f9a7-4c1c-b569-b1e4de56b2ff-config\") pod \"machine-api-operator-5694c8668f-hk7mf\" (UID: \"69303f47-f9a7-4c1c-b569-b1e4de56b2ff\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-hk7mf" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.317112 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.317327 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/0e4b5f84-4954-4d85-8bae-db18ed352464-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-2trmp\" (UID: \"0e4b5f84-4954-4d85-8bae-db18ed352464\") " pod="openshift-authentication/oauth-openshift-558db77b4-2trmp" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.317951 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26d0d35d-8516-46dd-8094-d72c27cfe267-config\") pod \"machine-approver-56656f9798-g5z94\" (UID: \"26d0d35d-8516-46dd-8094-d72c27cfe267\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-g5z94" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.318022 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cc2bd33e-fa09-4675-9d48-b546668b5b86-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-gxw89\" (UID: \"cc2bd33e-fa09-4675-9d48-b546668b5b86\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gxw89" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.318294 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4bf1a18e-ec09-4f71-b6a1-645698067dc1-client-ca\") pod \"controller-manager-879f6c89f-rnnlh\" (UID: \"4bf1a18e-ec09-4f71-b6a1-645698067dc1\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rnnlh" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.318809 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc2bd33e-fa09-4675-9d48-b546668b5b86-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-gxw89\" (UID: \"cc2bd33e-fa09-4675-9d48-b546668b5b86\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gxw89" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.319337 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/d7952588-d49d-41c1-b5ec-60e6ea4a1efa-etcd-ca\") pod \"etcd-operator-b45778765-mnn4f\" (UID: \"d7952588-d49d-41c1-b5ec-60e6ea4a1efa\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mnn4f" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.319809 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/21295e15-1d49-4261-9970-fb61c1d4854c-serving-cert\") pod \"route-controller-manager-6576b87f9c-lclf5\" (UID: \"21295e15-1d49-4261-9970-fb61c1d4854c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lclf5" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.320522 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4bf1a18e-ec09-4f71-b6a1-645698067dc1-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-rnnlh\" (UID: \"4bf1a18e-ec09-4f71-b6a1-645698067dc1\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rnnlh" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.321216 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18d8d58b-3dab-4949-b9d1-313870a96dd5-config\") pod \"console-operator-58897d9998-dfltg\" (UID: \"18d8d58b-3dab-4949-b9d1-313870a96dd5\") " pod="openshift-console-operator/console-operator-58897d9998-dfltg" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.346714 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/18d8d58b-3dab-4949-b9d1-313870a96dd5-trusted-ca\") pod \"console-operator-58897d9998-dfltg\" (UID: \"18d8d58b-3dab-4949-b9d1-313870a96dd5\") " pod="openshift-console-operator/console-operator-58897d9998-dfltg" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.347035 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/0e4b5f84-4954-4d85-8bae-db18ed352464-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-2trmp\" (UID: \"0e4b5f84-4954-4d85-8bae-db18ed352464\") " pod="openshift-authentication/oauth-openshift-558db77b4-2trmp" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.347217 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d7952588-d49d-41c1-b5ec-60e6ea4a1efa-etcd-client\") pod \"etcd-operator-b45778765-mnn4f\" (UID: \"d7952588-d49d-41c1-b5ec-60e6ea4a1efa\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mnn4f" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.347599 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/21295e15-1d49-4261-9970-fb61c1d4854c-client-ca\") pod \"route-controller-manager-6576b87f9c-lclf5\" (UID: \"21295e15-1d49-4261-9970-fb61c1d4854c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lclf5" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.347831 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/0e4b5f84-4954-4d85-8bae-db18ed352464-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-2trmp\" (UID: \"0e4b5f84-4954-4d85-8bae-db18ed352464\") " pod="openshift-authentication/oauth-openshift-558db77b4-2trmp" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.348181 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/26d0d35d-8516-46dd-8094-d72c27cfe267-machine-approver-tls\") pod \"machine-approver-56656f9798-g5z94\" (UID: \"26d0d35d-8516-46dd-8094-d72c27cfe267\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-g5z94" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.348751 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/d7952588-d49d-41c1-b5ec-60e6ea4a1efa-etcd-service-ca\") pod \"etcd-operator-b45778765-mnn4f\" (UID: \"d7952588-d49d-41c1-b5ec-60e6ea4a1efa\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mnn4f" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.350548 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.350866 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.351002 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-kf7zh"] Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.352040 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4bf1a18e-ec09-4f71-b6a1-645698067dc1-config\") pod \"controller-manager-879f6c89f-rnnlh\" (UID: \"4bf1a18e-ec09-4f71-b6a1-645698067dc1\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rnnlh" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.352327 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kf7zh" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.353107 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/911a0a7e-7a63-4b4a-8b74-6c21f8da7e31-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-d8pkd\" (UID: \"911a0a7e-7a63-4b4a-8b74-6c21f8da7e31\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-d8pkd" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.355137 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-nl2sk"] Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.356536 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/69303f47-f9a7-4c1c-b569-b1e4de56b2ff-images\") pod \"machine-api-operator-5694c8668f-hk7mf\" (UID: \"69303f47-f9a7-4c1c-b569-b1e4de56b2ff\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-hk7mf" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.356574 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-nl2sk" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.357238 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d7952588-d49d-41c1-b5ec-60e6ea4a1efa-serving-cert\") pod \"etcd-operator-b45778765-mnn4f\" (UID: \"d7952588-d49d-41c1-b5ec-60e6ea4a1efa\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mnn4f" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.357307 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0e4b5f84-4954-4d85-8bae-db18ed352464-audit-policies\") pod \"oauth-openshift-558db77b4-2trmp\" (UID: \"0e4b5f84-4954-4d85-8bae-db18ed352464\") " pod="openshift-authentication/oauth-openshift-558db77b4-2trmp" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.359872 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7952588-d49d-41c1-b5ec-60e6ea4a1efa-config\") pod \"etcd-operator-b45778765-mnn4f\" (UID: \"d7952588-d49d-41c1-b5ec-60e6ea4a1efa\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mnn4f" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.359925 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/0e4b5f84-4954-4d85-8bae-db18ed352464-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-2trmp\" (UID: \"0e4b5f84-4954-4d85-8bae-db18ed352464\") " pod="openshift-authentication/oauth-openshift-558db77b4-2trmp" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.360372 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/0e4b5f84-4954-4d85-8bae-db18ed352464-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-2trmp\" (UID: \"0e4b5f84-4954-4d85-8bae-db18ed352464\") " pod="openshift-authentication/oauth-openshift-558db77b4-2trmp" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.360498 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21295e15-1d49-4261-9970-fb61c1d4854c-config\") pod \"route-controller-manager-6576b87f9c-lclf5\" (UID: \"21295e15-1d49-4261-9970-fb61c1d4854c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lclf5" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.360727 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/0e4b5f84-4954-4d85-8bae-db18ed352464-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-2trmp\" (UID: \"0e4b5f84-4954-4d85-8bae-db18ed352464\") " pod="openshift-authentication/oauth-openshift-558db77b4-2trmp" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.362328 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0e4b5f84-4954-4d85-8bae-db18ed352464-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-2trmp\" (UID: \"0e4b5f84-4954-4d85-8bae-db18ed352464\") " pod="openshift-authentication/oauth-openshift-558db77b4-2trmp" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.363220 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a309e9aa-bf13-404b-90dd-8ddc81467822-metrics-tls\") pod \"dns-operator-744455d44c-pgf6s\" (UID: \"a309e9aa-bf13-404b-90dd-8ddc81467822\") " pod="openshift-dns-operator/dns-operator-744455d44c-pgf6s" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.363871 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/18d8d58b-3dab-4949-b9d1-313870a96dd5-serving-cert\") pod \"console-operator-58897d9998-dfltg\" (UID: \"18d8d58b-3dab-4949-b9d1-313870a96dd5\") " pod="openshift-console-operator/console-operator-58897d9998-dfltg" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.365424 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/69303f47-f9a7-4c1c-b569-b1e4de56b2ff-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-hk7mf\" (UID: \"69303f47-f9a7-4c1c-b569-b1e4de56b2ff\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-hk7mf" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.365458 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4bf1a18e-ec09-4f71-b6a1-645698067dc1-serving-cert\") pod \"controller-manager-879f6c89f-rnnlh\" (UID: \"4bf1a18e-ec09-4f71-b6a1-645698067dc1\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rnnlh" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.367915 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.368221 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-z9hjh"] Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.369519 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-z9hjh" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.369770 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29395560-pq7jf"] Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.374054 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/92733842-6b61-4327-b35c-e37cc833f7a3-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-p85cq\" (UID: \"92733842-6b61-4327-b35c-e37cc833f7a3\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p85cq" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.375381 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/0e4b5f84-4954-4d85-8bae-db18ed352464-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-2trmp\" (UID: \"0e4b5f84-4954-4d85-8bae-db18ed352464\") " pod="openshift-authentication/oauth-openshift-558db77b4-2trmp" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.375852 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/911a0a7e-7a63-4b4a-8b74-6c21f8da7e31-etcd-client\") pod \"apiserver-7bbb656c7d-d8pkd\" (UID: \"911a0a7e-7a63-4b4a-8b74-6c21f8da7e31\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-d8pkd" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.376240 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/0e4b5f84-4954-4d85-8bae-db18ed352464-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-2trmp\" (UID: \"0e4b5f84-4954-4d85-8bae-db18ed352464\") " pod="openshift-authentication/oauth-openshift-558db77b4-2trmp" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.376543 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/0e4b5f84-4954-4d85-8bae-db18ed352464-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-2trmp\" (UID: \"0e4b5f84-4954-4d85-8bae-db18ed352464\") " pod="openshift-authentication/oauth-openshift-558db77b4-2trmp" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.376662 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/107352c2-4357-43c2-9958-cb8a91714fa6-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-gspw5\" (UID: \"107352c2-4357-43c2-9958-cb8a91714fa6\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-gspw5" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.381328 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/911a0a7e-7a63-4b4a-8b74-6c21f8da7e31-encryption-config\") pod \"apiserver-7bbb656c7d-d8pkd\" (UID: \"911a0a7e-7a63-4b4a-8b74-6c21f8da7e31\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-d8pkd" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.391110 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/911a0a7e-7a63-4b4a-8b74-6c21f8da7e31-serving-cert\") pod \"apiserver-7bbb656c7d-d8pkd\" (UID: \"911a0a7e-7a63-4b4a-8b74-6c21f8da7e31\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-d8pkd" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.391844 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/0e4b5f84-4954-4d85-8bae-db18ed352464-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-2trmp\" (UID: \"0e4b5f84-4954-4d85-8bae-db18ed352464\") " pod="openshift-authentication/oauth-openshift-558db77b4-2trmp" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.391996 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29395560-pq7jf" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.394625 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-s7tww"] Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.395880 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-s7tww" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.397812 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.398653 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a59737d6-7609-4077-8fb8-d925f25a65f8-serving-cert\") pod \"authentication-operator-69f744f599-8kddm\" (UID: \"a59737d6-7609-4077-8fb8-d925f25a65f8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8kddm" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.398695 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a59737d6-7609-4077-8fb8-d925f25a65f8-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-8kddm\" (UID: \"a59737d6-7609-4077-8fb8-d925f25a65f8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8kddm" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.398773 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-skszv\" (UniqueName: \"kubernetes.io/projected/cfacdafa-2513-4d81-9222-c4b4fd8f8c10-kube-api-access-skszv\") pod \"migrator-59844c95c7-8s4px\" (UID: \"cfacdafa-2513-4d81-9222-c4b4fd8f8c10\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8s4px" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.398835 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/74ff4ad0-9479-4fa2-8df9-501c332e386c-config\") pod \"kube-apiserver-operator-766d6c64bb-5n549\" (UID: \"74ff4ad0-9479-4fa2-8df9-501c332e386c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5n549" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.398854 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a59737d6-7609-4077-8fb8-d925f25a65f8-config\") pod \"authentication-operator-69f744f599-8kddm\" (UID: \"a59737d6-7609-4077-8fb8-d925f25a65f8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8kddm" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.398888 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rtj8k\" (UniqueName: \"kubernetes.io/projected/a59737d6-7609-4077-8fb8-d925f25a65f8-kube-api-access-rtj8k\") pod \"authentication-operator-69f744f599-8kddm\" (UID: \"a59737d6-7609-4077-8fb8-d925f25a65f8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8kddm" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.398912 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/74ff4ad0-9479-4fa2-8df9-501c332e386c-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-5n549\" (UID: \"74ff4ad0-9479-4fa2-8df9-501c332e386c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5n549" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.398928 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2ldbq\" (UniqueName: \"kubernetes.io/projected/1fedc911-62a7-496f-93d1-f09769b89e86-kube-api-access-2ldbq\") pod \"openshift-apiserver-operator-796bbdcf4f-rv2nz\" (UID: \"1fedc911-62a7-496f-93d1-f09769b89e86\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rv2nz" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.398992 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a59737d6-7609-4077-8fb8-d925f25a65f8-service-ca-bundle\") pod \"authentication-operator-69f744f599-8kddm\" (UID: \"a59737d6-7609-4077-8fb8-d925f25a65f8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8kddm" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.399011 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/74ff4ad0-9479-4fa2-8df9-501c332e386c-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-5n549\" (UID: \"74ff4ad0-9479-4fa2-8df9-501c332e386c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5n549" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.399052 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1fedc911-62a7-496f-93d1-f09769b89e86-config\") pod \"openshift-apiserver-operator-796bbdcf4f-rv2nz\" (UID: \"1fedc911-62a7-496f-93d1-f09769b89e86\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rv2nz" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.399068 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1fedc911-62a7-496f-93d1-f09769b89e86-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-rv2nz\" (UID: \"1fedc911-62a7-496f-93d1-f09769b89e86\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rv2nz" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.401781 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/74ff4ad0-9479-4fa2-8df9-501c332e386c-config\") pod \"kube-apiserver-operator-766d6c64bb-5n549\" (UID: \"74ff4ad0-9479-4fa2-8df9-501c332e386c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5n549" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.401817 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a59737d6-7609-4077-8fb8-d925f25a65f8-service-ca-bundle\") pod \"authentication-operator-69f744f599-8kddm\" (UID: \"a59737d6-7609-4077-8fb8-d925f25a65f8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8kddm" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.401914 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p85cq"] Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.401954 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-d8pkd"] Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.402562 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1fedc911-62a7-496f-93d1-f09769b89e86-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-rv2nz\" (UID: \"1fedc911-62a7-496f-93d1-f09769b89e86\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rv2nz" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.402800 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1fedc911-62a7-496f-93d1-f09769b89e86-config\") pod \"openshift-apiserver-operator-796bbdcf4f-rv2nz\" (UID: \"1fedc911-62a7-496f-93d1-f09769b89e86\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rv2nz" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.403045 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/74ff4ad0-9479-4fa2-8df9-501c332e386c-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-5n549\" (UID: \"74ff4ad0-9479-4fa2-8df9-501c332e386c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5n549" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.406349 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a59737d6-7609-4077-8fb8-d925f25a65f8-serving-cert\") pod \"authentication-operator-69f744f599-8kddm\" (UID: \"a59737d6-7609-4077-8fb8-d925f25a65f8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8kddm" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.407530 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a59737d6-7609-4077-8fb8-d925f25a65f8-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-8kddm\" (UID: \"a59737d6-7609-4077-8fb8-d925f25a65f8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8kddm" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.407668 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.410730 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gxw89"] Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.410777 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-pgf6s"] Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.410789 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-hk7mf"] Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.411320 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a59737d6-7609-4077-8fb8-d925f25a65f8-config\") pod \"authentication-operator-69f744f599-8kddm\" (UID: \"a59737d6-7609-4077-8fb8-d925f25a65f8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8kddm" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.412496 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-bqktx"] Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.413776 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-bqktx" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.416652 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-dfltg"] Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.418267 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-rnnlh"] Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.420875 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-mnn4f"] Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.422780 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-gspw5"] Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.424643 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rv2nz"] Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.428376 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-2trmp"] Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.428461 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-lclf5"] Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.428892 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.429987 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-7fsr9"] Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.431659 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-8kddm"] Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.432797 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-szznp"] Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.434221 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-t5jsz"] Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.436031 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-8s4px"] Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.437875 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-lp8th"] Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.438698 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h59mj"] Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.439988 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4cx2q"] Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.442475 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-vgwv2"] Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.446898 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-lq8bp"] Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.448872 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-2hmqv"] Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.449421 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-lq8bp" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.449852 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-74phw"] Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.451603 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-kf7zh"] Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.452278 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.452448 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-v8bk6"] Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.453770 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6sfv4"] Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.454637 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5n549"] Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.456463 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-cc48c"] Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.456731 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-b9s66"] Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.457753 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-vf46x"] Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.459090 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-lq8bp"] Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.461883 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-s7tww"] Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.465059 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-5z687"] Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.466646 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-z9hjh"] Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.466762 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.468689 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29395560-pq7jf"] Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.470240 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5prl8"] Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.473316 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-bjdjf"] Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.474448 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-nl2sk"] Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.474629 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-bjdjf" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.475432 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-bjdjf"] Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.477411 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-fqd92"] Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.478096 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-fqd92"] Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.478266 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-fqd92" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.489098 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.506711 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.529681 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.546739 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.567026 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.586741 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.606844 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.626999 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.646482 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.666474 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.693328 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.706311 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.726453 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.746709 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.766316 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.806678 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.825849 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.847457 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.867555 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.887910 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.906947 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.927934 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.947104 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.967973 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Nov 21 14:11:08 crc kubenswrapper[4897]: I1121 14:11:08.987543 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Nov 21 14:11:09 crc kubenswrapper[4897]: I1121 14:11:09.007913 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Nov 21 14:11:09 crc kubenswrapper[4897]: I1121 14:11:09.027933 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Nov 21 14:11:09 crc kubenswrapper[4897]: I1121 14:11:09.047029 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Nov 21 14:11:09 crc kubenswrapper[4897]: I1121 14:11:09.067094 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Nov 21 14:11:09 crc kubenswrapper[4897]: I1121 14:11:09.087403 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Nov 21 14:11:09 crc kubenswrapper[4897]: I1121 14:11:09.107443 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Nov 21 14:11:09 crc kubenswrapper[4897]: I1121 14:11:09.126800 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Nov 21 14:11:09 crc kubenswrapper[4897]: I1121 14:11:09.147089 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Nov 21 14:11:09 crc kubenswrapper[4897]: I1121 14:11:09.167827 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Nov 21 14:11:09 crc kubenswrapper[4897]: I1121 14:11:09.188273 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Nov 21 14:11:09 crc kubenswrapper[4897]: I1121 14:11:09.207043 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Nov 21 14:11:09 crc kubenswrapper[4897]: I1121 14:11:09.228658 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Nov 21 14:11:09 crc kubenswrapper[4897]: I1121 14:11:09.246815 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Nov 21 14:11:09 crc kubenswrapper[4897]: I1121 14:11:09.268334 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Nov 21 14:11:09 crc kubenswrapper[4897]: I1121 14:11:09.287806 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Nov 21 14:11:09 crc kubenswrapper[4897]: I1121 14:11:09.305583 4897 request.go:700] Waited for 1.012783046s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-service-ca/secrets?fieldSelector=metadata.name%3Dservice-ca-dockercfg-pn86c&limit=500&resourceVersion=0 Nov 21 14:11:09 crc kubenswrapper[4897]: I1121 14:11:09.307616 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Nov 21 14:11:09 crc kubenswrapper[4897]: I1121 14:11:09.326039 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Nov 21 14:11:09 crc kubenswrapper[4897]: I1121 14:11:09.346243 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Nov 21 14:11:09 crc kubenswrapper[4897]: I1121 14:11:09.367457 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Nov 21 14:11:09 crc kubenswrapper[4897]: I1121 14:11:09.386892 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Nov 21 14:11:09 crc kubenswrapper[4897]: I1121 14:11:09.422411 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mtpkb\" (UniqueName: \"kubernetes.io/projected/21295e15-1d49-4261-9970-fb61c1d4854c-kube-api-access-mtpkb\") pod \"route-controller-manager-6576b87f9c-lclf5\" (UID: \"21295e15-1d49-4261-9970-fb61c1d4854c\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lclf5" Nov 21 14:11:09 crc kubenswrapper[4897]: I1121 14:11:09.443588 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6g94g\" (UniqueName: \"kubernetes.io/projected/d7952588-d49d-41c1-b5ec-60e6ea4a1efa-kube-api-access-6g94g\") pod \"etcd-operator-b45778765-mnn4f\" (UID: \"d7952588-d49d-41c1-b5ec-60e6ea4a1efa\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mnn4f" Nov 21 14:11:09 crc kubenswrapper[4897]: I1121 14:11:09.465681 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5qm7h\" (UniqueName: \"kubernetes.io/projected/18d8d58b-3dab-4949-b9d1-313870a96dd5-kube-api-access-5qm7h\") pod \"console-operator-58897d9998-dfltg\" (UID: \"18d8d58b-3dab-4949-b9d1-313870a96dd5\") " pod="openshift-console-operator/console-operator-58897d9998-dfltg" Nov 21 14:11:09 crc kubenswrapper[4897]: I1121 14:11:09.467653 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Nov 21 14:11:09 crc kubenswrapper[4897]: I1121 14:11:09.487057 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Nov 21 14:11:09 crc kubenswrapper[4897]: I1121 14:11:09.526843 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Nov 21 14:11:09 crc kubenswrapper[4897]: I1121 14:11:09.565792 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bw5l7\" (UniqueName: \"kubernetes.io/projected/26d0d35d-8516-46dd-8094-d72c27cfe267-kube-api-access-bw5l7\") pod \"machine-approver-56656f9798-g5z94\" (UID: \"26d0d35d-8516-46dd-8094-d72c27cfe267\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-g5z94" Nov 21 14:11:09 crc kubenswrapper[4897]: I1121 14:11:09.589039 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jw64h\" (UniqueName: \"kubernetes.io/projected/911a0a7e-7a63-4b4a-8b74-6c21f8da7e31-kube-api-access-jw64h\") pod \"apiserver-7bbb656c7d-d8pkd\" (UID: \"911a0a7e-7a63-4b4a-8b74-6c21f8da7e31\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-d8pkd" Nov 21 14:11:09 crc kubenswrapper[4897]: I1121 14:11:09.610545 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/92733842-6b61-4327-b35c-e37cc833f7a3-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-p85cq\" (UID: \"92733842-6b61-4327-b35c-e37cc833f7a3\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p85cq" Nov 21 14:11:09 crc kubenswrapper[4897]: I1121 14:11:09.626093 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-dfltg" Nov 21 14:11:09 crc kubenswrapper[4897]: I1121 14:11:09.634434 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-68nxr\" (UniqueName: \"kubernetes.io/projected/69303f47-f9a7-4c1c-b569-b1e4de56b2ff-kube-api-access-68nxr\") pod \"machine-api-operator-5694c8668f-hk7mf\" (UID: \"69303f47-f9a7-4c1c-b569-b1e4de56b2ff\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-hk7mf" Nov 21 14:11:09 crc kubenswrapper[4897]: I1121 14:11:09.649095 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-d8pkd" Nov 21 14:11:09 crc kubenswrapper[4897]: I1121 14:11:09.655119 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xjzmb\" (UniqueName: \"kubernetes.io/projected/a309e9aa-bf13-404b-90dd-8ddc81467822-kube-api-access-xjzmb\") pod \"dns-operator-744455d44c-pgf6s\" (UID: \"a309e9aa-bf13-404b-90dd-8ddc81467822\") " pod="openshift-dns-operator/dns-operator-744455d44c-pgf6s" Nov 21 14:11:09 crc kubenswrapper[4897]: I1121 14:11:09.655273 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lclf5" Nov 21 14:11:09 crc kubenswrapper[4897]: I1121 14:11:09.665498 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-mnn4f" Nov 21 14:11:09 crc kubenswrapper[4897]: I1121 14:11:09.670191 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dpwxx\" (UniqueName: \"kubernetes.io/projected/92733842-6b61-4327-b35c-e37cc833f7a3-kube-api-access-dpwxx\") pod \"cluster-image-registry-operator-dc59b4c8b-p85cq\" (UID: \"92733842-6b61-4327-b35c-e37cc833f7a3\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p85cq" Nov 21 14:11:09 crc kubenswrapper[4897]: I1121 14:11:09.674382 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-g5z94" Nov 21 14:11:09 crc kubenswrapper[4897]: W1121 14:11:09.689150 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod26d0d35d_8516_46dd_8094_d72c27cfe267.slice/crio-78f11a7c5a3214a3a311b1e9f4952a47c8b72b2ecd97e114af2c50acf76b6f2d WatchSource:0}: Error finding container 78f11a7c5a3214a3a311b1e9f4952a47c8b72b2ecd97e114af2c50acf76b6f2d: Status 404 returned error can't find the container with id 78f11a7c5a3214a3a311b1e9f4952a47c8b72b2ecd97e114af2c50acf76b6f2d Nov 21 14:11:09 crc kubenswrapper[4897]: I1121 14:11:09.689965 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qbtbc\" (UniqueName: \"kubernetes.io/projected/107352c2-4357-43c2-9958-cb8a91714fa6-kube-api-access-qbtbc\") pod \"cluster-samples-operator-665b6dd947-gspw5\" (UID: \"107352c2-4357-43c2-9958-cb8a91714fa6\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-gspw5" Nov 21 14:11:09 crc kubenswrapper[4897]: I1121 14:11:09.691065 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Nov 21 14:11:09 crc kubenswrapper[4897]: I1121 14:11:09.722036 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-gspw5" Nov 21 14:11:09 crc kubenswrapper[4897]: I1121 14:11:09.727057 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Nov 21 14:11:09 crc kubenswrapper[4897]: I1121 14:11:09.730277 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n9k72\" (UniqueName: \"kubernetes.io/projected/4bf1a18e-ec09-4f71-b6a1-645698067dc1-kube-api-access-n9k72\") pod \"controller-manager-879f6c89f-rnnlh\" (UID: \"4bf1a18e-ec09-4f71-b6a1-645698067dc1\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rnnlh" Nov 21 14:11:09 crc kubenswrapper[4897]: I1121 14:11:09.747674 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Nov 21 14:11:09 crc kubenswrapper[4897]: I1121 14:11:09.771413 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p85cq" Nov 21 14:11:09 crc kubenswrapper[4897]: I1121 14:11:09.771502 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Nov 21 14:11:09 crc kubenswrapper[4897]: I1121 14:11:09.786718 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Nov 21 14:11:09 crc kubenswrapper[4897]: I1121 14:11:09.816557 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-g5z94" event={"ID":"26d0d35d-8516-46dd-8094-d72c27cfe267","Type":"ContainerStarted","Data":"78f11a7c5a3214a3a311b1e9f4952a47c8b72b2ecd97e114af2c50acf76b6f2d"} Nov 21 14:11:09 crc kubenswrapper[4897]: I1121 14:11:09.822906 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmhkr\" (UniqueName: \"kubernetes.io/projected/0e4b5f84-4954-4d85-8bae-db18ed352464-kube-api-access-vmhkr\") pod \"oauth-openshift-558db77b4-2trmp\" (UID: \"0e4b5f84-4954-4d85-8bae-db18ed352464\") " pod="openshift-authentication/oauth-openshift-558db77b4-2trmp" Nov 21 14:11:09 crc kubenswrapper[4897]: I1121 14:11:09.844101 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-snt8h\" (UniqueName: \"kubernetes.io/projected/cc2bd33e-fa09-4675-9d48-b546668b5b86-kube-api-access-snt8h\") pod \"openshift-controller-manager-operator-756b6f6bc6-gxw89\" (UID: \"cc2bd33e-fa09-4675-9d48-b546668b5b86\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gxw89" Nov 21 14:11:09 crc kubenswrapper[4897]: I1121 14:11:09.850109 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Nov 21 14:11:09 crc kubenswrapper[4897]: I1121 14:11:09.868147 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Nov 21 14:11:09 crc kubenswrapper[4897]: I1121 14:11:09.888334 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Nov 21 14:11:09 crc kubenswrapper[4897]: I1121 14:11:09.894632 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-pgf6s" Nov 21 14:11:09 crc kubenswrapper[4897]: I1121 14:11:09.912917 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Nov 21 14:11:09 crc kubenswrapper[4897]: I1121 14:11:09.917565 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-hk7mf" Nov 21 14:11:09 crc kubenswrapper[4897]: I1121 14:11:09.926742 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Nov 21 14:11:09 crc kubenswrapper[4897]: I1121 14:11:09.936930 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-2trmp" Nov 21 14:11:09 crc kubenswrapper[4897]: I1121 14:11:09.946818 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 21 14:11:09 crc kubenswrapper[4897]: I1121 14:11:09.949908 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-rnnlh" Nov 21 14:11:09 crc kubenswrapper[4897]: I1121 14:11:09.967016 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 21 14:11:09 crc kubenswrapper[4897]: I1121 14:11:09.987287 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Nov 21 14:11:09 crc kubenswrapper[4897]: I1121 14:11:09.997515 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gxw89" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.009791 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.012302 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-gspw5"] Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.026894 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.027567 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p85cq"] Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.047638 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.068770 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.088165 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.101357 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-pgf6s"] Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.108669 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-dfltg"] Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.116059 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.128179 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.145665 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-mnn4f"] Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.145774 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.155948 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-lclf5"] Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.157565 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-d8pkd"] Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.166624 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Nov 21 14:11:10 crc kubenswrapper[4897]: W1121 14:11:10.184084 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod911a0a7e_7a63_4b4a_8b74_6c21f8da7e31.slice/crio-e9be315ce6221ffc901fcec23528e9b5614abf7cacd34e296164e080bbf7cea8 WatchSource:0}: Error finding container e9be315ce6221ffc901fcec23528e9b5614abf7cacd34e296164e080bbf7cea8: Status 404 returned error can't find the container with id e9be315ce6221ffc901fcec23528e9b5614abf7cacd34e296164e080bbf7cea8 Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.187893 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.221272 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-rnnlh"] Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.229206 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/74ff4ad0-9479-4fa2-8df9-501c332e386c-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-5n549\" (UID: \"74ff4ad0-9479-4fa2-8df9-501c332e386c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5n549" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.261895 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2ldbq\" (UniqueName: \"kubernetes.io/projected/1fedc911-62a7-496f-93d1-f09769b89e86-kube-api-access-2ldbq\") pod \"openshift-apiserver-operator-796bbdcf4f-rv2nz\" (UID: \"1fedc911-62a7-496f-93d1-f09769b89e86\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rv2nz" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.274076 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rv2nz" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.281884 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rtj8k\" (UniqueName: \"kubernetes.io/projected/a59737d6-7609-4077-8fb8-d925f25a65f8-kube-api-access-rtj8k\") pod \"authentication-operator-69f744f599-8kddm\" (UID: \"a59737d6-7609-4077-8fb8-d925f25a65f8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8kddm" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.284651 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-skszv\" (UniqueName: \"kubernetes.io/projected/cfacdafa-2513-4d81-9222-c4b4fd8f8c10-kube-api-access-skszv\") pod \"migrator-59844c95c7-8s4px\" (UID: \"cfacdafa-2513-4d81-9222-c4b4fd8f8c10\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8s4px" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.286523 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.294802 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-8kddm" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.306656 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.318245 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5n549" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.325637 4897 request.go:700] Waited for 1.911475106s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/secrets?fieldSelector=metadata.name%3Dnode-bootstrapper-token&limit=500&resourceVersion=0 Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.329620 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.330183 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-hk7mf"] Nov 21 14:11:10 crc kubenswrapper[4897]: W1121 14:11:10.344522 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod69303f47_f9a7_4c1c_b569_b1e4de56b2ff.slice/crio-43f575d1cba140342028964620eb494bfa30ac550548f93e3e8e2ade04d86150 WatchSource:0}: Error finding container 43f575d1cba140342028964620eb494bfa30ac550548f93e3e8e2ade04d86150: Status 404 returned error can't find the container with id 43f575d1cba140342028964620eb494bfa30ac550548f93e3e8e2ade04d86150 Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.346349 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.355642 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8s4px" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.367098 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.386321 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.406590 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.428606 4897 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.446603 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.468277 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.476003 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-2trmp"] Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.477886 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gxw89"] Nov 21 14:11:10 crc kubenswrapper[4897]: W1121 14:11:10.489476 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0e4b5f84_4954_4d85_8bae_db18ed352464.slice/crio-6476da7699e9ba7191eaa620c6d3b896c8ad8bc1907d7c323f28d02aa9d040ae WatchSource:0}: Error finding container 6476da7699e9ba7191eaa620c6d3b896c8ad8bc1907d7c323f28d02aa9d040ae: Status 404 returned error can't find the container with id 6476da7699e9ba7191eaa620c6d3b896c8ad8bc1907d7c323f28d02aa9d040ae Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.489585 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.508885 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.527359 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.563591 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rv2nz"] Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.602418 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-8kddm"] Nov 21 14:11:10 crc kubenswrapper[4897]: W1121 14:11:10.622432 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1fedc911_62a7_496f_93d1_f09769b89e86.slice/crio-0bc7f34d648c4bd69390702ca681c648ef462f40f6c9cba0dbf7a887ca2fc65c WatchSource:0}: Error finding container 0bc7f34d648c4bd69390702ca681c648ef462f40f6c9cba0dbf7a887ca2fc65c: Status 404 returned error can't find the container with id 0bc7f34d648c4bd69390702ca681c648ef462f40f6c9cba0dbf7a887ca2fc65c Nov 21 14:11:10 crc kubenswrapper[4897]: W1121 14:11:10.625286 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda59737d6_7609_4077_8fb8_d925f25a65f8.slice/crio-2da646a0805014d5381f6dc7fbf3315b03cc4fcca8a9065a6db2172d95b29d2f WatchSource:0}: Error finding container 2da646a0805014d5381f6dc7fbf3315b03cc4fcca8a9065a6db2172d95b29d2f: Status 404 returned error can't find the container with id 2da646a0805014d5381f6dc7fbf3315b03cc4fcca8a9065a6db2172d95b29d2f Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.634349 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-86gsj\" (UniqueName: \"kubernetes.io/projected/9f63b8dd-ffb7-4963-8fcc-0c133a2c045b-kube-api-access-86gsj\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.634378 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/8342c8a2-87f4-4bf9-a2ca-073c02b277ad-console-oauth-config\") pod \"console-f9d7485db-7fsr9\" (UID: \"8342c8a2-87f4-4bf9-a2ca-073c02b277ad\") " pod="openshift-console/console-f9d7485db-7fsr9" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.634439 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/9f63b8dd-ffb7-4963-8fcc-0c133a2c045b-registry-certificates\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.634455 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6jwzn\" (UniqueName: \"kubernetes.io/projected/5648ea03-bf08-419d-87fb-6287e11dd399-kube-api-access-6jwzn\") pod \"ingress-operator-5b745b69d9-2hmqv\" (UID: \"5648ea03-bf08-419d-87fb-6287e11dd399\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2hmqv" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.634472 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ml62r\" (UniqueName: \"kubernetes.io/projected/00cdea84-a65d-4422-8ecb-7bc8b912c557-kube-api-access-ml62r\") pod \"openshift-config-operator-7777fb866f-t5jsz\" (UID: \"00cdea84-a65d-4422-8ecb-7bc8b912c557\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-t5jsz" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.634514 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h7l9d\" (UniqueName: \"kubernetes.io/projected/e095a18e-3a0b-4418-95a8-a2cc3de5bd7b-kube-api-access-h7l9d\") pod \"router-default-5444994796-2kgg2\" (UID: \"e095a18e-3a0b-4418-95a8-a2cc3de5bd7b\") " pod="openshift-ingress/router-default-5444994796-2kgg2" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.634551 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8342c8a2-87f4-4bf9-a2ca-073c02b277ad-service-ca\") pod \"console-f9d7485db-7fsr9\" (UID: \"8342c8a2-87f4-4bf9-a2ca-073c02b277ad\") " pod="openshift-console/console-f9d7485db-7fsr9" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.634568 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5648ea03-bf08-419d-87fb-6287e11dd399-trusted-ca\") pod \"ingress-operator-5b745b69d9-2hmqv\" (UID: \"5648ea03-bf08-419d-87fb-6287e11dd399\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2hmqv" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.634583 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/e095a18e-3a0b-4418-95a8-a2cc3de5bd7b-default-certificate\") pod \"router-default-5444994796-2kgg2\" (UID: \"e095a18e-3a0b-4418-95a8-a2cc3de5bd7b\") " pod="openshift-ingress/router-default-5444994796-2kgg2" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.634627 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/9f63b8dd-ffb7-4963-8fcc-0c133a2c045b-installation-pull-secrets\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.634644 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cgg9b\" (UniqueName: \"kubernetes.io/projected/b76cdd1d-6a5d-42ac-8b03-1cc105eb1c1f-kube-api-access-cgg9b\") pod \"kube-storage-version-migrator-operator-b67b599dd-szznp\" (UID: \"b76cdd1d-6a5d-42ac-8b03-1cc105eb1c1f\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-szznp" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.634660 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9f63b8dd-ffb7-4963-8fcc-0c133a2c045b-trusted-ca\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.634675 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/9f63b8dd-ffb7-4963-8fcc-0c133a2c045b-ca-trust-extracted\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.634691 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b76cdd1d-6a5d-42ac-8b03-1cc105eb1c1f-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-szznp\" (UID: \"b76cdd1d-6a5d-42ac-8b03-1cc105eb1c1f\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-szznp" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.634771 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e095a18e-3a0b-4418-95a8-a2cc3de5bd7b-service-ca-bundle\") pod \"router-default-5444994796-2kgg2\" (UID: \"e095a18e-3a0b-4418-95a8-a2cc3de5bd7b\") " pod="openshift-ingress/router-default-5444994796-2kgg2" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.634787 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e095a18e-3a0b-4418-95a8-a2cc3de5bd7b-metrics-certs\") pod \"router-default-5444994796-2kgg2\" (UID: \"e095a18e-3a0b-4418-95a8-a2cc3de5bd7b\") " pod="openshift-ingress/router-default-5444994796-2kgg2" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.634819 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/9f63b8dd-ffb7-4963-8fcc-0c133a2c045b-registry-tls\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.634874 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/98941613-bca1-4b35-9dba-547294a1feb5-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-lp8th\" (UID: \"98941613-bca1-4b35-9dba-547294a1feb5\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-lp8th" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.634934 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/98941613-bca1-4b35-9dba-547294a1feb5-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-lp8th\" (UID: \"98941613-bca1-4b35-9dba-547294a1feb5\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-lp8th" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.634953 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lsdt5\" (UniqueName: \"kubernetes.io/projected/8342c8a2-87f4-4bf9-a2ca-073c02b277ad-kube-api-access-lsdt5\") pod \"console-f9d7485db-7fsr9\" (UID: \"8342c8a2-87f4-4bf9-a2ca-073c02b277ad\") " pod="openshift-console/console-f9d7485db-7fsr9" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.634994 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/00cdea84-a65d-4422-8ecb-7bc8b912c557-available-featuregates\") pod \"openshift-config-operator-7777fb866f-t5jsz\" (UID: \"00cdea84-a65d-4422-8ecb-7bc8b912c557\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-t5jsz" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.635030 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/98941613-bca1-4b35-9dba-547294a1feb5-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-lp8th\" (UID: \"98941613-bca1-4b35-9dba-547294a1feb5\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-lp8th" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.635103 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/00cdea84-a65d-4422-8ecb-7bc8b912c557-serving-cert\") pod \"openshift-config-operator-7777fb866f-t5jsz\" (UID: \"00cdea84-a65d-4422-8ecb-7bc8b912c557\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-t5jsz" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.635119 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b76cdd1d-6a5d-42ac-8b03-1cc105eb1c1f-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-szznp\" (UID: \"b76cdd1d-6a5d-42ac-8b03-1cc105eb1c1f\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-szznp" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.635141 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8342c8a2-87f4-4bf9-a2ca-073c02b277ad-trusted-ca-bundle\") pod \"console-f9d7485db-7fsr9\" (UID: \"8342c8a2-87f4-4bf9-a2ca-073c02b277ad\") " pod="openshift-console/console-f9d7485db-7fsr9" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.635215 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-57xcw\" (UniqueName: \"kubernetes.io/projected/82f81ac2-d2b2-49de-82f3-a51796dbff03-kube-api-access-57xcw\") pod \"downloads-7954f5f757-vgwv2\" (UID: \"82f81ac2-d2b2-49de-82f3-a51796dbff03\") " pod="openshift-console/downloads-7954f5f757-vgwv2" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.635229 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/5648ea03-bf08-419d-87fb-6287e11dd399-metrics-tls\") pod \"ingress-operator-5b745b69d9-2hmqv\" (UID: \"5648ea03-bf08-419d-87fb-6287e11dd399\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2hmqv" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.635260 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.635278 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/8342c8a2-87f4-4bf9-a2ca-073c02b277ad-console-serving-cert\") pod \"console-f9d7485db-7fsr9\" (UID: \"8342c8a2-87f4-4bf9-a2ca-073c02b277ad\") " pod="openshift-console/console-f9d7485db-7fsr9" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.635312 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/8342c8a2-87f4-4bf9-a2ca-073c02b277ad-console-config\") pod \"console-f9d7485db-7fsr9\" (UID: \"8342c8a2-87f4-4bf9-a2ca-073c02b277ad\") " pod="openshift-console/console-f9d7485db-7fsr9" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.635328 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/8342c8a2-87f4-4bf9-a2ca-073c02b277ad-oauth-serving-cert\") pod \"console-f9d7485db-7fsr9\" (UID: \"8342c8a2-87f4-4bf9-a2ca-073c02b277ad\") " pod="openshift-console/console-f9d7485db-7fsr9" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.635343 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5648ea03-bf08-419d-87fb-6287e11dd399-bound-sa-token\") pod \"ingress-operator-5b745b69d9-2hmqv\" (UID: \"5648ea03-bf08-419d-87fb-6287e11dd399\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2hmqv" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.635387 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9f63b8dd-ffb7-4963-8fcc-0c133a2c045b-bound-sa-token\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.635414 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/e095a18e-3a0b-4418-95a8-a2cc3de5bd7b-stats-auth\") pod \"router-default-5444994796-2kgg2\" (UID: \"e095a18e-3a0b-4418-95a8-a2cc3de5bd7b\") " pod="openshift-ingress/router-default-5444994796-2kgg2" Nov 21 14:11:10 crc kubenswrapper[4897]: E1121 14:11:10.636876 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-21 14:11:11.13686052 +0000 UTC m=+148.421454105 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5z687" (UID: "9f63b8dd-ffb7-4963-8fcc-0c133a2c045b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.641105 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5n549"] Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.677751 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-8s4px"] Nov 21 14:11:10 crc kubenswrapper[4897]: W1121 14:11:10.693443 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod74ff4ad0_9479_4fa2_8df9_501c332e386c.slice/crio-590e3ddad9a0e05c59eb4f2f06d039ffc57f637acad6bdd26bf2b80e2b06be44 WatchSource:0}: Error finding container 590e3ddad9a0e05c59eb4f2f06d039ffc57f637acad6bdd26bf2b80e2b06be44: Status 404 returned error can't find the container with id 590e3ddad9a0e05c59eb4f2f06d039ffc57f637acad6bdd26bf2b80e2b06be44 Nov 21 14:11:10 crc kubenswrapper[4897]: W1121 14:11:10.698912 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcfacdafa_2513_4d81_9222_c4b4fd8f8c10.slice/crio-30ff71eecd490c3b834cf2b90ff5b0a27eff05f753010afe2725399d7d2f4f05 WatchSource:0}: Error finding container 30ff71eecd490c3b834cf2b90ff5b0a27eff05f753010afe2725399d7d2f4f05: Status 404 returned error can't find the container with id 30ff71eecd490c3b834cf2b90ff5b0a27eff05f753010afe2725399d7d2f4f05 Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.737942 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.738211 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z588q\" (UniqueName: \"kubernetes.io/projected/5262754c-6c80-4da0-8d9f-a58d63cfeadb-kube-api-access-z588q\") pod \"package-server-manager-789f6589d5-cc48c\" (UID: \"5262754c-6c80-4da0-8d9f-a58d63cfeadb\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-cc48c" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.738248 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1d13c714-5451-44fd-9dad-8d613a6e3545-serving-cert\") pod \"service-ca-operator-777779d784-v8bk6\" (UID: \"1d13c714-5451-44fd-9dad-8d613a6e3545\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-v8bk6" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.738270 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/027b031c-1b21-40e8-aa7c-4a342aae265a-serving-cert\") pod \"apiserver-76f77b778f-s7tww\" (UID: \"027b031c-1b21-40e8-aa7c-4a342aae265a\") " pod="openshift-apiserver/apiserver-76f77b778f-s7tww" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.738290 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/6d1fd716-6991-43b6-9f1f-268576e864a0-signing-key\") pod \"service-ca-9c57cc56f-b9s66\" (UID: \"6d1fd716-6991-43b6-9f1f-268576e864a0\") " pod="openshift-service-ca/service-ca-9c57cc56f-b9s66" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.738310 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec19858a-32ef-443f-b307-df94c5d3e8de-config\") pod \"kube-controller-manager-operator-78b949d7b-5prl8\" (UID: \"ec19858a-32ef-443f-b307-df94c5d3e8de\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5prl8" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.738336 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/9f63b8dd-ffb7-4963-8fcc-0c133a2c045b-registry-certificates\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.738360 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/78db7e2e-ba45-464e-9b1d-0fc61191955a-plugins-dir\") pod \"csi-hostpathplugin-bjdjf\" (UID: \"78db7e2e-ba45-464e-9b1d-0fc61191955a\") " pod="hostpath-provisioner/csi-hostpathplugin-bjdjf" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.738385 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ml62r\" (UniqueName: \"kubernetes.io/projected/00cdea84-a65d-4422-8ecb-7bc8b912c557-kube-api-access-ml62r\") pod \"openshift-config-operator-7777fb866f-t5jsz\" (UID: \"00cdea84-a65d-4422-8ecb-7bc8b912c557\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-t5jsz" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.738407 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h7l9d\" (UniqueName: \"kubernetes.io/projected/e095a18e-3a0b-4418-95a8-a2cc3de5bd7b-kube-api-access-h7l9d\") pod \"router-default-5444994796-2kgg2\" (UID: \"e095a18e-3a0b-4418-95a8-a2cc3de5bd7b\") " pod="openshift-ingress/router-default-5444994796-2kgg2" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.738427 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/027b031c-1b21-40e8-aa7c-4a342aae265a-etcd-client\") pod \"apiserver-76f77b778f-s7tww\" (UID: \"027b031c-1b21-40e8-aa7c-4a342aae265a\") " pod="openshift-apiserver/apiserver-76f77b778f-s7tww" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.738452 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/e095a18e-3a0b-4418-95a8-a2cc3de5bd7b-default-certificate\") pod \"router-default-5444994796-2kgg2\" (UID: \"e095a18e-3a0b-4418-95a8-a2cc3de5bd7b\") " pod="openshift-ingress/router-default-5444994796-2kgg2" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.738476 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/56521238-a49d-446a-b4a9-e2809ff15826-srv-cert\") pod \"catalog-operator-68c6474976-4cx2q\" (UID: \"56521238-a49d-446a-b4a9-e2809ff15826\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4cx2q" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.738504 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/f2080446-c5c3-4c09-b8ae-54359bbb72ba-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-vf46x\" (UID: \"f2080446-c5c3-4c09-b8ae-54359bbb72ba\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-vf46x" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.738546 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cgg9b\" (UniqueName: \"kubernetes.io/projected/b76cdd1d-6a5d-42ac-8b03-1cc105eb1c1f-kube-api-access-cgg9b\") pod \"kube-storage-version-migrator-operator-b67b599dd-szznp\" (UID: \"b76cdd1d-6a5d-42ac-8b03-1cc105eb1c1f\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-szznp" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.738598 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b76cdd1d-6a5d-42ac-8b03-1cc105eb1c1f-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-szznp\" (UID: \"b76cdd1d-6a5d-42ac-8b03-1cc105eb1c1f\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-szznp" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.738621 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wts2z\" (UniqueName: \"kubernetes.io/projected/85811958-7716-47c2-b9ad-9392647bb172-kube-api-access-wts2z\") pod \"machine-config-server-bqktx\" (UID: \"85811958-7716-47c2-b9ad-9392647bb172\") " pod="openshift-machine-config-operator/machine-config-server-bqktx" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.738649 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e095a18e-3a0b-4418-95a8-a2cc3de5bd7b-metrics-certs\") pod \"router-default-5444994796-2kgg2\" (UID: \"e095a18e-3a0b-4418-95a8-a2cc3de5bd7b\") " pod="openshift-ingress/router-default-5444994796-2kgg2" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.738673 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/4c817f5a-6f7c-48d3-9d45-bdc401646bed-auth-proxy-config\") pod \"machine-config-operator-74547568cd-kf7zh\" (UID: \"4c817f5a-6f7c-48d3-9d45-bdc401646bed\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kf7zh" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.738697 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/9f63b8dd-ffb7-4963-8fcc-0c133a2c045b-registry-tls\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.738726 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/78db7e2e-ba45-464e-9b1d-0fc61191955a-socket-dir\") pod \"csi-hostpathplugin-bjdjf\" (UID: \"78db7e2e-ba45-464e-9b1d-0fc61191955a\") " pod="hostpath-provisioner/csi-hostpathplugin-bjdjf" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.738750 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/98941613-bca1-4b35-9dba-547294a1feb5-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-lp8th\" (UID: \"98941613-bca1-4b35-9dba-547294a1feb5\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-lp8th" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.738771 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/6d1fd716-6991-43b6-9f1f-268576e864a0-signing-cabundle\") pod \"service-ca-9c57cc56f-b9s66\" (UID: \"6d1fd716-6991-43b6-9f1f-268576e864a0\") " pod="openshift-service-ca/service-ca-9c57cc56f-b9s66" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.738802 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/85811958-7716-47c2-b9ad-9392647bb172-node-bootstrap-token\") pod \"machine-config-server-bqktx\" (UID: \"85811958-7716-47c2-b9ad-9392647bb172\") " pod="openshift-machine-config-operator/machine-config-server-bqktx" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.738825 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-666j9\" (UniqueName: \"kubernetes.io/projected/c14a1c93-8be4-42b7-aae1-9c31d53a6bfc-kube-api-access-666j9\") pod \"collect-profiles-29395560-pq7jf\" (UID: \"c14a1c93-8be4-42b7-aae1-9c31d53a6bfc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29395560-pq7jf" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.738847 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1d13c714-5451-44fd-9dad-8d613a6e3545-config\") pod \"service-ca-operator-777779d784-v8bk6\" (UID: \"1d13c714-5451-44fd-9dad-8d613a6e3545\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-v8bk6" Nov 21 14:11:10 crc kubenswrapper[4897]: E1121 14:11:10.738982 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-21 14:11:11.238953887 +0000 UTC m=+148.523547362 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.738871 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-525rj\" (UniqueName: \"kubernetes.io/projected/f2080446-c5c3-4c09-b8ae-54359bbb72ba-kube-api-access-525rj\") pod \"multus-admission-controller-857f4d67dd-vf46x\" (UID: \"f2080446-c5c3-4c09-b8ae-54359bbb72ba\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-vf46x" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.740093 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jc9p7\" (UniqueName: \"kubernetes.io/projected/027b031c-1b21-40e8-aa7c-4a342aae265a-kube-api-access-jc9p7\") pod \"apiserver-76f77b778f-s7tww\" (UID: \"027b031c-1b21-40e8-aa7c-4a342aae265a\") " pod="openshift-apiserver/apiserver-76f77b778f-s7tww" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.740141 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/00cdea84-a65d-4422-8ecb-7bc8b912c557-serving-cert\") pod \"openshift-config-operator-7777fb866f-t5jsz\" (UID: \"00cdea84-a65d-4422-8ecb-7bc8b912c557\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-t5jsz" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.740167 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/027b031c-1b21-40e8-aa7c-4a342aae265a-config\") pod \"apiserver-76f77b778f-s7tww\" (UID: \"027b031c-1b21-40e8-aa7c-4a342aae265a\") " pod="openshift-apiserver/apiserver-76f77b778f-s7tww" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.740190 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/56521238-a49d-446a-b4a9-e2809ff15826-profile-collector-cert\") pod \"catalog-operator-68c6474976-4cx2q\" (UID: \"56521238-a49d-446a-b4a9-e2809ff15826\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4cx2q" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.740212 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c14a1c93-8be4-42b7-aae1-9c31d53a6bfc-config-volume\") pod \"collect-profiles-29395560-pq7jf\" (UID: \"c14a1c93-8be4-42b7-aae1-9c31d53a6bfc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29395560-pq7jf" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.740238 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b76cdd1d-6a5d-42ac-8b03-1cc105eb1c1f-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-szznp\" (UID: \"b76cdd1d-6a5d-42ac-8b03-1cc105eb1c1f\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-szznp" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.740253 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-57xcw\" (UniqueName: \"kubernetes.io/projected/82f81ac2-d2b2-49de-82f3-a51796dbff03-kube-api-access-57xcw\") pod \"downloads-7954f5f757-vgwv2\" (UID: \"82f81ac2-d2b2-49de-82f3-a51796dbff03\") " pod="openshift-console/downloads-7954f5f757-vgwv2" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.740305 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c14a1c93-8be4-42b7-aae1-9c31d53a6bfc-secret-volume\") pod \"collect-profiles-29395560-pq7jf\" (UID: \"c14a1c93-8be4-42b7-aae1-9c31d53a6bfc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29395560-pq7jf" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.740335 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/8342c8a2-87f4-4bf9-a2ca-073c02b277ad-console-serving-cert\") pod \"console-f9d7485db-7fsr9\" (UID: \"8342c8a2-87f4-4bf9-a2ca-073c02b277ad\") " pod="openshift-console/console-f9d7485db-7fsr9" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.740353 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/027b031c-1b21-40e8-aa7c-4a342aae265a-node-pullsecrets\") pod \"apiserver-76f77b778f-s7tww\" (UID: \"027b031c-1b21-40e8-aa7c-4a342aae265a\") " pod="openshift-apiserver/apiserver-76f77b778f-s7tww" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.740370 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/027b031c-1b21-40e8-aa7c-4a342aae265a-audit-dir\") pod \"apiserver-76f77b778f-s7tww\" (UID: \"027b031c-1b21-40e8-aa7c-4a342aae265a\") " pod="openshift-apiserver/apiserver-76f77b778f-s7tww" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.740388 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/8342c8a2-87f4-4bf9-a2ca-073c02b277ad-oauth-serving-cert\") pod \"console-f9d7485db-7fsr9\" (UID: \"8342c8a2-87f4-4bf9-a2ca-073c02b277ad\") " pod="openshift-console/console-f9d7485db-7fsr9" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.740405 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/85811958-7716-47c2-b9ad-9392647bb172-certs\") pod \"machine-config-server-bqktx\" (UID: \"85811958-7716-47c2-b9ad-9392647bb172\") " pod="openshift-machine-config-operator/machine-config-server-bqktx" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.740475 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f7d61c37-621e-4812-bbaa-74006927176a-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-74phw\" (UID: \"f7d61c37-621e-4812-bbaa-74006927176a\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-74phw" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.740497 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-86gsj\" (UniqueName: \"kubernetes.io/projected/9f63b8dd-ffb7-4963-8fcc-0c133a2c045b-kube-api-access-86gsj\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.740702 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/9f63b8dd-ffb7-4963-8fcc-0c133a2c045b-registry-certificates\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.741392 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/98941613-bca1-4b35-9dba-547294a1feb5-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-lp8th\" (UID: \"98941613-bca1-4b35-9dba-547294a1feb5\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-lp8th" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.741551 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b0b1043c-75c9-4643-a400-a5c009dfb5b6-cert\") pod \"ingress-canary-fqd92\" (UID: \"b0b1043c-75c9-4643-a400-a5c009dfb5b6\") " pod="openshift-ingress-canary/ingress-canary-fqd92" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.741594 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j2ddt\" (UniqueName: \"kubernetes.io/projected/1d13c714-5451-44fd-9dad-8d613a6e3545-kube-api-access-j2ddt\") pod \"service-ca-operator-777779d784-v8bk6\" (UID: \"1d13c714-5451-44fd-9dad-8d613a6e3545\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-v8bk6" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.741635 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6jwzn\" (UniqueName: \"kubernetes.io/projected/5648ea03-bf08-419d-87fb-6287e11dd399-kube-api-access-6jwzn\") pod \"ingress-operator-5b745b69d9-2hmqv\" (UID: \"5648ea03-bf08-419d-87fb-6287e11dd399\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2hmqv" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.741663 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/20a1d330-d99c-480d-b44c-4d8a5b4c6453-metrics-tls\") pod \"dns-default-lq8bp\" (UID: \"20a1d330-d99c-480d-b44c-4d8a5b4c6453\") " pod="openshift-dns/dns-default-lq8bp" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.741735 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/f7d61c37-621e-4812-bbaa-74006927176a-proxy-tls\") pod \"machine-config-controller-84d6567774-74phw\" (UID: \"f7d61c37-621e-4812-bbaa-74006927176a\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-74phw" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.741762 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/11301ff8-fa01-48da-a95e-fc244fd0a575-tmpfs\") pod \"packageserver-d55dfcdfc-6sfv4\" (UID: \"11301ff8-fa01-48da-a95e-fc244fd0a575\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6sfv4" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.741790 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8342c8a2-87f4-4bf9-a2ca-073c02b277ad-service-ca\") pod \"console-f9d7485db-7fsr9\" (UID: \"8342c8a2-87f4-4bf9-a2ca-073c02b277ad\") " pod="openshift-console/console-f9d7485db-7fsr9" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.741838 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5648ea03-bf08-419d-87fb-6287e11dd399-trusted-ca\") pod \"ingress-operator-5b745b69d9-2hmqv\" (UID: \"5648ea03-bf08-419d-87fb-6287e11dd399\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2hmqv" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.741867 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4l4r5\" (UniqueName: \"kubernetes.io/projected/6a9aee9a-c47c-4ad2-9431-1a4ac5ce15fd-kube-api-access-4l4r5\") pod \"control-plane-machine-set-operator-78cbb6b69f-nl2sk\" (UID: \"6a9aee9a-c47c-4ad2-9431-1a4ac5ce15fd\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-nl2sk" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.741934 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tv5cm\" (UniqueName: \"kubernetes.io/projected/20a1d330-d99c-480d-b44c-4d8a5b4c6453-kube-api-access-tv5cm\") pod \"dns-default-lq8bp\" (UID: \"20a1d330-d99c-480d-b44c-4d8a5b4c6453\") " pod="openshift-dns/dns-default-lq8bp" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.742067 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/8342c8a2-87f4-4bf9-a2ca-073c02b277ad-oauth-serving-cert\") pod \"console-f9d7485db-7fsr9\" (UID: \"8342c8a2-87f4-4bf9-a2ca-073c02b277ad\") " pod="openshift-console/console-f9d7485db-7fsr9" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.742146 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/9f63b8dd-ffb7-4963-8fcc-0c133a2c045b-installation-pull-secrets\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.742196 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ec19858a-32ef-443f-b307-df94c5d3e8de-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-5prl8\" (UID: \"ec19858a-32ef-443f-b307-df94c5d3e8de\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5prl8" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.742456 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9f63b8dd-ffb7-4963-8fcc-0c133a2c045b-trusted-ca\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.742525 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/4c817f5a-6f7c-48d3-9d45-bdc401646bed-images\") pod \"machine-config-operator-74547568cd-kf7zh\" (UID: \"4c817f5a-6f7c-48d3-9d45-bdc401646bed\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kf7zh" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.742547 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b24e87a6-3cd8-4a37-b537-e9a5332cb6c7-srv-cert\") pod \"olm-operator-6b444d44fb-h59mj\" (UID: \"b24e87a6-3cd8-4a37-b537-e9a5332cb6c7\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h59mj" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.742583 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8342c8a2-87f4-4bf9-a2ca-073c02b277ad-service-ca\") pod \"console-f9d7485db-7fsr9\" (UID: \"8342c8a2-87f4-4bf9-a2ca-073c02b277ad\") " pod="openshift-console/console-f9d7485db-7fsr9" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.742700 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/9f63b8dd-ffb7-4963-8fcc-0c133a2c045b-ca-trust-extracted\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.742759 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8x66f\" (UniqueName: \"kubernetes.io/projected/f7d61c37-621e-4812-bbaa-74006927176a-kube-api-access-8x66f\") pod \"machine-config-controller-84d6567774-74phw\" (UID: \"f7d61c37-621e-4812-bbaa-74006927176a\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-74phw" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.742786 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ec19858a-32ef-443f-b307-df94c5d3e8de-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-5prl8\" (UID: \"ec19858a-32ef-443f-b307-df94c5d3e8de\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5prl8" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.742885 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cq46c\" (UniqueName: \"kubernetes.io/projected/78db7e2e-ba45-464e-9b1d-0fc61191955a-kube-api-access-cq46c\") pod \"csi-hostpathplugin-bjdjf\" (UID: \"78db7e2e-ba45-464e-9b1d-0fc61191955a\") " pod="hostpath-provisioner/csi-hostpathplugin-bjdjf" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.742920 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/4c817f5a-6f7c-48d3-9d45-bdc401646bed-proxy-tls\") pod \"machine-config-operator-74547568cd-kf7zh\" (UID: \"4c817f5a-6f7c-48d3-9d45-bdc401646bed\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kf7zh" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.743067 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5648ea03-bf08-419d-87fb-6287e11dd399-trusted-ca\") pod \"ingress-operator-5b745b69d9-2hmqv\" (UID: \"5648ea03-bf08-419d-87fb-6287e11dd399\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2hmqv" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.743079 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/9f63b8dd-ffb7-4963-8fcc-0c133a2c045b-ca-trust-extracted\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.743135 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e095a18e-3a0b-4418-95a8-a2cc3de5bd7b-service-ca-bundle\") pod \"router-default-5444994796-2kgg2\" (UID: \"e095a18e-3a0b-4418-95a8-a2cc3de5bd7b\") " pod="openshift-ingress/router-default-5444994796-2kgg2" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.743220 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/027b031c-1b21-40e8-aa7c-4a342aae265a-trusted-ca-bundle\") pod \"apiserver-76f77b778f-s7tww\" (UID: \"027b031c-1b21-40e8-aa7c-4a342aae265a\") " pod="openshift-apiserver/apiserver-76f77b778f-s7tww" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.743261 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9ckq9\" (UniqueName: \"kubernetes.io/projected/b0b1043c-75c9-4643-a400-a5c009dfb5b6-kube-api-access-9ckq9\") pod \"ingress-canary-fqd92\" (UID: \"b0b1043c-75c9-4643-a400-a5c009dfb5b6\") " pod="openshift-ingress-canary/ingress-canary-fqd92" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.743294 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e128526a-8489-47e1-a711-4977f2b0d46d-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-z9hjh\" (UID: \"e128526a-8489-47e1-a711-4977f2b0d46d\") " pod="openshift-marketplace/marketplace-operator-79b997595-z9hjh" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.743331 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-47p67\" (UniqueName: \"kubernetes.io/projected/11301ff8-fa01-48da-a95e-fc244fd0a575-kube-api-access-47p67\") pod \"packageserver-d55dfcdfc-6sfv4\" (UID: \"11301ff8-fa01-48da-a95e-fc244fd0a575\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6sfv4" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.743380 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gz5nl\" (UniqueName: \"kubernetes.io/projected/56521238-a49d-446a-b4a9-e2809ff15826-kube-api-access-gz5nl\") pod \"catalog-operator-68c6474976-4cx2q\" (UID: \"56521238-a49d-446a-b4a9-e2809ff15826\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4cx2q" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.743411 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9b76t\" (UniqueName: \"kubernetes.io/projected/b24e87a6-3cd8-4a37-b537-e9a5332cb6c7-kube-api-access-9b76t\") pod \"olm-operator-6b444d44fb-h59mj\" (UID: \"b24e87a6-3cd8-4a37-b537-e9a5332cb6c7\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h59mj" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.743431 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/027b031c-1b21-40e8-aa7c-4a342aae265a-audit\") pod \"apiserver-76f77b778f-s7tww\" (UID: \"027b031c-1b21-40e8-aa7c-4a342aae265a\") " pod="openshift-apiserver/apiserver-76f77b778f-s7tww" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.743466 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/e128526a-8489-47e1-a711-4977f2b0d46d-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-z9hjh\" (UID: \"e128526a-8489-47e1-a711-4977f2b0d46d\") " pod="openshift-marketplace/marketplace-operator-79b997595-z9hjh" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.743489 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mksx9\" (UniqueName: \"kubernetes.io/projected/6d1fd716-6991-43b6-9f1f-268576e864a0-kube-api-access-mksx9\") pod \"service-ca-9c57cc56f-b9s66\" (UID: \"6d1fd716-6991-43b6-9f1f-268576e864a0\") " pod="openshift-service-ca/service-ca-9c57cc56f-b9s66" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.743608 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/98941613-bca1-4b35-9dba-547294a1feb5-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-lp8th\" (UID: \"98941613-bca1-4b35-9dba-547294a1feb5\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-lp8th" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.743664 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lsdt5\" (UniqueName: \"kubernetes.io/projected/8342c8a2-87f4-4bf9-a2ca-073c02b277ad-kube-api-access-lsdt5\") pod \"console-f9d7485db-7fsr9\" (UID: \"8342c8a2-87f4-4bf9-a2ca-073c02b277ad\") " pod="openshift-console/console-f9d7485db-7fsr9" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.743692 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/00cdea84-a65d-4422-8ecb-7bc8b912c557-available-featuregates\") pod \"openshift-config-operator-7777fb866f-t5jsz\" (UID: \"00cdea84-a65d-4422-8ecb-7bc8b912c557\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-t5jsz" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.743740 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/11301ff8-fa01-48da-a95e-fc244fd0a575-webhook-cert\") pod \"packageserver-d55dfcdfc-6sfv4\" (UID: \"11301ff8-fa01-48da-a95e-fc244fd0a575\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6sfv4" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.743755 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e095a18e-3a0b-4418-95a8-a2cc3de5bd7b-service-ca-bundle\") pod \"router-default-5444994796-2kgg2\" (UID: \"e095a18e-3a0b-4418-95a8-a2cc3de5bd7b\") " pod="openshift-ingress/router-default-5444994796-2kgg2" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.743767 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/98941613-bca1-4b35-9dba-547294a1feb5-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-lp8th\" (UID: \"98941613-bca1-4b35-9dba-547294a1feb5\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-lp8th" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.743794 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/78db7e2e-ba45-464e-9b1d-0fc61191955a-mountpoint-dir\") pod \"csi-hostpathplugin-bjdjf\" (UID: \"78db7e2e-ba45-464e-9b1d-0fc61191955a\") " pod="hostpath-provisioner/csi-hostpathplugin-bjdjf" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.743823 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/11301ff8-fa01-48da-a95e-fc244fd0a575-apiservice-cert\") pod \"packageserver-d55dfcdfc-6sfv4\" (UID: \"11301ff8-fa01-48da-a95e-fc244fd0a575\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6sfv4" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.743881 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/78db7e2e-ba45-464e-9b1d-0fc61191955a-csi-data-dir\") pod \"csi-hostpathplugin-bjdjf\" (UID: \"78db7e2e-ba45-464e-9b1d-0fc61191955a\") " pod="hostpath-provisioner/csi-hostpathplugin-bjdjf" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.744086 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/00cdea84-a65d-4422-8ecb-7bc8b912c557-available-featuregates\") pod \"openshift-config-operator-7777fb866f-t5jsz\" (UID: \"00cdea84-a65d-4422-8ecb-7bc8b912c557\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-t5jsz" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.744555 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b76cdd1d-6a5d-42ac-8b03-1cc105eb1c1f-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-szznp\" (UID: \"b76cdd1d-6a5d-42ac-8b03-1cc105eb1c1f\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-szznp" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.744838 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8342c8a2-87f4-4bf9-a2ca-073c02b277ad-trusted-ca-bundle\") pod \"console-f9d7485db-7fsr9\" (UID: \"8342c8a2-87f4-4bf9-a2ca-073c02b277ad\") " pod="openshift-console/console-f9d7485db-7fsr9" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.744922 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9f63b8dd-ffb7-4963-8fcc-0c133a2c045b-trusted-ca\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.745381 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6a9aee9a-c47c-4ad2-9431-1a4ac5ce15fd-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-nl2sk\" (UID: \"6a9aee9a-c47c-4ad2-9431-1a4ac5ce15fd\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-nl2sk" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.745415 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s2xp5\" (UniqueName: \"kubernetes.io/projected/4c817f5a-6f7c-48d3-9d45-bdc401646bed-kube-api-access-s2xp5\") pod \"machine-config-operator-74547568cd-kf7zh\" (UID: \"4c817f5a-6f7c-48d3-9d45-bdc401646bed\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kf7zh" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.745439 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/027b031c-1b21-40e8-aa7c-4a342aae265a-image-import-ca\") pod \"apiserver-76f77b778f-s7tww\" (UID: \"027b031c-1b21-40e8-aa7c-4a342aae265a\") " pod="openshift-apiserver/apiserver-76f77b778f-s7tww" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.745523 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/5648ea03-bf08-419d-87fb-6287e11dd399-metrics-tls\") pod \"ingress-operator-5b745b69d9-2hmqv\" (UID: \"5648ea03-bf08-419d-87fb-6287e11dd399\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2hmqv" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.745562 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/20a1d330-d99c-480d-b44c-4d8a5b4c6453-config-volume\") pod \"dns-default-lq8bp\" (UID: \"20a1d330-d99c-480d-b44c-4d8a5b4c6453\") " pod="openshift-dns/dns-default-lq8bp" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.745620 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.745754 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8342c8a2-87f4-4bf9-a2ca-073c02b277ad-trusted-ca-bundle\") pod \"console-f9d7485db-7fsr9\" (UID: \"8342c8a2-87f4-4bf9-a2ca-073c02b277ad\") " pod="openshift-console/console-f9d7485db-7fsr9" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.745853 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/78db7e2e-ba45-464e-9b1d-0fc61191955a-registration-dir\") pod \"csi-hostpathplugin-bjdjf\" (UID: \"78db7e2e-ba45-464e-9b1d-0fc61191955a\") " pod="hostpath-provisioner/csi-hostpathplugin-bjdjf" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.745886 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2wp82\" (UniqueName: \"kubernetes.io/projected/e128526a-8489-47e1-a711-4977f2b0d46d-kube-api-access-2wp82\") pod \"marketplace-operator-79b997595-z9hjh\" (UID: \"e128526a-8489-47e1-a711-4977f2b0d46d\") " pod="openshift-marketplace/marketplace-operator-79b997595-z9hjh" Nov 21 14:11:10 crc kubenswrapper[4897]: E1121 14:11:10.745952 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-21 14:11:11.245938904 +0000 UTC m=+148.530532379 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5z687" (UID: "9f63b8dd-ffb7-4963-8fcc-0c133a2c045b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.745982 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/8342c8a2-87f4-4bf9-a2ca-073c02b277ad-console-config\") pod \"console-f9d7485db-7fsr9\" (UID: \"8342c8a2-87f4-4bf9-a2ca-073c02b277ad\") " pod="openshift-console/console-f9d7485db-7fsr9" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.746019 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5648ea03-bf08-419d-87fb-6287e11dd399-bound-sa-token\") pod \"ingress-operator-5b745b69d9-2hmqv\" (UID: \"5648ea03-bf08-419d-87fb-6287e11dd399\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2hmqv" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.746453 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/9f63b8dd-ffb7-4963-8fcc-0c133a2c045b-registry-tls\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.746492 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9f63b8dd-ffb7-4963-8fcc-0c133a2c045b-bound-sa-token\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.746809 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/5262754c-6c80-4da0-8d9f-a58d63cfeadb-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-cc48c\" (UID: \"5262754c-6c80-4da0-8d9f-a58d63cfeadb\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-cc48c" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.746843 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/027b031c-1b21-40e8-aa7c-4a342aae265a-etcd-serving-ca\") pod \"apiserver-76f77b778f-s7tww\" (UID: \"027b031c-1b21-40e8-aa7c-4a342aae265a\") " pod="openshift-apiserver/apiserver-76f77b778f-s7tww" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.746876 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/027b031c-1b21-40e8-aa7c-4a342aae265a-encryption-config\") pod \"apiserver-76f77b778f-s7tww\" (UID: \"027b031c-1b21-40e8-aa7c-4a342aae265a\") " pod="openshift-apiserver/apiserver-76f77b778f-s7tww" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.746910 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/e095a18e-3a0b-4418-95a8-a2cc3de5bd7b-stats-auth\") pod \"router-default-5444994796-2kgg2\" (UID: \"e095a18e-3a0b-4418-95a8-a2cc3de5bd7b\") " pod="openshift-ingress/router-default-5444994796-2kgg2" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.746935 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b24e87a6-3cd8-4a37-b537-e9a5332cb6c7-profile-collector-cert\") pod \"olm-operator-6b444d44fb-h59mj\" (UID: \"b24e87a6-3cd8-4a37-b537-e9a5332cb6c7\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h59mj" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.746961 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/8342c8a2-87f4-4bf9-a2ca-073c02b277ad-console-oauth-config\") pod \"console-f9d7485db-7fsr9\" (UID: \"8342c8a2-87f4-4bf9-a2ca-073c02b277ad\") " pod="openshift-console/console-f9d7485db-7fsr9" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.749124 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/5648ea03-bf08-419d-87fb-6287e11dd399-metrics-tls\") pod \"ingress-operator-5b745b69d9-2hmqv\" (UID: \"5648ea03-bf08-419d-87fb-6287e11dd399\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2hmqv" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.750744 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/8342c8a2-87f4-4bf9-a2ca-073c02b277ad-console-oauth-config\") pod \"console-f9d7485db-7fsr9\" (UID: \"8342c8a2-87f4-4bf9-a2ca-073c02b277ad\") " pod="openshift-console/console-f9d7485db-7fsr9" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.751215 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/8342c8a2-87f4-4bf9-a2ca-073c02b277ad-console-config\") pod \"console-f9d7485db-7fsr9\" (UID: \"8342c8a2-87f4-4bf9-a2ca-073c02b277ad\") " pod="openshift-console/console-f9d7485db-7fsr9" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.751221 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b76cdd1d-6a5d-42ac-8b03-1cc105eb1c1f-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-szznp\" (UID: \"b76cdd1d-6a5d-42ac-8b03-1cc105eb1c1f\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-szznp" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.751839 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/e095a18e-3a0b-4418-95a8-a2cc3de5bd7b-default-certificate\") pod \"router-default-5444994796-2kgg2\" (UID: \"e095a18e-3a0b-4418-95a8-a2cc3de5bd7b\") " pod="openshift-ingress/router-default-5444994796-2kgg2" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.752484 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/9f63b8dd-ffb7-4963-8fcc-0c133a2c045b-installation-pull-secrets\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.762219 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/00cdea84-a65d-4422-8ecb-7bc8b912c557-serving-cert\") pod \"openshift-config-operator-7777fb866f-t5jsz\" (UID: \"00cdea84-a65d-4422-8ecb-7bc8b912c557\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-t5jsz" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.763564 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/8342c8a2-87f4-4bf9-a2ca-073c02b277ad-console-serving-cert\") pod \"console-f9d7485db-7fsr9\" (UID: \"8342c8a2-87f4-4bf9-a2ca-073c02b277ad\") " pod="openshift-console/console-f9d7485db-7fsr9" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.767170 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/e095a18e-3a0b-4418-95a8-a2cc3de5bd7b-stats-auth\") pod \"router-default-5444994796-2kgg2\" (UID: \"e095a18e-3a0b-4418-95a8-a2cc3de5bd7b\") " pod="openshift-ingress/router-default-5444994796-2kgg2" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.770925 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e095a18e-3a0b-4418-95a8-a2cc3de5bd7b-metrics-certs\") pod \"router-default-5444994796-2kgg2\" (UID: \"e095a18e-3a0b-4418-95a8-a2cc3de5bd7b\") " pod="openshift-ingress/router-default-5444994796-2kgg2" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.772302 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/98941613-bca1-4b35-9dba-547294a1feb5-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-lp8th\" (UID: \"98941613-bca1-4b35-9dba-547294a1feb5\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-lp8th" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.781655 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cgg9b\" (UniqueName: \"kubernetes.io/projected/b76cdd1d-6a5d-42ac-8b03-1cc105eb1c1f-kube-api-access-cgg9b\") pod \"kube-storage-version-migrator-operator-b67b599dd-szznp\" (UID: \"b76cdd1d-6a5d-42ac-8b03-1cc105eb1c1f\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-szznp" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.804230 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h7l9d\" (UniqueName: \"kubernetes.io/projected/e095a18e-3a0b-4418-95a8-a2cc3de5bd7b-kube-api-access-h7l9d\") pod \"router-default-5444994796-2kgg2\" (UID: \"e095a18e-3a0b-4418-95a8-a2cc3de5bd7b\") " pod="openshift-ingress/router-default-5444994796-2kgg2" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.820802 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8s4px" event={"ID":"cfacdafa-2513-4d81-9222-c4b4fd8f8c10","Type":"ContainerStarted","Data":"30ff71eecd490c3b834cf2b90ff5b0a27eff05f753010afe2725399d7d2f4f05"} Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.821649 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-d8pkd" event={"ID":"911a0a7e-7a63-4b4a-8b74-6c21f8da7e31","Type":"ContainerStarted","Data":"e9be315ce6221ffc901fcec23528e9b5614abf7cacd34e296164e080bbf7cea8"} Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.822420 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5n549" event={"ID":"74ff4ad0-9479-4fa2-8df9-501c332e386c","Type":"ContainerStarted","Data":"590e3ddad9a0e05c59eb4f2f06d039ffc57f637acad6bdd26bf2b80e2b06be44"} Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.823854 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-g5z94" event={"ID":"26d0d35d-8516-46dd-8094-d72c27cfe267","Type":"ContainerStarted","Data":"5372cec3e4d0299f2dfa3865fd8511d95fae9a9dbd9e4bd77b59ffa691d49458"} Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.823891 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-g5z94" event={"ID":"26d0d35d-8516-46dd-8094-d72c27cfe267","Type":"ContainerStarted","Data":"1632076e49eae6e1ece3fac1a2f1a0cd655ae566021b18a78005bf8d688f7042"} Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.824213 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ml62r\" (UniqueName: \"kubernetes.io/projected/00cdea84-a65d-4422-8ecb-7bc8b912c557-kube-api-access-ml62r\") pod \"openshift-config-operator-7777fb866f-t5jsz\" (UID: \"00cdea84-a65d-4422-8ecb-7bc8b912c557\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-t5jsz" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.824766 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p85cq" event={"ID":"92733842-6b61-4327-b35c-e37cc833f7a3","Type":"ContainerStarted","Data":"8a13560fa66e9b5fdf660fb51427b28bfb2b168740e23997f4afebd6cad4f156"} Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.824794 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p85cq" event={"ID":"92733842-6b61-4327-b35c-e37cc833f7a3","Type":"ContainerStarted","Data":"8e447f94556ffa09fbb188f82ab32e0a77f75da3df0185d87e1c201116cf125e"} Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.826816 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-pgf6s" event={"ID":"a309e9aa-bf13-404b-90dd-8ddc81467822","Type":"ContainerStarted","Data":"d047c71206b8e8af17523afd3e196aee1c3177f0a38a27bb6083682276703854"} Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.826846 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-pgf6s" event={"ID":"a309e9aa-bf13-404b-90dd-8ddc81467822","Type":"ContainerStarted","Data":"b86714f1b896d202d622e5c0636215b7bbd9f8c013e4f31b7ea526484293fb4a"} Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.827988 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-8kddm" event={"ID":"a59737d6-7609-4077-8fb8-d925f25a65f8","Type":"ContainerStarted","Data":"2da646a0805014d5381f6dc7fbf3315b03cc4fcca8a9065a6db2172d95b29d2f"} Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.829413 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lclf5" event={"ID":"21295e15-1d49-4261-9970-fb61c1d4854c","Type":"ContainerStarted","Data":"86eaed5522081a2fff93adb245c025c95e874ea66e172ecaed1b5770a821a7bc"} Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.829441 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lclf5" event={"ID":"21295e15-1d49-4261-9970-fb61c1d4854c","Type":"ContainerStarted","Data":"3b8df1d580432d827538c096ed3b5f0a33e8962c87d588205ff955abcc71e1a4"} Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.831393 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-dfltg" event={"ID":"18d8d58b-3dab-4949-b9d1-313870a96dd5","Type":"ContainerStarted","Data":"bcdd062d396527c66da126146101f2593580724060ac97677aa09b2914f003c1"} Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.831420 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-dfltg" event={"ID":"18d8d58b-3dab-4949-b9d1-313870a96dd5","Type":"ContainerStarted","Data":"f1eae202c56e1f79e91666f857613214fe666ca09030bc9ad7f2f8a3d68408cb"} Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.832277 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-dfltg" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.833156 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gxw89" event={"ID":"cc2bd33e-fa09-4675-9d48-b546668b5b86","Type":"ContainerStarted","Data":"eff9848b28544def1da3df3b612a8c54082e080adb4acdbdf46d43655cab9776"} Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.834060 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-rnnlh" event={"ID":"4bf1a18e-ec09-4f71-b6a1-645698067dc1","Type":"ContainerStarted","Data":"fa8cb63bd527ad7c0cc4785027de1ca3217b1c4c8d640a8757cd888455fe1d99"} Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.834115 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-rnnlh" event={"ID":"4bf1a18e-ec09-4f71-b6a1-645698067dc1","Type":"ContainerStarted","Data":"bd87699e03c4d7fba1fe4336158f39a61691180e715516516c484eb3d45e40fd"} Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.834583 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-rnnlh" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.835399 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-2trmp" event={"ID":"0e4b5f84-4954-4d85-8bae-db18ed352464","Type":"ContainerStarted","Data":"6476da7699e9ba7191eaa620c6d3b896c8ad8bc1907d7c323f28d02aa9d040ae"} Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.835902 4897 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-rnnlh container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.11:8443/healthz\": dial tcp 10.217.0.11:8443: connect: connection refused" start-of-body= Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.835916 4897 patch_prober.go:28] interesting pod/console-operator-58897d9998-dfltg container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.10:8443/readyz\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.835927 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-rnnlh" podUID="4bf1a18e-ec09-4f71-b6a1-645698067dc1" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.11:8443/healthz\": dial tcp 10.217.0.11:8443: connect: connection refused" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.835944 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-dfltg" podUID="18d8d58b-3dab-4949-b9d1-313870a96dd5" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.10:8443/readyz\": dial tcp 10.217.0.10:8443: connect: connection refused" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.836456 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-hk7mf" event={"ID":"69303f47-f9a7-4c1c-b569-b1e4de56b2ff","Type":"ContainerStarted","Data":"843d0885fb5f0214b314d3e9bad7831af50735c4c73e3ad13a0c1ba6ca2a752c"} Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.836480 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-hk7mf" event={"ID":"69303f47-f9a7-4c1c-b569-b1e4de56b2ff","Type":"ContainerStarted","Data":"43f575d1cba140342028964620eb494bfa30ac550548f93e3e8e2ade04d86150"} Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.837340 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rv2nz" event={"ID":"1fedc911-62a7-496f-93d1-f09769b89e86","Type":"ContainerStarted","Data":"0bc7f34d648c4bd69390702ca681c648ef462f40f6c9cba0dbf7a887ca2fc65c"} Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.839059 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-gspw5" event={"ID":"107352c2-4357-43c2-9958-cb8a91714fa6","Type":"ContainerStarted","Data":"8fb629719297d00a7b227ee8a06350dfe70a741a13dd9de5bb7ae07c08514c69"} Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.839088 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-gspw5" event={"ID":"107352c2-4357-43c2-9958-cb8a91714fa6","Type":"ContainerStarted","Data":"c63483210f5b2f2a51d27c3c4c519e8f8578176a921a24bb65f90afa46f4555c"} Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.840407 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-mnn4f" event={"ID":"d7952588-d49d-41c1-b5ec-60e6ea4a1efa","Type":"ContainerStarted","Data":"2284148e0eae0b44be67cce26f73fc1e6d351aacf4fb35467d86197df2b1437c"} Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.847762 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-57xcw\" (UniqueName: \"kubernetes.io/projected/82f81ac2-d2b2-49de-82f3-a51796dbff03-kube-api-access-57xcw\") pod \"downloads-7954f5f757-vgwv2\" (UID: \"82f81ac2-d2b2-49de-82f3-a51796dbff03\") " pod="openshift-console/downloads-7954f5f757-vgwv2" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.851706 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.851936 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2xp5\" (UniqueName: \"kubernetes.io/projected/4c817f5a-6f7c-48d3-9d45-bdc401646bed-kube-api-access-s2xp5\") pod \"machine-config-operator-74547568cd-kf7zh\" (UID: \"4c817f5a-6f7c-48d3-9d45-bdc401646bed\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kf7zh" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.851976 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/027b031c-1b21-40e8-aa7c-4a342aae265a-image-import-ca\") pod \"apiserver-76f77b778f-s7tww\" (UID: \"027b031c-1b21-40e8-aa7c-4a342aae265a\") " pod="openshift-apiserver/apiserver-76f77b778f-s7tww" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.852000 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/20a1d330-d99c-480d-b44c-4d8a5b4c6453-config-volume\") pod \"dns-default-lq8bp\" (UID: \"20a1d330-d99c-480d-b44c-4d8a5b4c6453\") " pod="openshift-dns/dns-default-lq8bp" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.852043 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/78db7e2e-ba45-464e-9b1d-0fc61191955a-registration-dir\") pod \"csi-hostpathplugin-bjdjf\" (UID: \"78db7e2e-ba45-464e-9b1d-0fc61191955a\") " pod="hostpath-provisioner/csi-hostpathplugin-bjdjf" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.852065 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2wp82\" (UniqueName: \"kubernetes.io/projected/e128526a-8489-47e1-a711-4977f2b0d46d-kube-api-access-2wp82\") pod \"marketplace-operator-79b997595-z9hjh\" (UID: \"e128526a-8489-47e1-a711-4977f2b0d46d\") " pod="openshift-marketplace/marketplace-operator-79b997595-z9hjh" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.852092 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/5262754c-6c80-4da0-8d9f-a58d63cfeadb-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-cc48c\" (UID: \"5262754c-6c80-4da0-8d9f-a58d63cfeadb\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-cc48c" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.852114 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/027b031c-1b21-40e8-aa7c-4a342aae265a-etcd-serving-ca\") pod \"apiserver-76f77b778f-s7tww\" (UID: \"027b031c-1b21-40e8-aa7c-4a342aae265a\") " pod="openshift-apiserver/apiserver-76f77b778f-s7tww" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.852132 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/027b031c-1b21-40e8-aa7c-4a342aae265a-encryption-config\") pod \"apiserver-76f77b778f-s7tww\" (UID: \"027b031c-1b21-40e8-aa7c-4a342aae265a\") " pod="openshift-apiserver/apiserver-76f77b778f-s7tww" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.852153 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b24e87a6-3cd8-4a37-b537-e9a5332cb6c7-profile-collector-cert\") pod \"olm-operator-6b444d44fb-h59mj\" (UID: \"b24e87a6-3cd8-4a37-b537-e9a5332cb6c7\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h59mj" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.852174 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z588q\" (UniqueName: \"kubernetes.io/projected/5262754c-6c80-4da0-8d9f-a58d63cfeadb-kube-api-access-z588q\") pod \"package-server-manager-789f6589d5-cc48c\" (UID: \"5262754c-6c80-4da0-8d9f-a58d63cfeadb\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-cc48c" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.852196 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1d13c714-5451-44fd-9dad-8d613a6e3545-serving-cert\") pod \"service-ca-operator-777779d784-v8bk6\" (UID: \"1d13c714-5451-44fd-9dad-8d613a6e3545\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-v8bk6" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.852217 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/027b031c-1b21-40e8-aa7c-4a342aae265a-serving-cert\") pod \"apiserver-76f77b778f-s7tww\" (UID: \"027b031c-1b21-40e8-aa7c-4a342aae265a\") " pod="openshift-apiserver/apiserver-76f77b778f-s7tww" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.852243 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/78db7e2e-ba45-464e-9b1d-0fc61191955a-plugins-dir\") pod \"csi-hostpathplugin-bjdjf\" (UID: \"78db7e2e-ba45-464e-9b1d-0fc61191955a\") " pod="hostpath-provisioner/csi-hostpathplugin-bjdjf" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.852266 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/6d1fd716-6991-43b6-9f1f-268576e864a0-signing-key\") pod \"service-ca-9c57cc56f-b9s66\" (UID: \"6d1fd716-6991-43b6-9f1f-268576e864a0\") " pod="openshift-service-ca/service-ca-9c57cc56f-b9s66" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.852288 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec19858a-32ef-443f-b307-df94c5d3e8de-config\") pod \"kube-controller-manager-operator-78b949d7b-5prl8\" (UID: \"ec19858a-32ef-443f-b307-df94c5d3e8de\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5prl8" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.852311 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/027b031c-1b21-40e8-aa7c-4a342aae265a-etcd-client\") pod \"apiserver-76f77b778f-s7tww\" (UID: \"027b031c-1b21-40e8-aa7c-4a342aae265a\") " pod="openshift-apiserver/apiserver-76f77b778f-s7tww" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.852334 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/56521238-a49d-446a-b4a9-e2809ff15826-srv-cert\") pod \"catalog-operator-68c6474976-4cx2q\" (UID: \"56521238-a49d-446a-b4a9-e2809ff15826\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4cx2q" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.852357 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/f2080446-c5c3-4c09-b8ae-54359bbb72ba-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-vf46x\" (UID: \"f2080446-c5c3-4c09-b8ae-54359bbb72ba\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-vf46x" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.852383 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wts2z\" (UniqueName: \"kubernetes.io/projected/85811958-7716-47c2-b9ad-9392647bb172-kube-api-access-wts2z\") pod \"machine-config-server-bqktx\" (UID: \"85811958-7716-47c2-b9ad-9392647bb172\") " pod="openshift-machine-config-operator/machine-config-server-bqktx" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.852409 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/4c817f5a-6f7c-48d3-9d45-bdc401646bed-auth-proxy-config\") pod \"machine-config-operator-74547568cd-kf7zh\" (UID: \"4c817f5a-6f7c-48d3-9d45-bdc401646bed\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kf7zh" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.852435 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/78db7e2e-ba45-464e-9b1d-0fc61191955a-socket-dir\") pod \"csi-hostpathplugin-bjdjf\" (UID: \"78db7e2e-ba45-464e-9b1d-0fc61191955a\") " pod="hostpath-provisioner/csi-hostpathplugin-bjdjf" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.852458 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/6d1fd716-6991-43b6-9f1f-268576e864a0-signing-cabundle\") pod \"service-ca-9c57cc56f-b9s66\" (UID: \"6d1fd716-6991-43b6-9f1f-268576e864a0\") " pod="openshift-service-ca/service-ca-9c57cc56f-b9s66" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.852492 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/85811958-7716-47c2-b9ad-9392647bb172-node-bootstrap-token\") pod \"machine-config-server-bqktx\" (UID: \"85811958-7716-47c2-b9ad-9392647bb172\") " pod="openshift-machine-config-operator/machine-config-server-bqktx" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.852537 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-666j9\" (UniqueName: \"kubernetes.io/projected/c14a1c93-8be4-42b7-aae1-9c31d53a6bfc-kube-api-access-666j9\") pod \"collect-profiles-29395560-pq7jf\" (UID: \"c14a1c93-8be4-42b7-aae1-9c31d53a6bfc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29395560-pq7jf" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.852567 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1d13c714-5451-44fd-9dad-8d613a6e3545-config\") pod \"service-ca-operator-777779d784-v8bk6\" (UID: \"1d13c714-5451-44fd-9dad-8d613a6e3545\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-v8bk6" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.852592 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-525rj\" (UniqueName: \"kubernetes.io/projected/f2080446-c5c3-4c09-b8ae-54359bbb72ba-kube-api-access-525rj\") pod \"multus-admission-controller-857f4d67dd-vf46x\" (UID: \"f2080446-c5c3-4c09-b8ae-54359bbb72ba\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-vf46x" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.852615 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jc9p7\" (UniqueName: \"kubernetes.io/projected/027b031c-1b21-40e8-aa7c-4a342aae265a-kube-api-access-jc9p7\") pod \"apiserver-76f77b778f-s7tww\" (UID: \"027b031c-1b21-40e8-aa7c-4a342aae265a\") " pod="openshift-apiserver/apiserver-76f77b778f-s7tww" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.852648 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/027b031c-1b21-40e8-aa7c-4a342aae265a-config\") pod \"apiserver-76f77b778f-s7tww\" (UID: \"027b031c-1b21-40e8-aa7c-4a342aae265a\") " pod="openshift-apiserver/apiserver-76f77b778f-s7tww" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.852669 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/56521238-a49d-446a-b4a9-e2809ff15826-profile-collector-cert\") pod \"catalog-operator-68c6474976-4cx2q\" (UID: \"56521238-a49d-446a-b4a9-e2809ff15826\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4cx2q" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.852690 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c14a1c93-8be4-42b7-aae1-9c31d53a6bfc-config-volume\") pod \"collect-profiles-29395560-pq7jf\" (UID: \"c14a1c93-8be4-42b7-aae1-9c31d53a6bfc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29395560-pq7jf" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.852712 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c14a1c93-8be4-42b7-aae1-9c31d53a6bfc-secret-volume\") pod \"collect-profiles-29395560-pq7jf\" (UID: \"c14a1c93-8be4-42b7-aae1-9c31d53a6bfc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29395560-pq7jf" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.852735 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/027b031c-1b21-40e8-aa7c-4a342aae265a-node-pullsecrets\") pod \"apiserver-76f77b778f-s7tww\" (UID: \"027b031c-1b21-40e8-aa7c-4a342aae265a\") " pod="openshift-apiserver/apiserver-76f77b778f-s7tww" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.852759 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/027b031c-1b21-40e8-aa7c-4a342aae265a-audit-dir\") pod \"apiserver-76f77b778f-s7tww\" (UID: \"027b031c-1b21-40e8-aa7c-4a342aae265a\") " pod="openshift-apiserver/apiserver-76f77b778f-s7tww" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.852781 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/85811958-7716-47c2-b9ad-9392647bb172-certs\") pod \"machine-config-server-bqktx\" (UID: \"85811958-7716-47c2-b9ad-9392647bb172\") " pod="openshift-machine-config-operator/machine-config-server-bqktx" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.852805 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f7d61c37-621e-4812-bbaa-74006927176a-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-74phw\" (UID: \"f7d61c37-621e-4812-bbaa-74006927176a\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-74phw" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.852841 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b0b1043c-75c9-4643-a400-a5c009dfb5b6-cert\") pod \"ingress-canary-fqd92\" (UID: \"b0b1043c-75c9-4643-a400-a5c009dfb5b6\") " pod="openshift-ingress-canary/ingress-canary-fqd92" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.852865 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j2ddt\" (UniqueName: \"kubernetes.io/projected/1d13c714-5451-44fd-9dad-8d613a6e3545-kube-api-access-j2ddt\") pod \"service-ca-operator-777779d784-v8bk6\" (UID: \"1d13c714-5451-44fd-9dad-8d613a6e3545\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-v8bk6" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.852891 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/20a1d330-d99c-480d-b44c-4d8a5b4c6453-metrics-tls\") pod \"dns-default-lq8bp\" (UID: \"20a1d330-d99c-480d-b44c-4d8a5b4c6453\") " pod="openshift-dns/dns-default-lq8bp" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.852921 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/f7d61c37-621e-4812-bbaa-74006927176a-proxy-tls\") pod \"machine-config-controller-84d6567774-74phw\" (UID: \"f7d61c37-621e-4812-bbaa-74006927176a\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-74phw" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.852946 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4l4r5\" (UniqueName: \"kubernetes.io/projected/6a9aee9a-c47c-4ad2-9431-1a4ac5ce15fd-kube-api-access-4l4r5\") pod \"control-plane-machine-set-operator-78cbb6b69f-nl2sk\" (UID: \"6a9aee9a-c47c-4ad2-9431-1a4ac5ce15fd\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-nl2sk" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.852967 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/11301ff8-fa01-48da-a95e-fc244fd0a575-tmpfs\") pod \"packageserver-d55dfcdfc-6sfv4\" (UID: \"11301ff8-fa01-48da-a95e-fc244fd0a575\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6sfv4" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.852988 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tv5cm\" (UniqueName: \"kubernetes.io/projected/20a1d330-d99c-480d-b44c-4d8a5b4c6453-kube-api-access-tv5cm\") pod \"dns-default-lq8bp\" (UID: \"20a1d330-d99c-480d-b44c-4d8a5b4c6453\") " pod="openshift-dns/dns-default-lq8bp" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.853011 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ec19858a-32ef-443f-b307-df94c5d3e8de-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-5prl8\" (UID: \"ec19858a-32ef-443f-b307-df94c5d3e8de\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5prl8" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.853035 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/4c817f5a-6f7c-48d3-9d45-bdc401646bed-images\") pod \"machine-config-operator-74547568cd-kf7zh\" (UID: \"4c817f5a-6f7c-48d3-9d45-bdc401646bed\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kf7zh" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.853057 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b24e87a6-3cd8-4a37-b537-e9a5332cb6c7-srv-cert\") pod \"olm-operator-6b444d44fb-h59mj\" (UID: \"b24e87a6-3cd8-4a37-b537-e9a5332cb6c7\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h59mj" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.853079 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8x66f\" (UniqueName: \"kubernetes.io/projected/f7d61c37-621e-4812-bbaa-74006927176a-kube-api-access-8x66f\") pod \"machine-config-controller-84d6567774-74phw\" (UID: \"f7d61c37-621e-4812-bbaa-74006927176a\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-74phw" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.853236 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cq46c\" (UniqueName: \"kubernetes.io/projected/78db7e2e-ba45-464e-9b1d-0fc61191955a-kube-api-access-cq46c\") pod \"csi-hostpathplugin-bjdjf\" (UID: \"78db7e2e-ba45-464e-9b1d-0fc61191955a\") " pod="hostpath-provisioner/csi-hostpathplugin-bjdjf" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.853261 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/4c817f5a-6f7c-48d3-9d45-bdc401646bed-proxy-tls\") pod \"machine-config-operator-74547568cd-kf7zh\" (UID: \"4c817f5a-6f7c-48d3-9d45-bdc401646bed\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kf7zh" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.853288 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ec19858a-32ef-443f-b307-df94c5d3e8de-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-5prl8\" (UID: \"ec19858a-32ef-443f-b307-df94c5d3e8de\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5prl8" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.853336 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/027b031c-1b21-40e8-aa7c-4a342aae265a-trusted-ca-bundle\") pod \"apiserver-76f77b778f-s7tww\" (UID: \"027b031c-1b21-40e8-aa7c-4a342aae265a\") " pod="openshift-apiserver/apiserver-76f77b778f-s7tww" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.853360 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9ckq9\" (UniqueName: \"kubernetes.io/projected/b0b1043c-75c9-4643-a400-a5c009dfb5b6-kube-api-access-9ckq9\") pod \"ingress-canary-fqd92\" (UID: \"b0b1043c-75c9-4643-a400-a5c009dfb5b6\") " pod="openshift-ingress-canary/ingress-canary-fqd92" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.853384 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e128526a-8489-47e1-a711-4977f2b0d46d-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-z9hjh\" (UID: \"e128526a-8489-47e1-a711-4977f2b0d46d\") " pod="openshift-marketplace/marketplace-operator-79b997595-z9hjh" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.853407 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-47p67\" (UniqueName: \"kubernetes.io/projected/11301ff8-fa01-48da-a95e-fc244fd0a575-kube-api-access-47p67\") pod \"packageserver-d55dfcdfc-6sfv4\" (UID: \"11301ff8-fa01-48da-a95e-fc244fd0a575\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6sfv4" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.853442 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gz5nl\" (UniqueName: \"kubernetes.io/projected/56521238-a49d-446a-b4a9-e2809ff15826-kube-api-access-gz5nl\") pod \"catalog-operator-68c6474976-4cx2q\" (UID: \"56521238-a49d-446a-b4a9-e2809ff15826\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4cx2q" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.853464 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9b76t\" (UniqueName: \"kubernetes.io/projected/b24e87a6-3cd8-4a37-b537-e9a5332cb6c7-kube-api-access-9b76t\") pod \"olm-operator-6b444d44fb-h59mj\" (UID: \"b24e87a6-3cd8-4a37-b537-e9a5332cb6c7\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h59mj" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.853485 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/027b031c-1b21-40e8-aa7c-4a342aae265a-audit\") pod \"apiserver-76f77b778f-s7tww\" (UID: \"027b031c-1b21-40e8-aa7c-4a342aae265a\") " pod="openshift-apiserver/apiserver-76f77b778f-s7tww" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.853539 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/e128526a-8489-47e1-a711-4977f2b0d46d-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-z9hjh\" (UID: \"e128526a-8489-47e1-a711-4977f2b0d46d\") " pod="openshift-marketplace/marketplace-operator-79b997595-z9hjh" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.853564 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mksx9\" (UniqueName: \"kubernetes.io/projected/6d1fd716-6991-43b6-9f1f-268576e864a0-kube-api-access-mksx9\") pod \"service-ca-9c57cc56f-b9s66\" (UID: \"6d1fd716-6991-43b6-9f1f-268576e864a0\") " pod="openshift-service-ca/service-ca-9c57cc56f-b9s66" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.853585 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/11301ff8-fa01-48da-a95e-fc244fd0a575-webhook-cert\") pod \"packageserver-d55dfcdfc-6sfv4\" (UID: \"11301ff8-fa01-48da-a95e-fc244fd0a575\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6sfv4" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.853608 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/78db7e2e-ba45-464e-9b1d-0fc61191955a-mountpoint-dir\") pod \"csi-hostpathplugin-bjdjf\" (UID: \"78db7e2e-ba45-464e-9b1d-0fc61191955a\") " pod="hostpath-provisioner/csi-hostpathplugin-bjdjf" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.853630 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/11301ff8-fa01-48da-a95e-fc244fd0a575-apiservice-cert\") pod \"packageserver-d55dfcdfc-6sfv4\" (UID: \"11301ff8-fa01-48da-a95e-fc244fd0a575\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6sfv4" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.853654 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/78db7e2e-ba45-464e-9b1d-0fc61191955a-csi-data-dir\") pod \"csi-hostpathplugin-bjdjf\" (UID: \"78db7e2e-ba45-464e-9b1d-0fc61191955a\") " pod="hostpath-provisioner/csi-hostpathplugin-bjdjf" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.853678 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6a9aee9a-c47c-4ad2-9431-1a4ac5ce15fd-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-nl2sk\" (UID: \"6a9aee9a-c47c-4ad2-9431-1a4ac5ce15fd\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-nl2sk" Nov 21 14:11:10 crc kubenswrapper[4897]: E1121 14:11:10.853958 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-21 14:11:11.353932097 +0000 UTC m=+148.638525572 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.854265 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/6d1fd716-6991-43b6-9f1f-268576e864a0-signing-cabundle\") pod \"service-ca-9c57cc56f-b9s66\" (UID: \"6d1fd716-6991-43b6-9f1f-268576e864a0\") " pod="openshift-service-ca/service-ca-9c57cc56f-b9s66" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.855014 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/027b031c-1b21-40e8-aa7c-4a342aae265a-image-import-ca\") pod \"apiserver-76f77b778f-s7tww\" (UID: \"027b031c-1b21-40e8-aa7c-4a342aae265a\") " pod="openshift-apiserver/apiserver-76f77b778f-s7tww" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.855695 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/78db7e2e-ba45-464e-9b1d-0fc61191955a-csi-data-dir\") pod \"csi-hostpathplugin-bjdjf\" (UID: \"78db7e2e-ba45-464e-9b1d-0fc61191955a\") " pod="hostpath-provisioner/csi-hostpathplugin-bjdjf" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.857524 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/78db7e2e-ba45-464e-9b1d-0fc61191955a-plugins-dir\") pod \"csi-hostpathplugin-bjdjf\" (UID: \"78db7e2e-ba45-464e-9b1d-0fc61191955a\") " pod="hostpath-provisioner/csi-hostpathplugin-bjdjf" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.857770 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1d13c714-5451-44fd-9dad-8d613a6e3545-config\") pod \"service-ca-operator-777779d784-v8bk6\" (UID: \"1d13c714-5451-44fd-9dad-8d613a6e3545\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-v8bk6" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.858046 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c14a1c93-8be4-42b7-aae1-9c31d53a6bfc-config-volume\") pod \"collect-profiles-29395560-pq7jf\" (UID: \"c14a1c93-8be4-42b7-aae1-9c31d53a6bfc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29395560-pq7jf" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.858652 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/11301ff8-fa01-48da-a95e-fc244fd0a575-webhook-cert\") pod \"packageserver-d55dfcdfc-6sfv4\" (UID: \"11301ff8-fa01-48da-a95e-fc244fd0a575\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6sfv4" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.858808 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec19858a-32ef-443f-b307-df94c5d3e8de-config\") pod \"kube-controller-manager-operator-78b949d7b-5prl8\" (UID: \"ec19858a-32ef-443f-b307-df94c5d3e8de\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5prl8" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.859696 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/20a1d330-d99c-480d-b44c-4d8a5b4c6453-config-volume\") pod \"dns-default-lq8bp\" (UID: \"20a1d330-d99c-480d-b44c-4d8a5b4c6453\") " pod="openshift-dns/dns-default-lq8bp" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.859756 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/78db7e2e-ba45-464e-9b1d-0fc61191955a-mountpoint-dir\") pod \"csi-hostpathplugin-bjdjf\" (UID: \"78db7e2e-ba45-464e-9b1d-0fc61191955a\") " pod="hostpath-provisioner/csi-hostpathplugin-bjdjf" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.859941 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6a9aee9a-c47c-4ad2-9431-1a4ac5ce15fd-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-nl2sk\" (UID: \"6a9aee9a-c47c-4ad2-9431-1a4ac5ce15fd\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-nl2sk" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.860200 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/78db7e2e-ba45-464e-9b1d-0fc61191955a-registration-dir\") pod \"csi-hostpathplugin-bjdjf\" (UID: \"78db7e2e-ba45-464e-9b1d-0fc61191955a\") " pod="hostpath-provisioner/csi-hostpathplugin-bjdjf" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.860266 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/027b031c-1b21-40e8-aa7c-4a342aae265a-audit\") pod \"apiserver-76f77b778f-s7tww\" (UID: \"027b031c-1b21-40e8-aa7c-4a342aae265a\") " pod="openshift-apiserver/apiserver-76f77b778f-s7tww" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.860428 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/027b031c-1b21-40e8-aa7c-4a342aae265a-trusted-ca-bundle\") pod \"apiserver-76f77b778f-s7tww\" (UID: \"027b031c-1b21-40e8-aa7c-4a342aae265a\") " pod="openshift-apiserver/apiserver-76f77b778f-s7tww" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.861236 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/20a1d330-d99c-480d-b44c-4d8a5b4c6453-metrics-tls\") pod \"dns-default-lq8bp\" (UID: \"20a1d330-d99c-480d-b44c-4d8a5b4c6453\") " pod="openshift-dns/dns-default-lq8bp" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.862254 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/027b031c-1b21-40e8-aa7c-4a342aae265a-node-pullsecrets\") pod \"apiserver-76f77b778f-s7tww\" (UID: \"027b031c-1b21-40e8-aa7c-4a342aae265a\") " pod="openshift-apiserver/apiserver-76f77b778f-s7tww" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.862319 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/027b031c-1b21-40e8-aa7c-4a342aae265a-etcd-client\") pod \"apiserver-76f77b778f-s7tww\" (UID: \"027b031c-1b21-40e8-aa7c-4a342aae265a\") " pod="openshift-apiserver/apiserver-76f77b778f-s7tww" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.862393 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/027b031c-1b21-40e8-aa7c-4a342aae265a-audit-dir\") pod \"apiserver-76f77b778f-s7tww\" (UID: \"027b031c-1b21-40e8-aa7c-4a342aae265a\") " pod="openshift-apiserver/apiserver-76f77b778f-s7tww" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.862975 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/027b031c-1b21-40e8-aa7c-4a342aae265a-config\") pod \"apiserver-76f77b778f-s7tww\" (UID: \"027b031c-1b21-40e8-aa7c-4a342aae265a\") " pod="openshift-apiserver/apiserver-76f77b778f-s7tww" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.863322 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/e128526a-8489-47e1-a711-4977f2b0d46d-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-z9hjh\" (UID: \"e128526a-8489-47e1-a711-4977f2b0d46d\") " pod="openshift-marketplace/marketplace-operator-79b997595-z9hjh" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.865468 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/56521238-a49d-446a-b4a9-e2809ff15826-srv-cert\") pod \"catalog-operator-68c6474976-4cx2q\" (UID: \"56521238-a49d-446a-b4a9-e2809ff15826\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4cx2q" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.865569 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f7d61c37-621e-4812-bbaa-74006927176a-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-74phw\" (UID: \"f7d61c37-621e-4812-bbaa-74006927176a\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-74phw" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.866259 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b24e87a6-3cd8-4a37-b537-e9a5332cb6c7-srv-cert\") pod \"olm-operator-6b444d44fb-h59mj\" (UID: \"b24e87a6-3cd8-4a37-b537-e9a5332cb6c7\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h59mj" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.866311 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/78db7e2e-ba45-464e-9b1d-0fc61191955a-socket-dir\") pod \"csi-hostpathplugin-bjdjf\" (UID: \"78db7e2e-ba45-464e-9b1d-0fc61191955a\") " pod="hostpath-provisioner/csi-hostpathplugin-bjdjf" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.866346 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/4c817f5a-6f7c-48d3-9d45-bdc401646bed-auth-proxy-config\") pod \"machine-config-operator-74547568cd-kf7zh\" (UID: \"4c817f5a-6f7c-48d3-9d45-bdc401646bed\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kf7zh" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.866717 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/11301ff8-fa01-48da-a95e-fc244fd0a575-tmpfs\") pod \"packageserver-d55dfcdfc-6sfv4\" (UID: \"11301ff8-fa01-48da-a95e-fc244fd0a575\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6sfv4" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.866955 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/4c817f5a-6f7c-48d3-9d45-bdc401646bed-images\") pod \"machine-config-operator-74547568cd-kf7zh\" (UID: \"4c817f5a-6f7c-48d3-9d45-bdc401646bed\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kf7zh" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.867528 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/027b031c-1b21-40e8-aa7c-4a342aae265a-serving-cert\") pod \"apiserver-76f77b778f-s7tww\" (UID: \"027b031c-1b21-40e8-aa7c-4a342aae265a\") " pod="openshift-apiserver/apiserver-76f77b778f-s7tww" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.876248 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e128526a-8489-47e1-a711-4977f2b0d46d-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-z9hjh\" (UID: \"e128526a-8489-47e1-a711-4977f2b0d46d\") " pod="openshift-marketplace/marketplace-operator-79b997595-z9hjh" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.876781 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/f7d61c37-621e-4812-bbaa-74006927176a-proxy-tls\") pod \"machine-config-controller-84d6567774-74phw\" (UID: \"f7d61c37-621e-4812-bbaa-74006927176a\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-74phw" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.877391 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/11301ff8-fa01-48da-a95e-fc244fd0a575-apiservice-cert\") pod \"packageserver-d55dfcdfc-6sfv4\" (UID: \"11301ff8-fa01-48da-a95e-fc244fd0a575\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6sfv4" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.877865 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/85811958-7716-47c2-b9ad-9392647bb172-certs\") pod \"machine-config-server-bqktx\" (UID: \"85811958-7716-47c2-b9ad-9392647bb172\") " pod="openshift-machine-config-operator/machine-config-server-bqktx" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.878348 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/027b031c-1b21-40e8-aa7c-4a342aae265a-encryption-config\") pod \"apiserver-76f77b778f-s7tww\" (UID: \"027b031c-1b21-40e8-aa7c-4a342aae265a\") " pod="openshift-apiserver/apiserver-76f77b778f-s7tww" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.878953 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/85811958-7716-47c2-b9ad-9392647bb172-node-bootstrap-token\") pod \"machine-config-server-bqktx\" (UID: \"85811958-7716-47c2-b9ad-9392647bb172\") " pod="openshift-machine-config-operator/machine-config-server-bqktx" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.879317 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ec19858a-32ef-443f-b307-df94c5d3e8de-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-5prl8\" (UID: \"ec19858a-32ef-443f-b307-df94c5d3e8de\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5prl8" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.879437 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/6d1fd716-6991-43b6-9f1f-268576e864a0-signing-key\") pod \"service-ca-9c57cc56f-b9s66\" (UID: \"6d1fd716-6991-43b6-9f1f-268576e864a0\") " pod="openshift-service-ca/service-ca-9c57cc56f-b9s66" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.879919 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/56521238-a49d-446a-b4a9-e2809ff15826-profile-collector-cert\") pod \"catalog-operator-68c6474976-4cx2q\" (UID: \"56521238-a49d-446a-b4a9-e2809ff15826\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4cx2q" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.880178 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-86gsj\" (UniqueName: \"kubernetes.io/projected/9f63b8dd-ffb7-4963-8fcc-0c133a2c045b-kube-api-access-86gsj\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.880456 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b24e87a6-3cd8-4a37-b537-e9a5332cb6c7-profile-collector-cert\") pod \"olm-operator-6b444d44fb-h59mj\" (UID: \"b24e87a6-3cd8-4a37-b537-e9a5332cb6c7\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h59mj" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.880554 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1d13c714-5451-44fd-9dad-8d613a6e3545-serving-cert\") pod \"service-ca-operator-777779d784-v8bk6\" (UID: \"1d13c714-5451-44fd-9dad-8d613a6e3545\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-v8bk6" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.880968 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/f2080446-c5c3-4c09-b8ae-54359bbb72ba-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-vf46x\" (UID: \"f2080446-c5c3-4c09-b8ae-54359bbb72ba\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-vf46x" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.883145 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b0b1043c-75c9-4643-a400-a5c009dfb5b6-cert\") pod \"ingress-canary-fqd92\" (UID: \"b0b1043c-75c9-4643-a400-a5c009dfb5b6\") " pod="openshift-ingress-canary/ingress-canary-fqd92" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.884353 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/5262754c-6c80-4da0-8d9f-a58d63cfeadb-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-cc48c\" (UID: \"5262754c-6c80-4da0-8d9f-a58d63cfeadb\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-cc48c" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.884377 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c14a1c93-8be4-42b7-aae1-9c31d53a6bfc-secret-volume\") pod \"collect-profiles-29395560-pq7jf\" (UID: \"c14a1c93-8be4-42b7-aae1-9c31d53a6bfc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29395560-pq7jf" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.884413 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/4c817f5a-6f7c-48d3-9d45-bdc401646bed-proxy-tls\") pod \"machine-config-operator-74547568cd-kf7zh\" (UID: \"4c817f5a-6f7c-48d3-9d45-bdc401646bed\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kf7zh" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.887812 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-vgwv2" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.891045 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6jwzn\" (UniqueName: \"kubernetes.io/projected/5648ea03-bf08-419d-87fb-6287e11dd399-kube-api-access-6jwzn\") pod \"ingress-operator-5b745b69d9-2hmqv\" (UID: \"5648ea03-bf08-419d-87fb-6287e11dd399\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2hmqv" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.901909 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lsdt5\" (UniqueName: \"kubernetes.io/projected/8342c8a2-87f4-4bf9-a2ca-073c02b277ad-kube-api-access-lsdt5\") pod \"console-f9d7485db-7fsr9\" (UID: \"8342c8a2-87f4-4bf9-a2ca-073c02b277ad\") " pod="openshift-console/console-f9d7485db-7fsr9" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.902157 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-7fsr9" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.909003 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-t5jsz" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.910174 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/027b031c-1b21-40e8-aa7c-4a342aae265a-etcd-serving-ca\") pod \"apiserver-76f77b778f-s7tww\" (UID: \"027b031c-1b21-40e8-aa7c-4a342aae265a\") " pod="openshift-apiserver/apiserver-76f77b778f-s7tww" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.923911 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/98941613-bca1-4b35-9dba-547294a1feb5-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-lp8th\" (UID: \"98941613-bca1-4b35-9dba-547294a1feb5\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-lp8th" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.933771 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-szznp" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.941034 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-lp8th" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.949372 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5648ea03-bf08-419d-87fb-6287e11dd399-bound-sa-token\") pod \"ingress-operator-5b745b69d9-2hmqv\" (UID: \"5648ea03-bf08-419d-87fb-6287e11dd399\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2hmqv" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.955729 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:10 crc kubenswrapper[4897]: E1121 14:11:10.957952 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-21 14:11:11.457936818 +0000 UTC m=+148.742530293 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5z687" (UID: "9f63b8dd-ffb7-4963-8fcc-0c133a2c045b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.966710 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9f63b8dd-ffb7-4963-8fcc-0c133a2c045b-bound-sa-token\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:10 crc kubenswrapper[4897]: I1121 14:11:10.986923 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-2kgg2" Nov 21 14:11:11 crc kubenswrapper[4897]: I1121 14:11:11.005779 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2xp5\" (UniqueName: \"kubernetes.io/projected/4c817f5a-6f7c-48d3-9d45-bdc401646bed-kube-api-access-s2xp5\") pod \"machine-config-operator-74547568cd-kf7zh\" (UID: \"4c817f5a-6f7c-48d3-9d45-bdc401646bed\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kf7zh" Nov 21 14:11:11 crc kubenswrapper[4897]: I1121 14:11:11.026393 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-525rj\" (UniqueName: \"kubernetes.io/projected/f2080446-c5c3-4c09-b8ae-54359bbb72ba-kube-api-access-525rj\") pod \"multus-admission-controller-857f4d67dd-vf46x\" (UID: \"f2080446-c5c3-4c09-b8ae-54359bbb72ba\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-vf46x" Nov 21 14:11:11 crc kubenswrapper[4897]: I1121 14:11:11.037697 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kf7zh" Nov 21 14:11:11 crc kubenswrapper[4897]: I1121 14:11:11.043350 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ec19858a-32ef-443f-b307-df94c5d3e8de-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-5prl8\" (UID: \"ec19858a-32ef-443f-b307-df94c5d3e8de\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5prl8" Nov 21 14:11:11 crc kubenswrapper[4897]: I1121 14:11:11.056691 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 21 14:11:11 crc kubenswrapper[4897]: E1121 14:11:11.057124 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-21 14:11:11.557104262 +0000 UTC m=+148.841697737 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:11 crc kubenswrapper[4897]: I1121 14:11:11.064708 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9ckq9\" (UniqueName: \"kubernetes.io/projected/b0b1043c-75c9-4643-a400-a5c009dfb5b6-kube-api-access-9ckq9\") pod \"ingress-canary-fqd92\" (UID: \"b0b1043c-75c9-4643-a400-a5c009dfb5b6\") " pod="openshift-ingress-canary/ingress-canary-fqd92" Nov 21 14:11:11 crc kubenswrapper[4897]: I1121 14:11:11.084382 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-47p67\" (UniqueName: \"kubernetes.io/projected/11301ff8-fa01-48da-a95e-fc244fd0a575-kube-api-access-47p67\") pod \"packageserver-d55dfcdfc-6sfv4\" (UID: \"11301ff8-fa01-48da-a95e-fc244fd0a575\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6sfv4" Nov 21 14:11:11 crc kubenswrapper[4897]: I1121 14:11:11.105388 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8x66f\" (UniqueName: \"kubernetes.io/projected/f7d61c37-621e-4812-bbaa-74006927176a-kube-api-access-8x66f\") pod \"machine-config-controller-84d6567774-74phw\" (UID: \"f7d61c37-621e-4812-bbaa-74006927176a\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-74phw" Nov 21 14:11:11 crc kubenswrapper[4897]: I1121 14:11:11.121914 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-fqd92" Nov 21 14:11:11 crc kubenswrapper[4897]: I1121 14:11:11.134855 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gz5nl\" (UniqueName: \"kubernetes.io/projected/56521238-a49d-446a-b4a9-e2809ff15826-kube-api-access-gz5nl\") pod \"catalog-operator-68c6474976-4cx2q\" (UID: \"56521238-a49d-446a-b4a9-e2809ff15826\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4cx2q" Nov 21 14:11:11 crc kubenswrapper[4897]: I1121 14:11:11.146154 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cq46c\" (UniqueName: \"kubernetes.io/projected/78db7e2e-ba45-464e-9b1d-0fc61191955a-kube-api-access-cq46c\") pod \"csi-hostpathplugin-bjdjf\" (UID: \"78db7e2e-ba45-464e-9b1d-0fc61191955a\") " pod="hostpath-provisioner/csi-hostpathplugin-bjdjf" Nov 21 14:11:11 crc kubenswrapper[4897]: I1121 14:11:11.160349 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:11 crc kubenswrapper[4897]: E1121 14:11:11.160773 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-21 14:11:11.660757493 +0000 UTC m=+148.945350968 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5z687" (UID: "9f63b8dd-ffb7-4963-8fcc-0c133a2c045b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:11 crc kubenswrapper[4897]: I1121 14:11:11.171139 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9b76t\" (UniqueName: \"kubernetes.io/projected/b24e87a6-3cd8-4a37-b537-e9a5332cb6c7-kube-api-access-9b76t\") pod \"olm-operator-6b444d44fb-h59mj\" (UID: \"b24e87a6-3cd8-4a37-b537-e9a5332cb6c7\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h59mj" Nov 21 14:11:11 crc kubenswrapper[4897]: I1121 14:11:11.179790 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-vgwv2"] Nov 21 14:11:11 crc kubenswrapper[4897]: I1121 14:11:11.181486 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tv5cm\" (UniqueName: \"kubernetes.io/projected/20a1d330-d99c-480d-b44c-4d8a5b4c6453-kube-api-access-tv5cm\") pod \"dns-default-lq8bp\" (UID: \"20a1d330-d99c-480d-b44c-4d8a5b4c6453\") " pod="openshift-dns/dns-default-lq8bp" Nov 21 14:11:11 crc kubenswrapper[4897]: I1121 14:11:11.212917 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2wp82\" (UniqueName: \"kubernetes.io/projected/e128526a-8489-47e1-a711-4977f2b0d46d-kube-api-access-2wp82\") pod \"marketplace-operator-79b997595-z9hjh\" (UID: \"e128526a-8489-47e1-a711-4977f2b0d46d\") " pod="openshift-marketplace/marketplace-operator-79b997595-z9hjh" Nov 21 14:11:11 crc kubenswrapper[4897]: W1121 14:11:11.223274 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod82f81ac2_d2b2_49de_82f3_a51796dbff03.slice/crio-276f5c0ba20b2eb4081dc5dd4edfe75b18071d42f697ac11c01c9e43596ab576 WatchSource:0}: Error finding container 276f5c0ba20b2eb4081dc5dd4edfe75b18071d42f697ac11c01c9e43596ab576: Status 404 returned error can't find the container with id 276f5c0ba20b2eb4081dc5dd4edfe75b18071d42f697ac11c01c9e43596ab576 Nov 21 14:11:11 crc kubenswrapper[4897]: I1121 14:11:11.226555 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z588q\" (UniqueName: \"kubernetes.io/projected/5262754c-6c80-4da0-8d9f-a58d63cfeadb-kube-api-access-z588q\") pod \"package-server-manager-789f6589d5-cc48c\" (UID: \"5262754c-6c80-4da0-8d9f-a58d63cfeadb\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-cc48c" Nov 21 14:11:11 crc kubenswrapper[4897]: I1121 14:11:11.246026 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jc9p7\" (UniqueName: \"kubernetes.io/projected/027b031c-1b21-40e8-aa7c-4a342aae265a-kube-api-access-jc9p7\") pod \"apiserver-76f77b778f-s7tww\" (UID: \"027b031c-1b21-40e8-aa7c-4a342aae265a\") " pod="openshift-apiserver/apiserver-76f77b778f-s7tww" Nov 21 14:11:11 crc kubenswrapper[4897]: I1121 14:11:11.248936 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2hmqv" Nov 21 14:11:11 crc kubenswrapper[4897]: I1121 14:11:11.261364 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 21 14:11:11 crc kubenswrapper[4897]: E1121 14:11:11.261767 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-21 14:11:11.761745908 +0000 UTC m=+149.046339383 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:11 crc kubenswrapper[4897]: I1121 14:11:11.261764 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5prl8" Nov 21 14:11:11 crc kubenswrapper[4897]: I1121 14:11:11.262418 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j2ddt\" (UniqueName: \"kubernetes.io/projected/1d13c714-5451-44fd-9dad-8d613a6e3545-kube-api-access-j2ddt\") pod \"service-ca-operator-777779d784-v8bk6\" (UID: \"1d13c714-5451-44fd-9dad-8d613a6e3545\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-v8bk6" Nov 21 14:11:11 crc kubenswrapper[4897]: I1121 14:11:11.271172 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4cx2q" Nov 21 14:11:11 crc kubenswrapper[4897]: I1121 14:11:11.276922 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-v8bk6" Nov 21 14:11:11 crc kubenswrapper[4897]: I1121 14:11:11.285637 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wts2z\" (UniqueName: \"kubernetes.io/projected/85811958-7716-47c2-b9ad-9392647bb172-kube-api-access-wts2z\") pod \"machine-config-server-bqktx\" (UID: \"85811958-7716-47c2-b9ad-9392647bb172\") " pod="openshift-machine-config-operator/machine-config-server-bqktx" Nov 21 14:11:11 crc kubenswrapper[4897]: I1121 14:11:11.293922 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h59mj" Nov 21 14:11:11 crc kubenswrapper[4897]: I1121 14:11:11.299207 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-74phw" Nov 21 14:11:11 crc kubenswrapper[4897]: I1121 14:11:11.312741 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mksx9\" (UniqueName: \"kubernetes.io/projected/6d1fd716-6991-43b6-9f1f-268576e864a0-kube-api-access-mksx9\") pod \"service-ca-9c57cc56f-b9s66\" (UID: \"6d1fd716-6991-43b6-9f1f-268576e864a0\") " pod="openshift-service-ca/service-ca-9c57cc56f-b9s66" Nov 21 14:11:11 crc kubenswrapper[4897]: I1121 14:11:11.313186 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-cc48c" Nov 21 14:11:11 crc kubenswrapper[4897]: I1121 14:11:11.320129 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-vf46x" Nov 21 14:11:11 crc kubenswrapper[4897]: I1121 14:11:11.324959 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-666j9\" (UniqueName: \"kubernetes.io/projected/c14a1c93-8be4-42b7-aae1-9c31d53a6bfc-kube-api-access-666j9\") pod \"collect-profiles-29395560-pq7jf\" (UID: \"c14a1c93-8be4-42b7-aae1-9c31d53a6bfc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29395560-pq7jf" Nov 21 14:11:11 crc kubenswrapper[4897]: I1121 14:11:11.330645 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6sfv4" Nov 21 14:11:11 crc kubenswrapper[4897]: I1121 14:11:11.347373 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4l4r5\" (UniqueName: \"kubernetes.io/projected/6a9aee9a-c47c-4ad2-9431-1a4ac5ce15fd-kube-api-access-4l4r5\") pod \"control-plane-machine-set-operator-78cbb6b69f-nl2sk\" (UID: \"6a9aee9a-c47c-4ad2-9431-1a4ac5ce15fd\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-nl2sk" Nov 21 14:11:11 crc kubenswrapper[4897]: I1121 14:11:11.355311 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-z9hjh" Nov 21 14:11:11 crc kubenswrapper[4897]: I1121 14:11:11.363190 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29395560-pq7jf" Nov 21 14:11:11 crc kubenswrapper[4897]: I1121 14:11:11.366288 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:11 crc kubenswrapper[4897]: E1121 14:11:11.366902 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-21 14:11:11.866886271 +0000 UTC m=+149.151479746 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5z687" (UID: "9f63b8dd-ffb7-4963-8fcc-0c133a2c045b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:11 crc kubenswrapper[4897]: I1121 14:11:11.369976 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-s7tww" Nov 21 14:11:11 crc kubenswrapper[4897]: I1121 14:11:11.378734 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-bqktx" Nov 21 14:11:11 crc kubenswrapper[4897]: I1121 14:11:11.384991 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-lq8bp" Nov 21 14:11:11 crc kubenswrapper[4897]: I1121 14:11:11.416402 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-bjdjf" Nov 21 14:11:11 crc kubenswrapper[4897]: I1121 14:11:11.457441 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-7fsr9"] Nov 21 14:11:11 crc kubenswrapper[4897]: I1121 14:11:11.458892 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-lp8th"] Nov 21 14:11:11 crc kubenswrapper[4897]: I1121 14:11:11.472540 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-kf7zh"] Nov 21 14:11:11 crc kubenswrapper[4897]: I1121 14:11:11.472590 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 21 14:11:11 crc kubenswrapper[4897]: E1121 14:11:11.472886 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-21 14:11:11.972846337 +0000 UTC m=+149.257439812 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:11 crc kubenswrapper[4897]: I1121 14:11:11.472963 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:11 crc kubenswrapper[4897]: E1121 14:11:11.473309 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-21 14:11:11.973292179 +0000 UTC m=+149.257885654 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5z687" (UID: "9f63b8dd-ffb7-4963-8fcc-0c133a2c045b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:11 crc kubenswrapper[4897]: I1121 14:11:11.560485 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-fqd92"] Nov 21 14:11:11 crc kubenswrapper[4897]: I1121 14:11:11.574474 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 21 14:11:11 crc kubenswrapper[4897]: E1121 14:11:11.574654 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-21 14:11:12.074623925 +0000 UTC m=+149.359217400 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:11 crc kubenswrapper[4897]: I1121 14:11:11.575013 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:11 crc kubenswrapper[4897]: E1121 14:11:11.575414 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-21 14:11:12.075403226 +0000 UTC m=+149.359996901 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5z687" (UID: "9f63b8dd-ffb7-4963-8fcc-0c133a2c045b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:11 crc kubenswrapper[4897]: I1121 14:11:11.593050 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-szznp"] Nov 21 14:11:11 crc kubenswrapper[4897]: I1121 14:11:11.605812 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-b9s66" Nov 21 14:11:11 crc kubenswrapper[4897]: I1121 14:11:11.609359 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5prl8"] Nov 21 14:11:11 crc kubenswrapper[4897]: I1121 14:11:11.624411 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-t5jsz"] Nov 21 14:11:11 crc kubenswrapper[4897]: I1121 14:11:11.645170 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-nl2sk" Nov 21 14:11:11 crc kubenswrapper[4897]: I1121 14:11:11.678196 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 21 14:11:11 crc kubenswrapper[4897]: E1121 14:11:11.678586 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-21 14:11:12.178566794 +0000 UTC m=+149.463160269 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:11 crc kubenswrapper[4897]: W1121 14:11:11.710950 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod98941613_bca1_4b35_9dba_547294a1feb5.slice/crio-70b23a5f3b061f37289fb0134a50d4a5ffd5b8e2e04b697899e63a237b975d6a WatchSource:0}: Error finding container 70b23a5f3b061f37289fb0134a50d4a5ffd5b8e2e04b697899e63a237b975d6a: Status 404 returned error can't find the container with id 70b23a5f3b061f37289fb0134a50d4a5ffd5b8e2e04b697899e63a237b975d6a Nov 21 14:11:11 crc kubenswrapper[4897]: W1121 14:11:11.778643 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb76cdd1d_6a5d_42ac_8b03_1cc105eb1c1f.slice/crio-23b793c9cfc51a0777dc9adedb09aaebee7c7d7885c58af7e07ec7e399950b85 WatchSource:0}: Error finding container 23b793c9cfc51a0777dc9adedb09aaebee7c7d7885c58af7e07ec7e399950b85: Status 404 returned error can't find the container with id 23b793c9cfc51a0777dc9adedb09aaebee7c7d7885c58af7e07ec7e399950b85 Nov 21 14:11:11 crc kubenswrapper[4897]: I1121 14:11:11.779248 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:11 crc kubenswrapper[4897]: E1121 14:11:11.779563 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-21 14:11:12.279550319 +0000 UTC m=+149.564143794 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5z687" (UID: "9f63b8dd-ffb7-4963-8fcc-0c133a2c045b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:11 crc kubenswrapper[4897]: I1121 14:11:11.882065 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 21 14:11:11 crc kubenswrapper[4897]: E1121 14:11:11.883112 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-21 14:11:12.38248266 +0000 UTC m=+149.667076135 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:11 crc kubenswrapper[4897]: I1121 14:11:11.886078 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-8kddm" event={"ID":"a59737d6-7609-4077-8fb8-d925f25a65f8","Type":"ContainerStarted","Data":"6cb35dad21c8a1c9df06450bda4792e2031955bbc6239b25a4e9f66a115a8835"} Nov 21 14:11:11 crc kubenswrapper[4897]: I1121 14:11:11.887137 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p85cq" podStartSLOduration=123.88711433 podStartE2EDuration="2m3.88711433s" podCreationTimestamp="2025-11-21 14:09:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:11:11.844752103 +0000 UTC m=+149.129345588" watchObservedRunningTime="2025-11-21 14:11:11.88711433 +0000 UTC m=+149.171707805" Nov 21 14:11:11 crc kubenswrapper[4897]: I1121 14:11:11.893695 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-szznp" event={"ID":"b76cdd1d-6a5d-42ac-8b03-1cc105eb1c1f","Type":"ContainerStarted","Data":"23b793c9cfc51a0777dc9adedb09aaebee7c7d7885c58af7e07ec7e399950b85"} Nov 21 14:11:11 crc kubenswrapper[4897]: I1121 14:11:11.894628 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-v8bk6"] Nov 21 14:11:11 crc kubenswrapper[4897]: I1121 14:11:11.895597 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-7fsr9" event={"ID":"8342c8a2-87f4-4bf9-a2ca-073c02b277ad","Type":"ContainerStarted","Data":"59c4e9d029bb293f0c9708784213d9a311f8bc3f091225f8b32b8f91cd416fb6"} Nov 21 14:11:11 crc kubenswrapper[4897]: I1121 14:11:11.897740 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-2kgg2" event={"ID":"e095a18e-3a0b-4418-95a8-a2cc3de5bd7b","Type":"ContainerStarted","Data":"9b619a557b122bcc871179a93019eafa7f3ba7cfb6d7947d217b10b736796f59"} Nov 21 14:11:11 crc kubenswrapper[4897]: I1121 14:11:11.897795 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-2kgg2" event={"ID":"e095a18e-3a0b-4418-95a8-a2cc3de5bd7b","Type":"ContainerStarted","Data":"09327794588a566d17174e6c1dc3f81c8237f42eb800ca1df672f69f625592da"} Nov 21 14:11:11 crc kubenswrapper[4897]: I1121 14:11:11.904006 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8s4px" event={"ID":"cfacdafa-2513-4d81-9222-c4b4fd8f8c10","Type":"ContainerStarted","Data":"70886b8042c028ab9b49f8dc15f344199718a54e737a19dd8a67ab221a5b8e64"} Nov 21 14:11:11 crc kubenswrapper[4897]: I1121 14:11:11.905544 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-fqd92" event={"ID":"b0b1043c-75c9-4643-a400-a5c009dfb5b6","Type":"ContainerStarted","Data":"b471d3e80508388b162d2c30dababf085c32416e80c62d45b40a474fc025a2d2"} Nov 21 14:11:11 crc kubenswrapper[4897]: I1121 14:11:11.907848 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kf7zh" event={"ID":"4c817f5a-6f7c-48d3-9d45-bdc401646bed","Type":"ContainerStarted","Data":"bdfd8b87e354d401838be9e2d6e996febcc317ca54bd121b584284223eb9bd89"} Nov 21 14:11:11 crc kubenswrapper[4897]: I1121 14:11:11.919842 4897 generic.go:334] "Generic (PLEG): container finished" podID="911a0a7e-7a63-4b4a-8b74-6c21f8da7e31" containerID="c4eada5c380ab7e8c6d369e42b1f0ecd6c8d57b19064bf32f85533e93195f73b" exitCode=0 Nov 21 14:11:11 crc kubenswrapper[4897]: I1121 14:11:11.919941 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-d8pkd" event={"ID":"911a0a7e-7a63-4b4a-8b74-6c21f8da7e31","Type":"ContainerDied","Data":"c4eada5c380ab7e8c6d369e42b1f0ecd6c8d57b19064bf32f85533e93195f73b"} Nov 21 14:11:11 crc kubenswrapper[4897]: I1121 14:11:11.923357 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-2hmqv"] Nov 21 14:11:11 crc kubenswrapper[4897]: I1121 14:11:11.923393 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-lp8th" event={"ID":"98941613-bca1-4b35-9dba-547294a1feb5","Type":"ContainerStarted","Data":"70b23a5f3b061f37289fb0134a50d4a5ffd5b8e2e04b697899e63a237b975d6a"} Nov 21 14:11:11 crc kubenswrapper[4897]: I1121 14:11:11.930342 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-2trmp" event={"ID":"0e4b5f84-4954-4d85-8bae-db18ed352464","Type":"ContainerStarted","Data":"bb86b29391e447e7a95b1e90e43d0d41b3f73430bf10f101de634d819340dac0"} Nov 21 14:11:11 crc kubenswrapper[4897]: I1121 14:11:11.930851 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-2trmp" Nov 21 14:11:11 crc kubenswrapper[4897]: I1121 14:11:11.931924 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-vgwv2" event={"ID":"82f81ac2-d2b2-49de-82f3-a51796dbff03","Type":"ContainerStarted","Data":"276f5c0ba20b2eb4081dc5dd4edfe75b18071d42f697ac11c01c9e43596ab576"} Nov 21 14:11:11 crc kubenswrapper[4897]: I1121 14:11:11.933075 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-hk7mf" event={"ID":"69303f47-f9a7-4c1c-b569-b1e4de56b2ff","Type":"ContainerStarted","Data":"362289d158035abdcd8346f054fd0db9eeb33b35b681a3778c6c06fb0f49493d"} Nov 21 14:11:11 crc kubenswrapper[4897]: I1121 14:11:11.939733 4897 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-2trmp container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.13:6443/healthz\": dial tcp 10.217.0.13:6443: connect: connection refused" start-of-body= Nov 21 14:11:11 crc kubenswrapper[4897]: I1121 14:11:11.939779 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-2trmp" podUID="0e4b5f84-4954-4d85-8bae-db18ed352464" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.13:6443/healthz\": dial tcp 10.217.0.13:6443: connect: connection refused" Nov 21 14:11:11 crc kubenswrapper[4897]: I1121 14:11:11.943083 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rv2nz" event={"ID":"1fedc911-62a7-496f-93d1-f09769b89e86","Type":"ContainerStarted","Data":"a1792c7d8baeb9bc3d73b2f3b05a91fd631f488f4d7023f642964184619a907e"} Nov 21 14:11:11 crc kubenswrapper[4897]: W1121 14:11:11.974814 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5648ea03_bf08_419d_87fb_6287e11dd399.slice/crio-e67e7e9146b5ebf524bc7ad554dc99fff752d9914b9b734f4b169168f49c5035 WatchSource:0}: Error finding container e67e7e9146b5ebf524bc7ad554dc99fff752d9914b9b734f4b169168f49c5035: Status 404 returned error can't find the container with id e67e7e9146b5ebf524bc7ad554dc99fff752d9914b9b734f4b169168f49c5035 Nov 21 14:11:11 crc kubenswrapper[4897]: I1121 14:11:11.975832 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-gspw5" event={"ID":"107352c2-4357-43c2-9958-cb8a91714fa6","Type":"ContainerStarted","Data":"0567f206f0c68f5e8efa6c1fa68400fe513e8c833aa0366dee33011d042f975e"} Nov 21 14:11:12 crc kubenswrapper[4897]: I1121 14:11:11.987856 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:12 crc kubenswrapper[4897]: E1121 14:11:11.988431 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-21 14:11:12.488408294 +0000 UTC m=+149.773001769 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5z687" (UID: "9f63b8dd-ffb7-4963-8fcc-0c133a2c045b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:12 crc kubenswrapper[4897]: I1121 14:11:12.012011 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-mnn4f" event={"ID":"d7952588-d49d-41c1-b5ec-60e6ea4a1efa","Type":"ContainerStarted","Data":"68de337e862f5490466ea70ec63b871a68a2af89cbb07daa8ec261daac41c158"} Nov 21 14:11:12 crc kubenswrapper[4897]: I1121 14:11:12.018031 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5prl8" event={"ID":"ec19858a-32ef-443f-b307-df94c5d3e8de","Type":"ContainerStarted","Data":"8f0835ae5bacd6d213029cb7b4d90e8294181918aff45aa3bfa6a0a8289e4ced"} Nov 21 14:11:12 crc kubenswrapper[4897]: I1121 14:11:12.023717 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-t5jsz" event={"ID":"00cdea84-a65d-4422-8ecb-7bc8b912c557","Type":"ContainerStarted","Data":"ae106edbe8f1b40eba25ed8c6d66aff470ee11510a3f3455a6e3475f4464e990"} Nov 21 14:11:12 crc kubenswrapper[4897]: I1121 14:11:12.026854 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5n549" event={"ID":"74ff4ad0-9479-4fa2-8df9-501c332e386c","Type":"ContainerStarted","Data":"dfd0486b1017311d931b60a6b636d4f9c75511f6cf831eba77872890d3045098"} Nov 21 14:11:12 crc kubenswrapper[4897]: I1121 14:11:12.037531 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gxw89" event={"ID":"cc2bd33e-fa09-4675-9d48-b546668b5b86","Type":"ContainerStarted","Data":"c55c8e81823aad1d7aadc6c181a3f18c52f86d1c6cf10de0f820cab6c703a4e0"} Nov 21 14:11:12 crc kubenswrapper[4897]: I1121 14:11:12.044461 4897 patch_prober.go:28] interesting pod/console-operator-58897d9998-dfltg container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.10:8443/readyz\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Nov 21 14:11:12 crc kubenswrapper[4897]: I1121 14:11:12.044527 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-dfltg" podUID="18d8d58b-3dab-4949-b9d1-313870a96dd5" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.10:8443/readyz\": dial tcp 10.217.0.10:8443: connect: connection refused" Nov 21 14:11:12 crc kubenswrapper[4897]: I1121 14:11:12.051656 4897 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-rnnlh container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.11:8443/healthz\": dial tcp 10.217.0.11:8443: connect: connection refused" start-of-body= Nov 21 14:11:12 crc kubenswrapper[4897]: I1121 14:11:12.051719 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-rnnlh" podUID="4bf1a18e-ec09-4f71-b6a1-645698067dc1" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.11:8443/healthz\": dial tcp 10.217.0.11:8443: connect: connection refused" Nov 21 14:11:12 crc kubenswrapper[4897]: I1121 14:11:12.057607 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h59mj"] Nov 21 14:11:12 crc kubenswrapper[4897]: I1121 14:11:12.066558 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-dfltg" podStartSLOduration=124.066526833 podStartE2EDuration="2m4.066526833s" podCreationTimestamp="2025-11-21 14:09:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:11:12.051929461 +0000 UTC m=+149.336522936" watchObservedRunningTime="2025-11-21 14:11:12.066526833 +0000 UTC m=+149.351120328" Nov 21 14:11:12 crc kubenswrapper[4897]: I1121 14:11:12.094868 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 21 14:11:12 crc kubenswrapper[4897]: E1121 14:11:12.096132 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-21 14:11:12.596106239 +0000 UTC m=+149.880699714 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:12 crc kubenswrapper[4897]: I1121 14:11:12.143211 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4cx2q"] Nov 21 14:11:12 crc kubenswrapper[4897]: I1121 14:11:12.158654 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-s7tww"] Nov 21 14:11:12 crc kubenswrapper[4897]: I1121 14:11:12.197471 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:12 crc kubenswrapper[4897]: E1121 14:11:12.203325 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-21 14:11:12.70330385 +0000 UTC m=+149.987897325 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5z687" (UID: "9f63b8dd-ffb7-4963-8fcc-0c133a2c045b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:12 crc kubenswrapper[4897]: W1121 14:11:12.267057 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod56521238_a49d_446a_b4a9_e2809ff15826.slice/crio-2469e5fb856a07c6e06c32adae55dbf1fc742af1e8faca55166b71db836b4fb0 WatchSource:0}: Error finding container 2469e5fb856a07c6e06c32adae55dbf1fc742af1e8faca55166b71db836b4fb0: Status 404 returned error can't find the container with id 2469e5fb856a07c6e06c32adae55dbf1fc742af1e8faca55166b71db836b4fb0 Nov 21 14:11:12 crc kubenswrapper[4897]: I1121 14:11:12.322491 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 21 14:11:12 crc kubenswrapper[4897]: E1121 14:11:12.322852 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-21 14:11:12.82282424 +0000 UTC m=+150.107417705 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:12 crc kubenswrapper[4897]: I1121 14:11:12.323027 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:12 crc kubenswrapper[4897]: I1121 14:11:12.323803 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-g5z94" podStartSLOduration=125.323786737 podStartE2EDuration="2m5.323786737s" podCreationTimestamp="2025-11-21 14:09:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:11:12.318463646 +0000 UTC m=+149.603057121" watchObservedRunningTime="2025-11-21 14:11:12.323786737 +0000 UTC m=+149.608380212" Nov 21 14:11:12 crc kubenswrapper[4897]: I1121 14:11:12.331946 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-z9hjh"] Nov 21 14:11:12 crc kubenswrapper[4897]: E1121 14:11:12.334420 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-21 14:11:12.834393246 +0000 UTC m=+150.118986711 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5z687" (UID: "9f63b8dd-ffb7-4963-8fcc-0c133a2c045b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:12 crc kubenswrapper[4897]: I1121 14:11:12.357823 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-cc48c"] Nov 21 14:11:12 crc kubenswrapper[4897]: I1121 14:11:12.430758 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 21 14:11:12 crc kubenswrapper[4897]: E1121 14:11:12.432771 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-21 14:11:12.932686116 +0000 UTC m=+150.217279591 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:12 crc kubenswrapper[4897]: W1121 14:11:12.487949 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5262754c_6c80_4da0_8d9f_a58d63cfeadb.slice/crio-78d7339fbf894d21d22ff73fa9a2576ef25496ad9f4cec48a038cb8b85a62620 WatchSource:0}: Error finding container 78d7339fbf894d21d22ff73fa9a2576ef25496ad9f4cec48a038cb8b85a62620: Status 404 returned error can't find the container with id 78d7339fbf894d21d22ff73fa9a2576ef25496ad9f4cec48a038cb8b85a62620 Nov 21 14:11:12 crc kubenswrapper[4897]: I1121 14:11:12.534101 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:12 crc kubenswrapper[4897]: E1121 14:11:12.534689 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-21 14:11:13.034672719 +0000 UTC m=+150.319266194 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5z687" (UID: "9f63b8dd-ffb7-4963-8fcc-0c133a2c045b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:12 crc kubenswrapper[4897]: I1121 14:11:12.608386 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-74phw"] Nov 21 14:11:12 crc kubenswrapper[4897]: I1121 14:11:12.637625 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29395560-pq7jf"] Nov 21 14:11:12 crc kubenswrapper[4897]: I1121 14:11:12.639233 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 21 14:11:12 crc kubenswrapper[4897]: E1121 14:11:12.639895 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-21 14:11:13.139619126 +0000 UTC m=+150.424212741 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:12 crc kubenswrapper[4897]: I1121 14:11:12.646296 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6sfv4"] Nov 21 14:11:12 crc kubenswrapper[4897]: I1121 14:11:12.659019 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-lq8bp"] Nov 21 14:11:12 crc kubenswrapper[4897]: W1121 14:11:12.668083 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf7d61c37_621e_4812_bbaa_74006927176a.slice/crio-7bafb73cbc2fa1a1ca3a5f3e753d80615959111db72bb9362a3315dcc87ee78d WatchSource:0}: Error finding container 7bafb73cbc2fa1a1ca3a5f3e753d80615959111db72bb9362a3315dcc87ee78d: Status 404 returned error can't find the container with id 7bafb73cbc2fa1a1ca3a5f3e753d80615959111db72bb9362a3315dcc87ee78d Nov 21 14:11:12 crc kubenswrapper[4897]: I1121 14:11:12.668646 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-bjdjf"] Nov 21 14:11:12 crc kubenswrapper[4897]: I1121 14:11:12.740695 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:12 crc kubenswrapper[4897]: E1121 14:11:12.741048 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-21 14:11:13.241029454 +0000 UTC m=+150.525622929 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5z687" (UID: "9f63b8dd-ffb7-4963-8fcc-0c133a2c045b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:12 crc kubenswrapper[4897]: I1121 14:11:12.800479 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-nl2sk"] Nov 21 14:11:12 crc kubenswrapper[4897]: I1121 14:11:12.811388 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-rnnlh" podStartSLOduration=124.811362503 podStartE2EDuration="2m4.811362503s" podCreationTimestamp="2025-11-21 14:09:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:11:12.808893283 +0000 UTC m=+150.093486778" watchObservedRunningTime="2025-11-21 14:11:12.811362503 +0000 UTC m=+150.095955978" Nov 21 14:11:12 crc kubenswrapper[4897]: I1121 14:11:12.812743 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-b9s66"] Nov 21 14:11:12 crc kubenswrapper[4897]: I1121 14:11:12.828029 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-vf46x"] Nov 21 14:11:12 crc kubenswrapper[4897]: I1121 14:11:12.841293 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 21 14:11:12 crc kubenswrapper[4897]: E1121 14:11:12.841650 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-21 14:11:13.341628119 +0000 UTC m=+150.626221594 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:12 crc kubenswrapper[4897]: I1121 14:11:12.945959 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:12 crc kubenswrapper[4897]: E1121 14:11:12.947374 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-21 14:11:13.447355028 +0000 UTC m=+150.731948503 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5z687" (UID: "9f63b8dd-ffb7-4963-8fcc-0c133a2c045b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:13 crc kubenswrapper[4897]: I1121 14:11:13.026014 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-mnn4f" podStartSLOduration=125.025993521 podStartE2EDuration="2m5.025993521s" podCreationTimestamp="2025-11-21 14:09:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:11:13.023285214 +0000 UTC m=+150.307878699" watchObservedRunningTime="2025-11-21 14:11:13.025993521 +0000 UTC m=+150.310586996" Nov 21 14:11:13 crc kubenswrapper[4897]: I1121 14:11:13.036956 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-5n549" podStartSLOduration=124.0369245 podStartE2EDuration="2m4.0369245s" podCreationTimestamp="2025-11-21 14:09:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:11:12.963108313 +0000 UTC m=+150.247701808" watchObservedRunningTime="2025-11-21 14:11:13.0369245 +0000 UTC m=+150.321517975" Nov 21 14:11:13 crc kubenswrapper[4897]: I1121 14:11:13.050543 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 21 14:11:13 crc kubenswrapper[4897]: E1121 14:11:13.055117 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-21 14:11:13.555080414 +0000 UTC m=+150.839674069 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:13 crc kubenswrapper[4897]: I1121 14:11:13.098185 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-d8pkd" event={"ID":"911a0a7e-7a63-4b4a-8b74-6c21f8da7e31","Type":"ContainerStarted","Data":"0a91af45ac7010bb64c46a9ef1ecfa1a1da935fb8a4f4efff6bc9154489c1a3c"} Nov 21 14:11:13 crc kubenswrapper[4897]: I1121 14:11:13.117843 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-gspw5" podStartSLOduration=125.117826278 podStartE2EDuration="2m5.117826278s" podCreationTimestamp="2025-11-21 14:09:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:11:13.117164839 +0000 UTC m=+150.401758314" watchObservedRunningTime="2025-11-21 14:11:13.117826278 +0000 UTC m=+150.402419753" Nov 21 14:11:13 crc kubenswrapper[4897]: I1121 14:11:13.118366 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-2trmp" podStartSLOduration=124.118360303 podStartE2EDuration="2m4.118360303s" podCreationTimestamp="2025-11-21 14:09:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:11:13.069139101 +0000 UTC m=+150.353732596" watchObservedRunningTime="2025-11-21 14:11:13.118360303 +0000 UTC m=+150.402953778" Nov 21 14:11:13 crc kubenswrapper[4897]: I1121 14:11:13.120966 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2hmqv" event={"ID":"5648ea03-bf08-419d-87fb-6287e11dd399","Type":"ContainerStarted","Data":"714240581e7580533d0293011f2e14cff1e437c31e97c88837de8b0ca595c7bd"} Nov 21 14:11:13 crc kubenswrapper[4897]: I1121 14:11:13.121117 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2hmqv" event={"ID":"5648ea03-bf08-419d-87fb-6287e11dd399","Type":"ContainerStarted","Data":"e67e7e9146b5ebf524bc7ad554dc99fff752d9914b9b734f4b169168f49c5035"} Nov 21 14:11:13 crc kubenswrapper[4897]: I1121 14:11:13.128341 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-vf46x" event={"ID":"f2080446-c5c3-4c09-b8ae-54359bbb72ba","Type":"ContainerStarted","Data":"f1d0c283d1803cfd72a1fe9ecf1fa486012c471075ae7c15ade76ec03d680456"} Nov 21 14:11:13 crc kubenswrapper[4897]: I1121 14:11:13.155747 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:13 crc kubenswrapper[4897]: E1121 14:11:13.156116 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-21 14:11:13.65610281 +0000 UTC m=+150.940696285 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5z687" (UID: "9f63b8dd-ffb7-4963-8fcc-0c133a2c045b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:13 crc kubenswrapper[4897]: I1121 14:11:13.158049 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29395560-pq7jf" event={"ID":"c14a1c93-8be4-42b7-aae1-9c31d53a6bfc","Type":"ContainerStarted","Data":"f11286e8fc2e30ab9359b216a9923028c1bfc07c1ca7b16681bcd9b4acdcefae"} Nov 21 14:11:13 crc kubenswrapper[4897]: I1121 14:11:13.168991 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-vgwv2" event={"ID":"82f81ac2-d2b2-49de-82f3-a51796dbff03","Type":"ContainerStarted","Data":"fb2f5da04cb53f0f6c52a1de0459590c7227b4b9d96196cee1224495857a6db4"} Nov 21 14:11:13 crc kubenswrapper[4897]: I1121 14:11:13.169468 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-vgwv2" Nov 21 14:11:13 crc kubenswrapper[4897]: I1121 14:11:13.170527 4897 patch_prober.go:28] interesting pod/downloads-7954f5f757-vgwv2 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.26:8080/\": dial tcp 10.217.0.26:8080: connect: connection refused" start-of-body= Nov 21 14:11:13 crc kubenswrapper[4897]: I1121 14:11:13.170561 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-vgwv2" podUID="82f81ac2-d2b2-49de-82f3-a51796dbff03" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.26:8080/\": dial tcp 10.217.0.26:8080: connect: connection refused" Nov 21 14:11:13 crc kubenswrapper[4897]: I1121 14:11:13.170649 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-lq8bp" event={"ID":"20a1d330-d99c-480d-b44c-4d8a5b4c6453","Type":"ContainerStarted","Data":"78a6ec40d9a5863998dca8d27acd2c9a71e93907d9c6f5c633223fdc96097e13"} Nov 21 14:11:13 crc kubenswrapper[4897]: I1121 14:11:13.174084 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-b9s66" event={"ID":"6d1fd716-6991-43b6-9f1f-268576e864a0","Type":"ContainerStarted","Data":"1512ead145adabaa0ff354da4140aec9e9b00d9dece075ababf4ac9a8f579e3d"} Nov 21 14:11:13 crc kubenswrapper[4897]: I1121 14:11:13.175540 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-bjdjf" event={"ID":"78db7e2e-ba45-464e-9b1d-0fc61191955a","Type":"ContainerStarted","Data":"cbab3b28068fca0f0bed8ce7a9be1f138f0c461a310375273808082fb377c2d3"} Nov 21 14:11:13 crc kubenswrapper[4897]: I1121 14:11:13.181223 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-szznp" event={"ID":"b76cdd1d-6a5d-42ac-8b03-1cc105eb1c1f","Type":"ContainerStarted","Data":"2562e8f2d2a6893e7364b92adf8287e9f3d2cada3a0a3fff451eeb95d6f86e88"} Nov 21 14:11:13 crc kubenswrapper[4897]: I1121 14:11:13.193583 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gxw89" podStartSLOduration=125.193561539 podStartE2EDuration="2m5.193561539s" podCreationTimestamp="2025-11-21 14:09:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:11:13.158420946 +0000 UTC m=+150.443014451" watchObservedRunningTime="2025-11-21 14:11:13.193561539 +0000 UTC m=+150.478155014" Nov 21 14:11:13 crc kubenswrapper[4897]: I1121 14:11:13.208294 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kf7zh" event={"ID":"4c817f5a-6f7c-48d3-9d45-bdc401646bed","Type":"ContainerStarted","Data":"0d1707c9fa55582bc2f60026715cebbca4a166d958a64740a0e71ee4554c9e3b"} Nov 21 14:11:13 crc kubenswrapper[4897]: I1121 14:11:13.211943 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-lp8th" event={"ID":"98941613-bca1-4b35-9dba-547294a1feb5","Type":"ContainerStarted","Data":"1839ed5d270f03cab7bb28f923f943692077db3d0a9a9c1cad1f06eb2e374247"} Nov 21 14:11:13 crc kubenswrapper[4897]: I1121 14:11:13.213199 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-nl2sk" event={"ID":"6a9aee9a-c47c-4ad2-9431-1a4ac5ce15fd","Type":"ContainerStarted","Data":"15892a75aad1cab2491c9606e16e4343ca7d462862d2f489385b579ddaa7112c"} Nov 21 14:11:13 crc kubenswrapper[4897]: I1121 14:11:13.214325 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h59mj" event={"ID":"b24e87a6-3cd8-4a37-b537-e9a5332cb6c7","Type":"ContainerStarted","Data":"32ac98918b81f8bff2153224de885a5b3fb55bf040e96ce5693f2286b536d705"} Nov 21 14:11:13 crc kubenswrapper[4897]: I1121 14:11:13.214350 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h59mj" event={"ID":"b24e87a6-3cd8-4a37-b537-e9a5332cb6c7","Type":"ContainerStarted","Data":"1d1d287b3311c7b5cf1c63bec01f2d74fb1c9cba6278cbc3bfccbfa647e23cc1"} Nov 21 14:11:13 crc kubenswrapper[4897]: I1121 14:11:13.215464 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h59mj" Nov 21 14:11:13 crc kubenswrapper[4897]: I1121 14:11:13.217894 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-74phw" event={"ID":"f7d61c37-621e-4812-bbaa-74006927176a","Type":"ContainerStarted","Data":"7bafb73cbc2fa1a1ca3a5f3e753d80615959111db72bb9362a3315dcc87ee78d"} Nov 21 14:11:13 crc kubenswrapper[4897]: I1121 14:11:13.220313 4897 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-h59mj container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.25:8443/healthz\": dial tcp 10.217.0.25:8443: connect: connection refused" start-of-body= Nov 21 14:11:13 crc kubenswrapper[4897]: I1121 14:11:13.220410 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h59mj" podUID="b24e87a6-3cd8-4a37-b537-e9a5332cb6c7" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.25:8443/healthz\": dial tcp 10.217.0.25:8443: connect: connection refused" Nov 21 14:11:13 crc kubenswrapper[4897]: I1121 14:11:13.225643 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-cc48c" event={"ID":"5262754c-6c80-4da0-8d9f-a58d63cfeadb","Type":"ContainerStarted","Data":"78d7339fbf894d21d22ff73fa9a2576ef25496ad9f4cec48a038cb8b85a62620"} Nov 21 14:11:13 crc kubenswrapper[4897]: I1121 14:11:13.227093 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-8kddm" podStartSLOduration=124.227028905 podStartE2EDuration="2m4.227028905s" podCreationTimestamp="2025-11-21 14:09:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:11:13.1960748 +0000 UTC m=+150.480668275" watchObservedRunningTime="2025-11-21 14:11:13.227028905 +0000 UTC m=+150.511622380" Nov 21 14:11:13 crc kubenswrapper[4897]: I1121 14:11:13.236146 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-7fsr9" event={"ID":"8342c8a2-87f4-4bf9-a2ca-073c02b277ad","Type":"ContainerStarted","Data":"8a96362523e6d8f3d19ad31a419973897ae10647ff86cb917812b0afe685363b"} Nov 21 14:11:13 crc kubenswrapper[4897]: I1121 14:11:13.243248 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-bqktx" event={"ID":"85811958-7716-47c2-b9ad-9392647bb172","Type":"ContainerStarted","Data":"ef16c95a06fe27123eccdc0a5ebe22499bdf19e3b6ce0e51f169c27940e94b50"} Nov 21 14:11:13 crc kubenswrapper[4897]: I1121 14:11:13.243308 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-bqktx" event={"ID":"85811958-7716-47c2-b9ad-9392647bb172","Type":"ContainerStarted","Data":"18e926e8300e09b621eb179eab0d52fd292689bc0d200b5f5a31c7822ab4915d"} Nov 21 14:11:13 crc kubenswrapper[4897]: I1121 14:11:13.247246 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lclf5" podStartSLOduration=124.247222896 podStartE2EDuration="2m4.247222896s" podCreationTimestamp="2025-11-21 14:09:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:11:13.244367035 +0000 UTC m=+150.528960510" watchObservedRunningTime="2025-11-21 14:11:13.247222896 +0000 UTC m=+150.531816371" Nov 21 14:11:13 crc kubenswrapper[4897]: I1121 14:11:13.248953 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-hk7mf" podStartSLOduration=124.248944975 podStartE2EDuration="2m4.248944975s" podCreationTimestamp="2025-11-21 14:09:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:11:13.225575674 +0000 UTC m=+150.510169149" watchObservedRunningTime="2025-11-21 14:11:13.248944975 +0000 UTC m=+150.533538450" Nov 21 14:11:13 crc kubenswrapper[4897]: I1121 14:11:13.250132 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8s4px" event={"ID":"cfacdafa-2513-4d81-9222-c4b4fd8f8c10","Type":"ContainerStarted","Data":"0b2a9212550e571bc7784a4eff851203ec0405340188c66c4c8e12a6ebcdb9fb"} Nov 21 14:11:13 crc kubenswrapper[4897]: I1121 14:11:13.256039 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4cx2q" event={"ID":"56521238-a49d-446a-b4a9-e2809ff15826","Type":"ContainerStarted","Data":"32e767a0f6b431dec2850079211fe817e099ba024296281eec7dcca8751cce35"} Nov 21 14:11:13 crc kubenswrapper[4897]: I1121 14:11:13.256094 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4cx2q" event={"ID":"56521238-a49d-446a-b4a9-e2809ff15826","Type":"ContainerStarted","Data":"2469e5fb856a07c6e06c32adae55dbf1fc742af1e8faca55166b71db836b4fb0"} Nov 21 14:11:13 crc kubenswrapper[4897]: I1121 14:11:13.257305 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4cx2q" Nov 21 14:11:13 crc kubenswrapper[4897]: I1121 14:11:13.258764 4897 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-4cx2q container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.24:8443/healthz\": dial tcp 10.217.0.24:8443: connect: connection refused" start-of-body= Nov 21 14:11:13 crc kubenswrapper[4897]: I1121 14:11:13.258824 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4cx2q" podUID="56521238-a49d-446a-b4a9-e2809ff15826" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.24:8443/healthz\": dial tcp 10.217.0.24:8443: connect: connection refused" Nov 21 14:11:13 crc kubenswrapper[4897]: I1121 14:11:13.259495 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 21 14:11:13 crc kubenswrapper[4897]: E1121 14:11:13.259903 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-21 14:11:13.759868774 +0000 UTC m=+151.044462249 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:13 crc kubenswrapper[4897]: I1121 14:11:13.260819 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:13 crc kubenswrapper[4897]: E1121 14:11:13.267938 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-21 14:11:13.767915901 +0000 UTC m=+151.052509376 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5z687" (UID: "9f63b8dd-ffb7-4963-8fcc-0c133a2c045b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:13 crc kubenswrapper[4897]: I1121 14:11:13.269359 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-pgf6s" event={"ID":"a309e9aa-bf13-404b-90dd-8ddc81467822","Type":"ContainerStarted","Data":"36d27e699a19666e2484b3e591aa4380e1f118c3086a7e79355c9cbaa470d750"} Nov 21 14:11:13 crc kubenswrapper[4897]: I1121 14:11:13.272181 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-z9hjh" event={"ID":"e128526a-8489-47e1-a711-4977f2b0d46d","Type":"ContainerStarted","Data":"c045494f8df7b7987645b495a3806d7fe2c61470dfc6a9479ad0ae14c20f251d"} Nov 21 14:11:13 crc kubenswrapper[4897]: I1121 14:11:13.274749 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5prl8" event={"ID":"ec19858a-32ef-443f-b307-df94c5d3e8de","Type":"ContainerStarted","Data":"affdd1d1c563c326ae9222f0fecd540bb6922bc697b2559e70d8763a02f938bd"} Nov 21 14:11:13 crc kubenswrapper[4897]: I1121 14:11:13.276571 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-s7tww" event={"ID":"027b031c-1b21-40e8-aa7c-4a342aae265a","Type":"ContainerStarted","Data":"768f5bbadef3ed802193984d690144015eb350027540c1da66b45453d93ac4fc"} Nov 21 14:11:13 crc kubenswrapper[4897]: I1121 14:11:13.282036 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-v8bk6" event={"ID":"1d13c714-5451-44fd-9dad-8d613a6e3545","Type":"ContainerStarted","Data":"aabce05fdd91f43e7fcf9ddcf06c499aa7854637675589eacbadbf929fe06c58"} Nov 21 14:11:13 crc kubenswrapper[4897]: I1121 14:11:13.286472 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-fqd92" event={"ID":"b0b1043c-75c9-4643-a400-a5c009dfb5b6","Type":"ContainerStarted","Data":"e1826c2ea7e4f410c69554f4e2c94212076ed01147e10c1e07ff6d3f09c0a06d"} Nov 21 14:11:13 crc kubenswrapper[4897]: I1121 14:11:13.292403 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6sfv4" event={"ID":"11301ff8-fa01-48da-a95e-fc244fd0a575","Type":"ContainerStarted","Data":"0f32f3909ee5b0f4d9f7675db0b2904e9aaaedb01f9141fbd53f33838d3bd8a3"} Nov 21 14:11:13 crc kubenswrapper[4897]: I1121 14:11:13.292854 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rv2nz" podStartSLOduration=124.29260676 podStartE2EDuration="2m4.29260676s" podCreationTimestamp="2025-11-21 14:09:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:11:13.287180186 +0000 UTC m=+150.571773661" watchObservedRunningTime="2025-11-21 14:11:13.29260676 +0000 UTC m=+150.577200225" Nov 21 14:11:13 crc kubenswrapper[4897]: I1121 14:11:13.327743 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-t5jsz" event={"ID":"00cdea84-a65d-4422-8ecb-7bc8b912c557","Type":"ContainerStarted","Data":"61bfcae24f7fa11ea34ac4b520b1a34612209cf320cfdd9aa98b2462556b76fa"} Nov 21 14:11:13 crc kubenswrapper[4897]: I1121 14:11:13.328703 4897 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-rnnlh container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.11:8443/healthz\": dial tcp 10.217.0.11:8443: connect: connection refused" start-of-body= Nov 21 14:11:13 crc kubenswrapper[4897]: I1121 14:11:13.328803 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-rnnlh" podUID="4bf1a18e-ec09-4f71-b6a1-645698067dc1" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.11:8443/healthz\": dial tcp 10.217.0.11:8443: connect: connection refused" Nov 21 14:11:13 crc kubenswrapper[4897]: I1121 14:11:13.329230 4897 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-2trmp container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.13:6443/healthz\": dial tcp 10.217.0.13:6443: connect: connection refused" start-of-body= Nov 21 14:11:13 crc kubenswrapper[4897]: I1121 14:11:13.329291 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-2trmp" podUID="0e4b5f84-4954-4d85-8bae-db18ed352464" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.13:6443/healthz\": dial tcp 10.217.0.13:6443: connect: connection refused" Nov 21 14:11:13 crc kubenswrapper[4897]: I1121 14:11:13.329791 4897 patch_prober.go:28] interesting pod/console-operator-58897d9998-dfltg container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.10:8443/readyz\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Nov 21 14:11:13 crc kubenswrapper[4897]: I1121 14:11:13.329853 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-dfltg" podUID="18d8d58b-3dab-4949-b9d1-313870a96dd5" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.10:8443/readyz\": dial tcp 10.217.0.10:8443: connect: connection refused" Nov 21 14:11:13 crc kubenswrapper[4897]: I1121 14:11:13.362487 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-7fsr9" podStartSLOduration=125.362466795 podStartE2EDuration="2m5.362466795s" podCreationTimestamp="2025-11-21 14:09:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:11:13.32798157 +0000 UTC m=+150.612575045" watchObservedRunningTime="2025-11-21 14:11:13.362466795 +0000 UTC m=+150.647060270" Nov 21 14:11:13 crc kubenswrapper[4897]: I1121 14:11:13.366913 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 21 14:11:13 crc kubenswrapper[4897]: E1121 14:11:13.369066 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-21 14:11:13.86903363 +0000 UTC m=+151.153627105 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:13 crc kubenswrapper[4897]: I1121 14:11:13.375211 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:13 crc kubenswrapper[4897]: E1121 14:11:13.378542 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-21 14:11:13.878519969 +0000 UTC m=+151.163113474 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5z687" (UID: "9f63b8dd-ffb7-4963-8fcc-0c133a2c045b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:13 crc kubenswrapper[4897]: I1121 14:11:13.405727 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4cx2q" podStartSLOduration=124.405706588 podStartE2EDuration="2m4.405706588s" podCreationTimestamp="2025-11-21 14:09:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:11:13.363871505 +0000 UTC m=+150.648464980" watchObservedRunningTime="2025-11-21 14:11:13.405706588 +0000 UTC m=+150.690300063" Nov 21 14:11:13 crc kubenswrapper[4897]: I1121 14:11:13.409390 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-pgf6s" podStartSLOduration=125.409371381 podStartE2EDuration="2m5.409371381s" podCreationTimestamp="2025-11-21 14:09:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:11:13.405334727 +0000 UTC m=+150.689928212" watchObservedRunningTime="2025-11-21 14:11:13.409371381 +0000 UTC m=+150.693964876" Nov 21 14:11:13 crc kubenswrapper[4897]: I1121 14:11:13.445625 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5prl8" podStartSLOduration=124.445584695 podStartE2EDuration="2m4.445584695s" podCreationTimestamp="2025-11-21 14:09:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:11:13.436764216 +0000 UTC m=+150.721357691" watchObservedRunningTime="2025-11-21 14:11:13.445584695 +0000 UTC m=+150.730178210" Nov 21 14:11:13 crc kubenswrapper[4897]: I1121 14:11:13.482569 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 21 14:11:13 crc kubenswrapper[4897]: E1121 14:11:13.485929 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-21 14:11:13.985905695 +0000 UTC m=+151.270499240 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:13 crc kubenswrapper[4897]: I1121 14:11:13.499972 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-lp8th" podStartSLOduration=124.499950402 podStartE2EDuration="2m4.499950402s" podCreationTimestamp="2025-11-21 14:09:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:11:13.485698169 +0000 UTC m=+150.770291644" watchObservedRunningTime="2025-11-21 14:11:13.499950402 +0000 UTC m=+150.784543877" Nov 21 14:11:13 crc kubenswrapper[4897]: I1121 14:11:13.522247 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-vgwv2" podStartSLOduration=125.522224412 podStartE2EDuration="2m5.522224412s" podCreationTimestamp="2025-11-21 14:09:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:11:13.520873564 +0000 UTC m=+150.805467039" watchObservedRunningTime="2025-11-21 14:11:13.522224412 +0000 UTC m=+150.806817887" Nov 21 14:11:13 crc kubenswrapper[4897]: I1121 14:11:13.588750 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:13 crc kubenswrapper[4897]: E1121 14:11:13.589335 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-21 14:11:14.089316729 +0000 UTC m=+151.373910204 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5z687" (UID: "9f63b8dd-ffb7-4963-8fcc-0c133a2c045b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:13 crc kubenswrapper[4897]: I1121 14:11:13.591119 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8s4px" podStartSLOduration=124.591102529 podStartE2EDuration="2m4.591102529s" podCreationTimestamp="2025-11-21 14:09:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:11:13.589872095 +0000 UTC m=+150.874465590" watchObservedRunningTime="2025-11-21 14:11:13.591102529 +0000 UTC m=+150.875696004" Nov 21 14:11:13 crc kubenswrapper[4897]: I1121 14:11:13.624008 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-bqktx" podStartSLOduration=5.623968158 podStartE2EDuration="5.623968158s" podCreationTimestamp="2025-11-21 14:11:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:11:13.620933633 +0000 UTC m=+150.905527108" watchObservedRunningTime="2025-11-21 14:11:13.623968158 +0000 UTC m=+150.908561633" Nov 21 14:11:13 crc kubenswrapper[4897]: I1121 14:11:13.646308 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-fqd92" podStartSLOduration=5.646281009 podStartE2EDuration="5.646281009s" podCreationTimestamp="2025-11-21 14:11:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:11:13.644016256 +0000 UTC m=+150.928609731" watchObservedRunningTime="2025-11-21 14:11:13.646281009 +0000 UTC m=+150.930874484" Nov 21 14:11:13 crc kubenswrapper[4897]: I1121 14:11:13.690388 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 21 14:11:13 crc kubenswrapper[4897]: E1121 14:11:13.691342 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-21 14:11:14.191321233 +0000 UTC m=+151.475914708 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:13 crc kubenswrapper[4897]: I1121 14:11:13.718197 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-szznp" podStartSLOduration=124.718179783 podStartE2EDuration="2m4.718179783s" podCreationTimestamp="2025-11-21 14:09:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:11:13.680429475 +0000 UTC m=+150.965022970" watchObservedRunningTime="2025-11-21 14:11:13.718179783 +0000 UTC m=+151.002773258" Nov 21 14:11:13 crc kubenswrapper[4897]: I1121 14:11:13.758900 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h59mj" podStartSLOduration=124.758878563 podStartE2EDuration="2m4.758878563s" podCreationTimestamp="2025-11-21 14:09:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:11:13.717441851 +0000 UTC m=+151.002035316" watchObservedRunningTime="2025-11-21 14:11:13.758878563 +0000 UTC m=+151.043472038" Nov 21 14:11:13 crc kubenswrapper[4897]: I1121 14:11:13.793771 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:13 crc kubenswrapper[4897]: E1121 14:11:13.794798 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-21 14:11:14.294778498 +0000 UTC m=+151.579371973 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5z687" (UID: "9f63b8dd-ffb7-4963-8fcc-0c133a2c045b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:13 crc kubenswrapper[4897]: I1121 14:11:13.802292 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-2kgg2" podStartSLOduration=124.80226542 podStartE2EDuration="2m4.80226542s" podCreationTimestamp="2025-11-21 14:09:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:11:13.800321685 +0000 UTC m=+151.084915180" watchObservedRunningTime="2025-11-21 14:11:13.80226542 +0000 UTC m=+151.086858895" Nov 21 14:11:13 crc kubenswrapper[4897]: I1121 14:11:13.895755 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 21 14:11:13 crc kubenswrapper[4897]: E1121 14:11:13.895996 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-21 14:11:14.395961739 +0000 UTC m=+151.680555214 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:13 crc kubenswrapper[4897]: I1121 14:11:13.896202 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:13 crc kubenswrapper[4897]: E1121 14:11:13.896630 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-21 14:11:14.396615557 +0000 UTC m=+151.681209022 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5z687" (UID: "9f63b8dd-ffb7-4963-8fcc-0c133a2c045b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:13 crc kubenswrapper[4897]: I1121 14:11:13.988909 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-2kgg2" Nov 21 14:11:13 crc kubenswrapper[4897]: I1121 14:11:13.992242 4897 patch_prober.go:28] interesting pod/router-default-5444994796-2kgg2 container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Nov 21 14:11:13 crc kubenswrapper[4897]: I1121 14:11:13.992307 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-2kgg2" podUID="e095a18e-3a0b-4418-95a8-a2cc3de5bd7b" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Nov 21 14:11:13 crc kubenswrapper[4897]: I1121 14:11:13.997291 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 21 14:11:13 crc kubenswrapper[4897]: E1121 14:11:13.997589 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-21 14:11:14.497545491 +0000 UTC m=+151.782138976 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:13 crc kubenswrapper[4897]: I1121 14:11:13.998052 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:13 crc kubenswrapper[4897]: E1121 14:11:13.998811 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-21 14:11:14.498799507 +0000 UTC m=+151.783392992 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5z687" (UID: "9f63b8dd-ffb7-4963-8fcc-0c133a2c045b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:14 crc kubenswrapper[4897]: I1121 14:11:14.099467 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 21 14:11:14 crc kubenswrapper[4897]: E1121 14:11:14.099648 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-21 14:11:14.599612877 +0000 UTC m=+151.884206352 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:14 crc kubenswrapper[4897]: I1121 14:11:14.099723 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:14 crc kubenswrapper[4897]: E1121 14:11:14.100059 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-21 14:11:14.600050939 +0000 UTC m=+151.884644414 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5z687" (UID: "9f63b8dd-ffb7-4963-8fcc-0c133a2c045b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:14 crc kubenswrapper[4897]: I1121 14:11:14.200460 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 21 14:11:14 crc kubenswrapper[4897]: E1121 14:11:14.200617 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-21 14:11:14.700583752 +0000 UTC m=+151.985177227 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:14 crc kubenswrapper[4897]: I1121 14:11:14.200866 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:14 crc kubenswrapper[4897]: E1121 14:11:14.201209 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-21 14:11:14.70119058 +0000 UTC m=+151.985784055 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5z687" (UID: "9f63b8dd-ffb7-4963-8fcc-0c133a2c045b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:14 crc kubenswrapper[4897]: I1121 14:11:14.301492 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 21 14:11:14 crc kubenswrapper[4897]: E1121 14:11:14.301690 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-21 14:11:14.801651639 +0000 UTC m=+152.086245114 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:14 crc kubenswrapper[4897]: I1121 14:11:14.302023 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:14 crc kubenswrapper[4897]: E1121 14:11:14.302404 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-21 14:11:14.802395341 +0000 UTC m=+152.086988816 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5z687" (UID: "9f63b8dd-ffb7-4963-8fcc-0c133a2c045b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:14 crc kubenswrapper[4897]: I1121 14:11:14.332963 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-nl2sk" event={"ID":"6a9aee9a-c47c-4ad2-9431-1a4ac5ce15fd","Type":"ContainerStarted","Data":"6b24a93b66a6548ee625293462d5d9c0af6532e1398d2fc3cbf86cd145b7fafe"} Nov 21 14:11:14 crc kubenswrapper[4897]: I1121 14:11:14.334290 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-lq8bp" event={"ID":"20a1d330-d99c-480d-b44c-4d8a5b4c6453","Type":"ContainerStarted","Data":"e3bc5a45a94066aa5533ad6f341e64ee9f281711b5e4d6dd4064f364d3595438"} Nov 21 14:11:14 crc kubenswrapper[4897]: I1121 14:11:14.335576 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29395560-pq7jf" event={"ID":"c14a1c93-8be4-42b7-aae1-9c31d53a6bfc","Type":"ContainerStarted","Data":"7c71e97295b9fe2b2d30a0502ffb2637c6dd199e5c02c6709abf1958f64afa5a"} Nov 21 14:11:14 crc kubenswrapper[4897]: I1121 14:11:14.337359 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-vf46x" event={"ID":"f2080446-c5c3-4c09-b8ae-54359bbb72ba","Type":"ContainerStarted","Data":"1917ef9634d935c816e2bb0331f15900f2483151f86bc34991e429a9da206bd5"} Nov 21 14:11:14 crc kubenswrapper[4897]: I1121 14:11:14.338661 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-74phw" event={"ID":"f7d61c37-621e-4812-bbaa-74006927176a","Type":"ContainerStarted","Data":"36b21496bd89aa682ece671b40e78e2a9797ea0cac619a92b2d2223f3c1a7fcd"} Nov 21 14:11:14 crc kubenswrapper[4897]: I1121 14:11:14.339902 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-z9hjh" event={"ID":"e128526a-8489-47e1-a711-4977f2b0d46d","Type":"ContainerStarted","Data":"fd8a88f7278af250e4cea7de45d4fc4a9196d7593856d54edf70fd466c2f0ea0"} Nov 21 14:11:14 crc kubenswrapper[4897]: I1121 14:11:14.340841 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-z9hjh" Nov 21 14:11:14 crc kubenswrapper[4897]: I1121 14:11:14.347455 4897 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-z9hjh container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.30:8080/healthz\": dial tcp 10.217.0.30:8080: connect: connection refused" start-of-body= Nov 21 14:11:14 crc kubenswrapper[4897]: I1121 14:11:14.347667 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-z9hjh" podUID="e128526a-8489-47e1-a711-4977f2b0d46d" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.30:8080/healthz\": dial tcp 10.217.0.30:8080: connect: connection refused" Nov 21 14:11:14 crc kubenswrapper[4897]: I1121 14:11:14.352110 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-v8bk6" event={"ID":"1d13c714-5451-44fd-9dad-8d613a6e3545","Type":"ContainerStarted","Data":"dfd571dbe40b59ab355af52ba465a8264eb78e85963c3e8c752c93f5e0602081"} Nov 21 14:11:14 crc kubenswrapper[4897]: I1121 14:11:14.357087 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-b9s66" event={"ID":"6d1fd716-6991-43b6-9f1f-268576e864a0","Type":"ContainerStarted","Data":"bd41211a38a3df746a4422fb7ee5b5f5a760909d94bb7e7d3aaf5fdc160d8671"} Nov 21 14:11:14 crc kubenswrapper[4897]: I1121 14:11:14.360259 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-nl2sk" podStartSLOduration=125.360241026 podStartE2EDuration="2m5.360241026s" podCreationTimestamp="2025-11-21 14:09:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:11:14.35826063 +0000 UTC m=+151.642854115" watchObservedRunningTime="2025-11-21 14:11:14.360241026 +0000 UTC m=+151.644834501" Nov 21 14:11:14 crc kubenswrapper[4897]: I1121 14:11:14.368744 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6sfv4" event={"ID":"11301ff8-fa01-48da-a95e-fc244fd0a575","Type":"ContainerStarted","Data":"ba10c41814445283d3274af082a06d79354626f6f5929b9e3f5ae82682f06fa1"} Nov 21 14:11:14 crc kubenswrapper[4897]: I1121 14:11:14.369756 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6sfv4" Nov 21 14:11:14 crc kubenswrapper[4897]: I1121 14:11:14.373641 4897 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-6sfv4 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.29:5443/healthz\": dial tcp 10.217.0.29:5443: connect: connection refused" start-of-body= Nov 21 14:11:14 crc kubenswrapper[4897]: I1121 14:11:14.373707 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6sfv4" podUID="11301ff8-fa01-48da-a95e-fc244fd0a575" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.29:5443/healthz\": dial tcp 10.217.0.29:5443: connect: connection refused" Nov 21 14:11:14 crc kubenswrapper[4897]: I1121 14:11:14.385044 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kf7zh" event={"ID":"4c817f5a-6f7c-48d3-9d45-bdc401646bed","Type":"ContainerStarted","Data":"99774e58809dbc57c79917280294b4f5f3f180f5e5dd0126a3ae6b45a9150b4d"} Nov 21 14:11:14 crc kubenswrapper[4897]: I1121 14:11:14.386389 4897 generic.go:334] "Generic (PLEG): container finished" podID="027b031c-1b21-40e8-aa7c-4a342aae265a" containerID="3b30e13bcb2b0cb5abcd3190d26b79a72ad24055545054b81e7c13e86a678c36" exitCode=0 Nov 21 14:11:14 crc kubenswrapper[4897]: I1121 14:11:14.386561 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-s7tww" event={"ID":"027b031c-1b21-40e8-aa7c-4a342aae265a","Type":"ContainerDied","Data":"3b30e13bcb2b0cb5abcd3190d26b79a72ad24055545054b81e7c13e86a678c36"} Nov 21 14:11:14 crc kubenswrapper[4897]: I1121 14:11:14.393651 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-cc48c" event={"ID":"5262754c-6c80-4da0-8d9f-a58d63cfeadb","Type":"ContainerStarted","Data":"f5da08ccf8a79a2930b0cddd886ab20722c4ce9409b447b09022d781d699afeb"} Nov 21 14:11:14 crc kubenswrapper[4897]: I1121 14:11:14.393746 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-cc48c" event={"ID":"5262754c-6c80-4da0-8d9f-a58d63cfeadb","Type":"ContainerStarted","Data":"b18826c86bd3852cfdc9b1b612819ad6d0c83b3a4c2afb94df61b6b205b864b2"} Nov 21 14:11:14 crc kubenswrapper[4897]: I1121 14:11:14.404622 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 21 14:11:14 crc kubenswrapper[4897]: E1121 14:11:14.405765 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-21 14:11:14.905740543 +0000 UTC m=+152.190334018 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:14 crc kubenswrapper[4897]: I1121 14:11:14.408300 4897 generic.go:334] "Generic (PLEG): container finished" podID="00cdea84-a65d-4422-8ecb-7bc8b912c557" containerID="61bfcae24f7fa11ea34ac4b520b1a34612209cf320cfdd9aa98b2462556b76fa" exitCode=0 Nov 21 14:11:14 crc kubenswrapper[4897]: I1121 14:11:14.408404 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-t5jsz" event={"ID":"00cdea84-a65d-4422-8ecb-7bc8b912c557","Type":"ContainerDied","Data":"61bfcae24f7fa11ea34ac4b520b1a34612209cf320cfdd9aa98b2462556b76fa"} Nov 21 14:11:14 crc kubenswrapper[4897]: I1121 14:11:14.408432 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-t5jsz" event={"ID":"00cdea84-a65d-4422-8ecb-7bc8b912c557","Type":"ContainerStarted","Data":"db29b8cae763b4e7a0f306ccd1dd07ef9d263bb09a63ce48f06ddb02fc57c9f6"} Nov 21 14:11:14 crc kubenswrapper[4897]: I1121 14:11:14.411532 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2hmqv" event={"ID":"5648ea03-bf08-419d-87fb-6287e11dd399","Type":"ContainerStarted","Data":"e802c0728e5be404e93e9a4c9b0f793ad9b3e9f15a1931404502c81df9d39b2b"} Nov 21 14:11:14 crc kubenswrapper[4897]: I1121 14:11:14.412132 4897 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-4cx2q container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.24:8443/healthz\": dial tcp 10.217.0.24:8443: connect: connection refused" start-of-body= Nov 21 14:11:14 crc kubenswrapper[4897]: I1121 14:11:14.412185 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4cx2q" podUID="56521238-a49d-446a-b4a9-e2809ff15826" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.24:8443/healthz\": dial tcp 10.217.0.24:8443: connect: connection refused" Nov 21 14:11:14 crc kubenswrapper[4897]: I1121 14:11:14.412355 4897 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-h59mj container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.25:8443/healthz\": dial tcp 10.217.0.25:8443: connect: connection refused" start-of-body= Nov 21 14:11:14 crc kubenswrapper[4897]: I1121 14:11:14.412465 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h59mj" podUID="b24e87a6-3cd8-4a37-b537-e9a5332cb6c7" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.25:8443/healthz\": dial tcp 10.217.0.25:8443: connect: connection refused" Nov 21 14:11:14 crc kubenswrapper[4897]: I1121 14:11:14.412774 4897 patch_prober.go:28] interesting pod/downloads-7954f5f757-vgwv2 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.26:8080/\": dial tcp 10.217.0.26:8080: connect: connection refused" start-of-body= Nov 21 14:11:14 crc kubenswrapper[4897]: I1121 14:11:14.412820 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-vgwv2" podUID="82f81ac2-d2b2-49de-82f3-a51796dbff03" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.26:8080/\": dial tcp 10.217.0.26:8080: connect: connection refused" Nov 21 14:11:14 crc kubenswrapper[4897]: I1121 14:11:14.431254 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29395560-pq7jf" podStartSLOduration=126.431237104 podStartE2EDuration="2m6.431237104s" podCreationTimestamp="2025-11-21 14:09:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:11:14.430941646 +0000 UTC m=+151.715535121" watchObservedRunningTime="2025-11-21 14:11:14.431237104 +0000 UTC m=+151.715830579" Nov 21 14:11:14 crc kubenswrapper[4897]: I1121 14:11:14.432576 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-b9s66" podStartSLOduration=125.432572041 podStartE2EDuration="2m5.432572041s" podCreationTimestamp="2025-11-21 14:09:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:11:14.397951222 +0000 UTC m=+151.682544697" watchObservedRunningTime="2025-11-21 14:11:14.432572041 +0000 UTC m=+151.717165516" Nov 21 14:11:14 crc kubenswrapper[4897]: I1121 14:11:14.460936 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-z9hjh" podStartSLOduration=125.460919763 podStartE2EDuration="2m5.460919763s" podCreationTimestamp="2025-11-21 14:09:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:11:14.46007836 +0000 UTC m=+151.744671835" watchObservedRunningTime="2025-11-21 14:11:14.460919763 +0000 UTC m=+151.745513238" Nov 21 14:11:14 crc kubenswrapper[4897]: I1121 14:11:14.479038 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-v8bk6" podStartSLOduration=125.479022815 podStartE2EDuration="2m5.479022815s" podCreationTimestamp="2025-11-21 14:09:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:11:14.476301608 +0000 UTC m=+151.760895083" watchObservedRunningTime="2025-11-21 14:11:14.479022815 +0000 UTC m=+151.763616290" Nov 21 14:11:14 crc kubenswrapper[4897]: I1121 14:11:14.507542 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:14 crc kubenswrapper[4897]: I1121 14:11:14.509563 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-d8pkd" podStartSLOduration=125.509540508 podStartE2EDuration="2m5.509540508s" podCreationTimestamp="2025-11-21 14:09:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:11:14.507435538 +0000 UTC m=+151.792029013" watchObservedRunningTime="2025-11-21 14:11:14.509540508 +0000 UTC m=+151.794133983" Nov 21 14:11:14 crc kubenswrapper[4897]: E1121 14:11:14.510576 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-21 14:11:15.010562517 +0000 UTC m=+152.295155992 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5z687" (UID: "9f63b8dd-ffb7-4963-8fcc-0c133a2c045b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:14 crc kubenswrapper[4897]: I1121 14:11:14.529434 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kf7zh" podStartSLOduration=125.52941855 podStartE2EDuration="2m5.52941855s" podCreationTimestamp="2025-11-21 14:09:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:11:14.526697693 +0000 UTC m=+151.811291168" watchObservedRunningTime="2025-11-21 14:11:14.52941855 +0000 UTC m=+151.814012025" Nov 21 14:11:14 crc kubenswrapper[4897]: I1121 14:11:14.567868 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6sfv4" podStartSLOduration=125.567848707 podStartE2EDuration="2m5.567848707s" podCreationTimestamp="2025-11-21 14:09:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:11:14.556979939 +0000 UTC m=+151.841573414" watchObservedRunningTime="2025-11-21 14:11:14.567848707 +0000 UTC m=+151.852442182" Nov 21 14:11:14 crc kubenswrapper[4897]: I1121 14:11:14.612313 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 21 14:11:14 crc kubenswrapper[4897]: E1121 14:11:14.612496 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-21 14:11:15.112463147 +0000 UTC m=+152.397056622 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:14 crc kubenswrapper[4897]: I1121 14:11:14.612606 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:14 crc kubenswrapper[4897]: E1121 14:11:14.613231 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-21 14:11:15.113209059 +0000 UTC m=+152.397802534 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5z687" (UID: "9f63b8dd-ffb7-4963-8fcc-0c133a2c045b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:14 crc kubenswrapper[4897]: I1121 14:11:14.650887 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-d8pkd" Nov 21 14:11:14 crc kubenswrapper[4897]: I1121 14:11:14.650953 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-d8pkd" Nov 21 14:11:14 crc kubenswrapper[4897]: I1121 14:11:14.653383 4897 patch_prober.go:28] interesting pod/apiserver-7bbb656c7d-d8pkd container/oauth-apiserver namespace/openshift-oauth-apiserver: Startup probe status=failure output="Get \"https://10.217.0.5:8443/livez\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Nov 21 14:11:14 crc kubenswrapper[4897]: I1121 14:11:14.653484 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-d8pkd" podUID="911a0a7e-7a63-4b4a-8b74-6c21f8da7e31" containerName="oauth-apiserver" probeResult="failure" output="Get \"https://10.217.0.5:8443/livez\": dial tcp 10.217.0.5:8443: connect: connection refused" Nov 21 14:11:14 crc kubenswrapper[4897]: I1121 14:11:14.714109 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 21 14:11:14 crc kubenswrapper[4897]: E1121 14:11:14.714474 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-21 14:11:15.214442621 +0000 UTC m=+152.499036106 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:14 crc kubenswrapper[4897]: I1121 14:11:14.714649 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:14 crc kubenswrapper[4897]: E1121 14:11:14.715066 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-21 14:11:15.215049899 +0000 UTC m=+152.499643374 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5z687" (UID: "9f63b8dd-ffb7-4963-8fcc-0c133a2c045b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:14 crc kubenswrapper[4897]: I1121 14:11:14.815326 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 21 14:11:14 crc kubenswrapper[4897]: E1121 14:11:14.815548 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-21 14:11:15.315512818 +0000 UTC m=+152.600106293 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:14 crc kubenswrapper[4897]: I1121 14:11:14.817928 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:14 crc kubenswrapper[4897]: E1121 14:11:14.818386 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-21 14:11:15.318362919 +0000 UTC m=+152.602956474 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5z687" (UID: "9f63b8dd-ffb7-4963-8fcc-0c133a2c045b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:14 crc kubenswrapper[4897]: I1121 14:11:14.918891 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 21 14:11:14 crc kubenswrapper[4897]: E1121 14:11:14.919247 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-21 14:11:15.419211581 +0000 UTC m=+152.703805076 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:14 crc kubenswrapper[4897]: I1121 14:11:14.993765 4897 patch_prober.go:28] interesting pod/router-default-5444994796-2kgg2 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 21 14:11:14 crc kubenswrapper[4897]: [-]has-synced failed: reason withheld Nov 21 14:11:14 crc kubenswrapper[4897]: [+]process-running ok Nov 21 14:11:14 crc kubenswrapper[4897]: healthz check failed Nov 21 14:11:14 crc kubenswrapper[4897]: I1121 14:11:14.993891 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-2kgg2" podUID="e095a18e-3a0b-4418-95a8-a2cc3de5bd7b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 21 14:11:15 crc kubenswrapper[4897]: I1121 14:11:15.020481 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:15 crc kubenswrapper[4897]: E1121 14:11:15.021550 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-21 14:11:15.521499833 +0000 UTC m=+152.806093298 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5z687" (UID: "9f63b8dd-ffb7-4963-8fcc-0c133a2c045b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:15 crc kubenswrapper[4897]: I1121 14:11:15.122198 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 21 14:11:15 crc kubenswrapper[4897]: E1121 14:11:15.122786 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-21 14:11:15.622766806 +0000 UTC m=+152.907360281 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:15 crc kubenswrapper[4897]: I1121 14:11:15.224679 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:15 crc kubenswrapper[4897]: E1121 14:11:15.225317 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-21 14:11:15.725288005 +0000 UTC m=+153.009881680 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5z687" (UID: "9f63b8dd-ffb7-4963-8fcc-0c133a2c045b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:15 crc kubenswrapper[4897]: I1121 14:11:15.326205 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 21 14:11:15 crc kubenswrapper[4897]: E1121 14:11:15.326552 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-21 14:11:15.826531897 +0000 UTC m=+153.111125362 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:15 crc kubenswrapper[4897]: I1121 14:11:15.433946 4897 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-4cx2q container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.24:8443/healthz\": dial tcp 10.217.0.24:8443: connect: connection refused" start-of-body= Nov 21 14:11:15 crc kubenswrapper[4897]: I1121 14:11:15.434036 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4cx2q" podUID="56521238-a49d-446a-b4a9-e2809ff15826" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.24:8443/healthz\": dial tcp 10.217.0.24:8443: connect: connection refused" Nov 21 14:11:15 crc kubenswrapper[4897]: I1121 14:11:15.434442 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-cc48c" Nov 21 14:11:15 crc kubenswrapper[4897]: I1121 14:11:15.435225 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:15 crc kubenswrapper[4897]: E1121 14:11:15.435831 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-21 14:11:15.935814957 +0000 UTC m=+153.220408432 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5z687" (UID: "9f63b8dd-ffb7-4963-8fcc-0c133a2c045b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:15 crc kubenswrapper[4897]: I1121 14:11:15.437085 4897 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-z9hjh container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.30:8080/healthz\": dial tcp 10.217.0.30:8080: connect: connection refused" start-of-body= Nov 21 14:11:15 crc kubenswrapper[4897]: I1121 14:11:15.437158 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-z9hjh" podUID="e128526a-8489-47e1-a711-4977f2b0d46d" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.30:8080/healthz\": dial tcp 10.217.0.30:8080: connect: connection refused" Nov 21 14:11:15 crc kubenswrapper[4897]: I1121 14:11:15.437238 4897 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-6sfv4 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.29:5443/healthz\": dial tcp 10.217.0.29:5443: connect: connection refused" start-of-body= Nov 21 14:11:15 crc kubenswrapper[4897]: I1121 14:11:15.437344 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6sfv4" podUID="11301ff8-fa01-48da-a95e-fc244fd0a575" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.29:5443/healthz\": dial tcp 10.217.0.29:5443: connect: connection refused" Nov 21 14:11:15 crc kubenswrapper[4897]: I1121 14:11:15.457569 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2hmqv" podStartSLOduration=126.457533652 podStartE2EDuration="2m6.457533652s" podCreationTimestamp="2025-11-21 14:09:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:11:15.45039524 +0000 UTC m=+152.734988715" watchObservedRunningTime="2025-11-21 14:11:15.457533652 +0000 UTC m=+152.742127117" Nov 21 14:11:15 crc kubenswrapper[4897]: I1121 14:11:15.475956 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-cc48c" podStartSLOduration=126.475936312 podStartE2EDuration="2m6.475936312s" podCreationTimestamp="2025-11-21 14:09:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:11:15.472766392 +0000 UTC m=+152.757359857" watchObservedRunningTime="2025-11-21 14:11:15.475936312 +0000 UTC m=+152.760529787" Nov 21 14:11:15 crc kubenswrapper[4897]: I1121 14:11:15.496096 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-t5jsz" podStartSLOduration=127.496073762 podStartE2EDuration="2m7.496073762s" podCreationTimestamp="2025-11-21 14:09:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:11:15.495520216 +0000 UTC m=+152.780113691" watchObservedRunningTime="2025-11-21 14:11:15.496073762 +0000 UTC m=+152.780667227" Nov 21 14:11:15 crc kubenswrapper[4897]: I1121 14:11:15.536968 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 21 14:11:15 crc kubenswrapper[4897]: E1121 14:11:15.537718 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-21 14:11:16.037668077 +0000 UTC m=+153.322261562 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:15 crc kubenswrapper[4897]: I1121 14:11:15.638862 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:15 crc kubenswrapper[4897]: E1121 14:11:15.639293 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-21 14:11:16.139278131 +0000 UTC m=+153.423871606 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5z687" (UID: "9f63b8dd-ffb7-4963-8fcc-0c133a2c045b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:15 crc kubenswrapper[4897]: I1121 14:11:15.739902 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 21 14:11:15 crc kubenswrapper[4897]: E1121 14:11:15.740112 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-21 14:11:16.240080011 +0000 UTC m=+153.524673476 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:15 crc kubenswrapper[4897]: I1121 14:11:15.740449 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:15 crc kubenswrapper[4897]: E1121 14:11:15.740819 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-21 14:11:16.240803941 +0000 UTC m=+153.525397416 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5z687" (UID: "9f63b8dd-ffb7-4963-8fcc-0c133a2c045b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:15 crc kubenswrapper[4897]: I1121 14:11:15.841969 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 21 14:11:15 crc kubenswrapper[4897]: E1121 14:11:15.842119 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-21 14:11:16.342089075 +0000 UTC m=+153.626682550 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:15 crc kubenswrapper[4897]: I1121 14:11:15.842354 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:15 crc kubenswrapper[4897]: E1121 14:11:15.842752 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-21 14:11:16.342737773 +0000 UTC m=+153.627331248 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5z687" (UID: "9f63b8dd-ffb7-4963-8fcc-0c133a2c045b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:15 crc kubenswrapper[4897]: I1121 14:11:15.943425 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 21 14:11:15 crc kubenswrapper[4897]: E1121 14:11:15.943671 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-21 14:11:16.443637726 +0000 UTC m=+153.728231201 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:15 crc kubenswrapper[4897]: I1121 14:11:15.944232 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:15 crc kubenswrapper[4897]: E1121 14:11:15.944640 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-21 14:11:16.444631304 +0000 UTC m=+153.729224869 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5z687" (UID: "9f63b8dd-ffb7-4963-8fcc-0c133a2c045b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:15 crc kubenswrapper[4897]: I1121 14:11:15.993091 4897 patch_prober.go:28] interesting pod/router-default-5444994796-2kgg2 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 21 14:11:15 crc kubenswrapper[4897]: [-]has-synced failed: reason withheld Nov 21 14:11:15 crc kubenswrapper[4897]: [+]process-running ok Nov 21 14:11:15 crc kubenswrapper[4897]: healthz check failed Nov 21 14:11:15 crc kubenswrapper[4897]: I1121 14:11:15.993174 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-2kgg2" podUID="e095a18e-3a0b-4418-95a8-a2cc3de5bd7b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 21 14:11:16 crc kubenswrapper[4897]: I1121 14:11:16.045973 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 21 14:11:16 crc kubenswrapper[4897]: E1121 14:11:16.046555 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-21 14:11:16.546486404 +0000 UTC m=+153.831079889 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:16 crc kubenswrapper[4897]: I1121 14:11:16.104782 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" Nov 21 14:11:16 crc kubenswrapper[4897]: I1121 14:11:16.149597 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:16 crc kubenswrapper[4897]: E1121 14:11:16.150011 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-21 14:11:16.64997005 +0000 UTC m=+153.934563515 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5z687" (UID: "9f63b8dd-ffb7-4963-8fcc-0c133a2c045b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:16 crc kubenswrapper[4897]: I1121 14:11:16.251656 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 21 14:11:16 crc kubenswrapper[4897]: E1121 14:11:16.252272 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-21 14:11:16.752249122 +0000 UTC m=+154.036842597 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:16 crc kubenswrapper[4897]: I1121 14:11:16.354095 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:16 crc kubenswrapper[4897]: E1121 14:11:16.354700 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-21 14:11:16.854677528 +0000 UTC m=+154.139271003 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5z687" (UID: "9f63b8dd-ffb7-4963-8fcc-0c133a2c045b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:16 crc kubenswrapper[4897]: I1121 14:11:16.439234 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-vf46x" event={"ID":"f2080446-c5c3-4c09-b8ae-54359bbb72ba","Type":"ContainerStarted","Data":"53d897ed05d8ff2f473ac8b9c8a72991c25ab1215df778307a048f9767e9d8d2"} Nov 21 14:11:16 crc kubenswrapper[4897]: I1121 14:11:16.440791 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-74phw" event={"ID":"f7d61c37-621e-4812-bbaa-74006927176a","Type":"ContainerStarted","Data":"a49bab3a4673e7fc6cda745cf9202a243785cbf2170bc82c347dcebe3d067ccd"} Nov 21 14:11:16 crc kubenswrapper[4897]: I1121 14:11:16.442299 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-s7tww" event={"ID":"027b031c-1b21-40e8-aa7c-4a342aae265a","Type":"ContainerStarted","Data":"ddd695113de8ff63b5565063cd2771f850257c356d2ff77965c1765ade28cae7"} Nov 21 14:11:16 crc kubenswrapper[4897]: I1121 14:11:16.443732 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-lq8bp" event={"ID":"20a1d330-d99c-480d-b44c-4d8a5b4c6453","Type":"ContainerStarted","Data":"0d2036844bc1a98a1fbf8b4ed95d74e908a27f7686abad195a725b9d7c6a58fc"} Nov 21 14:11:16 crc kubenswrapper[4897]: I1121 14:11:16.444379 4897 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-z9hjh container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.30:8080/healthz\": dial tcp 10.217.0.30:8080: connect: connection refused" start-of-body= Nov 21 14:11:16 crc kubenswrapper[4897]: I1121 14:11:16.444443 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-z9hjh" podUID="e128526a-8489-47e1-a711-4977f2b0d46d" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.30:8080/healthz\": dial tcp 10.217.0.30:8080: connect: connection refused" Nov 21 14:11:16 crc kubenswrapper[4897]: I1121 14:11:16.445013 4897 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-6sfv4 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.29:5443/healthz\": dial tcp 10.217.0.29:5443: connect: connection refused" start-of-body= Nov 21 14:11:16 crc kubenswrapper[4897]: I1121 14:11:16.445067 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6sfv4" podUID="11301ff8-fa01-48da-a95e-fc244fd0a575" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.29:5443/healthz\": dial tcp 10.217.0.29:5443: connect: connection refused" Nov 21 14:11:16 crc kubenswrapper[4897]: I1121 14:11:16.455688 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 21 14:11:16 crc kubenswrapper[4897]: E1121 14:11:16.455927 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-21 14:11:16.955877619 +0000 UTC m=+154.240471094 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:16 crc kubenswrapper[4897]: I1121 14:11:16.456276 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:16 crc kubenswrapper[4897]: E1121 14:11:16.456677 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-21 14:11:16.956659962 +0000 UTC m=+154.241253437 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5z687" (UID: "9f63b8dd-ffb7-4963-8fcc-0c133a2c045b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:16 crc kubenswrapper[4897]: I1121 14:11:16.461003 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-vf46x" podStartSLOduration=127.460979184 podStartE2EDuration="2m7.460979184s" podCreationTimestamp="2025-11-21 14:09:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:11:16.459857262 +0000 UTC m=+153.744450737" watchObservedRunningTime="2025-11-21 14:11:16.460979184 +0000 UTC m=+153.745572659" Nov 21 14:11:16 crc kubenswrapper[4897]: I1121 14:11:16.502991 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-74phw" podStartSLOduration=127.50291924 podStartE2EDuration="2m7.50291924s" podCreationTimestamp="2025-11-21 14:09:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:11:16.502633331 +0000 UTC m=+153.787226806" watchObservedRunningTime="2025-11-21 14:11:16.50291924 +0000 UTC m=+153.787512715" Nov 21 14:11:16 crc kubenswrapper[4897]: I1121 14:11:16.557406 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 21 14:11:16 crc kubenswrapper[4897]: E1121 14:11:16.557664 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-21 14:11:17.057626616 +0000 UTC m=+154.342220091 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:16 crc kubenswrapper[4897]: I1121 14:11:16.559418 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:16 crc kubenswrapper[4897]: E1121 14:11:16.560112 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-21 14:11:17.060098285 +0000 UTC m=+154.344691760 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5z687" (UID: "9f63b8dd-ffb7-4963-8fcc-0c133a2c045b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:16 crc kubenswrapper[4897]: I1121 14:11:16.661047 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 21 14:11:16 crc kubenswrapper[4897]: E1121 14:11:16.661167 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-21 14:11:17.161144263 +0000 UTC m=+154.445737728 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:16 crc kubenswrapper[4897]: I1121 14:11:16.661487 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:16 crc kubenswrapper[4897]: E1121 14:11:16.661993 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-21 14:11:17.161970316 +0000 UTC m=+154.446563791 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5z687" (UID: "9f63b8dd-ffb7-4963-8fcc-0c133a2c045b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:16 crc kubenswrapper[4897]: I1121 14:11:16.762917 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 21 14:11:16 crc kubenswrapper[4897]: E1121 14:11:16.763173 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-21 14:11:17.263136566 +0000 UTC m=+154.547730041 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:16 crc kubenswrapper[4897]: I1121 14:11:16.763580 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:16 crc kubenswrapper[4897]: E1121 14:11:16.763934 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-21 14:11:17.263918819 +0000 UTC m=+154.548512294 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5z687" (UID: "9f63b8dd-ffb7-4963-8fcc-0c133a2c045b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:16 crc kubenswrapper[4897]: I1121 14:11:16.864750 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 21 14:11:16 crc kubenswrapper[4897]: E1121 14:11:16.865358 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-21 14:11:17.365333536 +0000 UTC m=+154.649927001 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:16 crc kubenswrapper[4897]: I1121 14:11:16.865471 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:16 crc kubenswrapper[4897]: E1121 14:11:16.865957 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-21 14:11:17.365950174 +0000 UTC m=+154.650543649 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5z687" (UID: "9f63b8dd-ffb7-4963-8fcc-0c133a2c045b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:16 crc kubenswrapper[4897]: I1121 14:11:16.909905 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-t5jsz" Nov 21 14:11:16 crc kubenswrapper[4897]: I1121 14:11:16.911372 4897 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-t5jsz container/openshift-config-operator namespace/openshift-config-operator: Liveness probe status=failure output="Get \"https://10.217.0.22:8443/healthz\": dial tcp 10.217.0.22:8443: connect: connection refused" start-of-body= Nov 21 14:11:16 crc kubenswrapper[4897]: I1121 14:11:16.911417 4897 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-t5jsz" podUID="00cdea84-a65d-4422-8ecb-7bc8b912c557" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.22:8443/healthz\": dial tcp 10.217.0.22:8443: connect: connection refused" Nov 21 14:11:16 crc kubenswrapper[4897]: I1121 14:11:16.911713 4897 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-t5jsz container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.22:8443/healthz\": dial tcp 10.217.0.22:8443: connect: connection refused" start-of-body= Nov 21 14:11:16 crc kubenswrapper[4897]: I1121 14:11:16.911738 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-t5jsz" podUID="00cdea84-a65d-4422-8ecb-7bc8b912c557" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.22:8443/healthz\": dial tcp 10.217.0.22:8443: connect: connection refused" Nov 21 14:11:16 crc kubenswrapper[4897]: I1121 14:11:16.911986 4897 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-t5jsz container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.22:8443/healthz\": dial tcp 10.217.0.22:8443: connect: connection refused" start-of-body= Nov 21 14:11:16 crc kubenswrapper[4897]: I1121 14:11:16.912013 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-t5jsz" podUID="00cdea84-a65d-4422-8ecb-7bc8b912c557" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.22:8443/healthz\": dial tcp 10.217.0.22:8443: connect: connection refused" Nov 21 14:11:16 crc kubenswrapper[4897]: I1121 14:11:16.967262 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 21 14:11:16 crc kubenswrapper[4897]: E1121 14:11:16.968239 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-21 14:11:17.468197465 +0000 UTC m=+154.752790940 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:16 crc kubenswrapper[4897]: I1121 14:11:16.996376 4897 patch_prober.go:28] interesting pod/router-default-5444994796-2kgg2 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 21 14:11:16 crc kubenswrapper[4897]: [-]has-synced failed: reason withheld Nov 21 14:11:16 crc kubenswrapper[4897]: [+]process-running ok Nov 21 14:11:16 crc kubenswrapper[4897]: healthz check failed Nov 21 14:11:16 crc kubenswrapper[4897]: I1121 14:11:16.996492 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-2kgg2" podUID="e095a18e-3a0b-4418-95a8-a2cc3de5bd7b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 21 14:11:17 crc kubenswrapper[4897]: I1121 14:11:17.068971 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 21 14:11:17 crc kubenswrapper[4897]: I1121 14:11:17.069054 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 21 14:11:17 crc kubenswrapper[4897]: I1121 14:11:17.069116 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:17 crc kubenswrapper[4897]: E1121 14:11:17.069531 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-21 14:11:17.569512659 +0000 UTC m=+154.854106144 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5z687" (UID: "9f63b8dd-ffb7-4963-8fcc-0c133a2c045b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:17 crc kubenswrapper[4897]: I1121 14:11:17.071672 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 21 14:11:17 crc kubenswrapper[4897]: I1121 14:11:17.078319 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 21 14:11:17 crc kubenswrapper[4897]: I1121 14:11:17.170190 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 21 14:11:17 crc kubenswrapper[4897]: I1121 14:11:17.170444 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 21 14:11:17 crc kubenswrapper[4897]: I1121 14:11:17.170495 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 21 14:11:17 crc kubenswrapper[4897]: E1121 14:11:17.172114 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-21 14:11:17.67207878 +0000 UTC m=+154.956672255 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:17 crc kubenswrapper[4897]: I1121 14:11:17.176995 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 21 14:11:17 crc kubenswrapper[4897]: I1121 14:11:17.184156 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 21 14:11:17 crc kubenswrapper[4897]: I1121 14:11:17.271526 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:17 crc kubenswrapper[4897]: E1121 14:11:17.271997 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-21 14:11:17.771977474 +0000 UTC m=+155.056570949 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5z687" (UID: "9f63b8dd-ffb7-4963-8fcc-0c133a2c045b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:17 crc kubenswrapper[4897]: I1121 14:11:17.329357 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 21 14:11:17 crc kubenswrapper[4897]: I1121 14:11:17.346377 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 21 14:11:17 crc kubenswrapper[4897]: I1121 14:11:17.356723 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 21 14:11:17 crc kubenswrapper[4897]: I1121 14:11:17.372302 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 21 14:11:17 crc kubenswrapper[4897]: E1121 14:11:17.372491 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-21 14:11:17.872450395 +0000 UTC m=+155.157043870 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:17 crc kubenswrapper[4897]: I1121 14:11:17.372823 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:17 crc kubenswrapper[4897]: E1121 14:11:17.373777 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-21 14:11:17.873762822 +0000 UTC m=+155.158356297 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5z687" (UID: "9f63b8dd-ffb7-4963-8fcc-0c133a2c045b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:17 crc kubenswrapper[4897]: I1121 14:11:17.385901 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-lq8bp" Nov 21 14:11:17 crc kubenswrapper[4897]: I1121 14:11:17.465753 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-s7tww" event={"ID":"027b031c-1b21-40e8-aa7c-4a342aae265a","Type":"ContainerStarted","Data":"4e1d96efc81ee92c9801110edd69b85524b2bc7c996af8fec1092b78fef33aee"} Nov 21 14:11:17 crc kubenswrapper[4897]: I1121 14:11:17.474074 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 21 14:11:17 crc kubenswrapper[4897]: E1121 14:11:17.474297 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-21 14:11:17.974264443 +0000 UTC m=+155.258857918 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:17 crc kubenswrapper[4897]: I1121 14:11:17.474392 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:17 crc kubenswrapper[4897]: E1121 14:11:17.474751 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-21 14:11:17.974736127 +0000 UTC m=+155.259329592 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5z687" (UID: "9f63b8dd-ffb7-4963-8fcc-0c133a2c045b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:17 crc kubenswrapper[4897]: I1121 14:11:17.494354 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-lq8bp" podStartSLOduration=9.494327761 podStartE2EDuration="9.494327761s" podCreationTimestamp="2025-11-21 14:11:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:11:16.528742719 +0000 UTC m=+153.813336204" watchObservedRunningTime="2025-11-21 14:11:17.494327761 +0000 UTC m=+154.778921246" Nov 21 14:11:17 crc kubenswrapper[4897]: I1121 14:11:17.496076 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-s7tww" podStartSLOduration=128.49606779 podStartE2EDuration="2m8.49606779s" podCreationTimestamp="2025-11-21 14:09:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:11:17.491955854 +0000 UTC m=+154.776549329" watchObservedRunningTime="2025-11-21 14:11:17.49606779 +0000 UTC m=+154.780661265" Nov 21 14:11:17 crc kubenswrapper[4897]: I1121 14:11:17.575161 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 21 14:11:17 crc kubenswrapper[4897]: E1121 14:11:17.575383 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-21 14:11:18.075348322 +0000 UTC m=+155.359941797 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:17 crc kubenswrapper[4897]: I1121 14:11:17.575957 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:17 crc kubenswrapper[4897]: E1121 14:11:17.577824 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-21 14:11:18.077802211 +0000 UTC m=+155.362395686 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5z687" (UID: "9f63b8dd-ffb7-4963-8fcc-0c133a2c045b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:17 crc kubenswrapper[4897]: I1121 14:11:17.677227 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 21 14:11:17 crc kubenswrapper[4897]: E1121 14:11:17.677766 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-21 14:11:18.177742207 +0000 UTC m=+155.462335682 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:17 crc kubenswrapper[4897]: I1121 14:11:17.779280 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:17 crc kubenswrapper[4897]: E1121 14:11:17.779669 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-21 14:11:18.279654908 +0000 UTC m=+155.564248373 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5z687" (UID: "9f63b8dd-ffb7-4963-8fcc-0c133a2c045b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:17 crc kubenswrapper[4897]: I1121 14:11:17.880340 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 21 14:11:17 crc kubenswrapper[4897]: E1121 14:11:17.880573 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-21 14:11:18.380540811 +0000 UTC m=+155.665134276 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:17 crc kubenswrapper[4897]: I1121 14:11:17.880643 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:17 crc kubenswrapper[4897]: E1121 14:11:17.881065 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-21 14:11:18.381045175 +0000 UTC m=+155.665638710 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5z687" (UID: "9f63b8dd-ffb7-4963-8fcc-0c133a2c045b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:17 crc kubenswrapper[4897]: I1121 14:11:17.982147 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 21 14:11:17 crc kubenswrapper[4897]: E1121 14:11:17.982376 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-21 14:11:18.482340939 +0000 UTC m=+155.766934414 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:17 crc kubenswrapper[4897]: I1121 14:11:17.982574 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:17 crc kubenswrapper[4897]: E1121 14:11:17.982915 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-21 14:11:18.482907685 +0000 UTC m=+155.767501160 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5z687" (UID: "9f63b8dd-ffb7-4963-8fcc-0c133a2c045b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:17 crc kubenswrapper[4897]: I1121 14:11:17.997336 4897 patch_prober.go:28] interesting pod/router-default-5444994796-2kgg2 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 21 14:11:17 crc kubenswrapper[4897]: [-]has-synced failed: reason withheld Nov 21 14:11:17 crc kubenswrapper[4897]: [+]process-running ok Nov 21 14:11:17 crc kubenswrapper[4897]: healthz check failed Nov 21 14:11:17 crc kubenswrapper[4897]: I1121 14:11:17.997415 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-2kgg2" podUID="e095a18e-3a0b-4418-95a8-a2cc3de5bd7b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 21 14:11:18 crc kubenswrapper[4897]: I1121 14:11:18.084275 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 21 14:11:18 crc kubenswrapper[4897]: E1121 14:11:18.084475 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-21 14:11:18.584433305 +0000 UTC m=+155.869026780 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:18 crc kubenswrapper[4897]: I1121 14:11:18.084655 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:18 crc kubenswrapper[4897]: E1121 14:11:18.084969 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-21 14:11:18.584956761 +0000 UTC m=+155.869550236 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5z687" (UID: "9f63b8dd-ffb7-4963-8fcc-0c133a2c045b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:18 crc kubenswrapper[4897]: W1121 14:11:18.112438 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b6479f0_333b_4a96_9adf_2099afdc2447.slice/crio-e0cc8c53dc65762f35fa24122098425eb3e5a0805f01f613934ad1c6f7960bb9 WatchSource:0}: Error finding container e0cc8c53dc65762f35fa24122098425eb3e5a0805f01f613934ad1c6f7960bb9: Status 404 returned error can't find the container with id e0cc8c53dc65762f35fa24122098425eb3e5a0805f01f613934ad1c6f7960bb9 Nov 21 14:11:18 crc kubenswrapper[4897]: W1121 14:11:18.188232 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d751cbb_f2e2_430d_9754_c882a5e924a5.slice/crio-bf8c9ac66b84bf6eaf59f43b8c772d648f383bbeeeb1f91644b568c3f1811ddd WatchSource:0}: Error finding container bf8c9ac66b84bf6eaf59f43b8c772d648f383bbeeeb1f91644b568c3f1811ddd: Status 404 returned error can't find the container with id bf8c9ac66b84bf6eaf59f43b8c772d648f383bbeeeb1f91644b568c3f1811ddd Nov 21 14:11:18 crc kubenswrapper[4897]: I1121 14:11:18.188680 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 21 14:11:18 crc kubenswrapper[4897]: E1121 14:11:18.189042 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-21 14:11:18.689026353 +0000 UTC m=+155.973619828 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:18 crc kubenswrapper[4897]: I1121 14:11:18.290031 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:18 crc kubenswrapper[4897]: E1121 14:11:18.290671 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-21 14:11:18.790647626 +0000 UTC m=+156.075241101 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5z687" (UID: "9f63b8dd-ffb7-4963-8fcc-0c133a2c045b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:18 crc kubenswrapper[4897]: I1121 14:11:18.397640 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 21 14:11:18 crc kubenswrapper[4897]: E1121 14:11:18.397864 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-21 14:11:18.897830067 +0000 UTC m=+156.182423542 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:18 crc kubenswrapper[4897]: I1121 14:11:18.397905 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:18 crc kubenswrapper[4897]: E1121 14:11:18.398296 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-21 14:11:18.898283909 +0000 UTC m=+156.182877454 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5z687" (UID: "9f63b8dd-ffb7-4963-8fcc-0c133a2c045b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:18 crc kubenswrapper[4897]: I1121 14:11:18.472525 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-bjdjf" event={"ID":"78db7e2e-ba45-464e-9b1d-0fc61191955a","Type":"ContainerStarted","Data":"aa606e2c8d38bdcc7883905b840825a781d1c0dabbd2987b95c118f05154ed6b"} Nov 21 14:11:18 crc kubenswrapper[4897]: I1121 14:11:18.474090 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"46ae2a3b88274a3d731bd0938560394c75d9fc036b4122472ddf0dc26bf472bb"} Nov 21 14:11:18 crc kubenswrapper[4897]: I1121 14:11:18.474121 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"4ad70ee69b72523080e93ddf5c50a8806e65466023c8acc2db28f0c980c4d29e"} Nov 21 14:11:18 crc kubenswrapper[4897]: I1121 14:11:18.476159 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"cd0889b0f73a0ecda531c2b7e6d89cac1ccb0fa9d4a14b2290a713218cca3376"} Nov 21 14:11:18 crc kubenswrapper[4897]: I1121 14:11:18.476198 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"bf8c9ac66b84bf6eaf59f43b8c772d648f383bbeeeb1f91644b568c3f1811ddd"} Nov 21 14:11:18 crc kubenswrapper[4897]: I1121 14:11:18.478656 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"4de947cea4da50493f6fdc6c3547313e47e37f594ca35af34d12d714424c3041"} Nov 21 14:11:18 crc kubenswrapper[4897]: I1121 14:11:18.478685 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"e0cc8c53dc65762f35fa24122098425eb3e5a0805f01f613934ad1c6f7960bb9"} Nov 21 14:11:18 crc kubenswrapper[4897]: I1121 14:11:18.478975 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 21 14:11:18 crc kubenswrapper[4897]: I1121 14:11:18.497494 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 21 14:11:18 crc kubenswrapper[4897]: I1121 14:11:18.498407 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 21 14:11:18 crc kubenswrapper[4897]: E1121 14:11:18.498664 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-21 14:11:18.998650847 +0000 UTC m=+156.283244322 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:18 crc kubenswrapper[4897]: I1121 14:11:18.498796 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 21 14:11:18 crc kubenswrapper[4897]: I1121 14:11:18.500844 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Nov 21 14:11:18 crc kubenswrapper[4897]: I1121 14:11:18.501067 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Nov 21 14:11:18 crc kubenswrapper[4897]: I1121 14:11:18.514142 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 21 14:11:18 crc kubenswrapper[4897]: I1121 14:11:18.603620 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7cd98bfc-77a5-4ef5-86f1-7614b0b2ad2a-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"7cd98bfc-77a5-4ef5-86f1-7614b0b2ad2a\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 21 14:11:18 crc kubenswrapper[4897]: I1121 14:11:18.604761 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7cd98bfc-77a5-4ef5-86f1-7614b0b2ad2a-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"7cd98bfc-77a5-4ef5-86f1-7614b0b2ad2a\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 21 14:11:18 crc kubenswrapper[4897]: I1121 14:11:18.606809 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:18 crc kubenswrapper[4897]: E1121 14:11:18.611249 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-21 14:11:19.11122657 +0000 UTC m=+156.395820035 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5z687" (UID: "9f63b8dd-ffb7-4963-8fcc-0c133a2c045b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:18 crc kubenswrapper[4897]: I1121 14:11:18.711968 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 21 14:11:18 crc kubenswrapper[4897]: I1121 14:11:18.712252 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7cd98bfc-77a5-4ef5-86f1-7614b0b2ad2a-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"7cd98bfc-77a5-4ef5-86f1-7614b0b2ad2a\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 21 14:11:18 crc kubenswrapper[4897]: I1121 14:11:18.712294 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7cd98bfc-77a5-4ef5-86f1-7614b0b2ad2a-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"7cd98bfc-77a5-4ef5-86f1-7614b0b2ad2a\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 21 14:11:18 crc kubenswrapper[4897]: I1121 14:11:18.712370 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7cd98bfc-77a5-4ef5-86f1-7614b0b2ad2a-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"7cd98bfc-77a5-4ef5-86f1-7614b0b2ad2a\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 21 14:11:18 crc kubenswrapper[4897]: E1121 14:11:18.712439 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-21 14:11:19.212424762 +0000 UTC m=+156.497018237 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:18 crc kubenswrapper[4897]: I1121 14:11:18.744732 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7cd98bfc-77a5-4ef5-86f1-7614b0b2ad2a-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"7cd98bfc-77a5-4ef5-86f1-7614b0b2ad2a\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 21 14:11:18 crc kubenswrapper[4897]: I1121 14:11:18.814084 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:18 crc kubenswrapper[4897]: E1121 14:11:18.814522 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-21 14:11:19.314489488 +0000 UTC m=+156.599082963 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5z687" (UID: "9f63b8dd-ffb7-4963-8fcc-0c133a2c045b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:18 crc kubenswrapper[4897]: I1121 14:11:18.879798 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 21 14:11:18 crc kubenswrapper[4897]: I1121 14:11:18.915816 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 21 14:11:18 crc kubenswrapper[4897]: E1121 14:11:18.916011 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-21 14:11:19.415980197 +0000 UTC m=+156.700573672 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:18 crc kubenswrapper[4897]: I1121 14:11:18.916526 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:18 crc kubenswrapper[4897]: E1121 14:11:18.916893 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-21 14:11:19.416884953 +0000 UTC m=+156.701478428 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5z687" (UID: "9f63b8dd-ffb7-4963-8fcc-0c133a2c045b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:19 crc kubenswrapper[4897]: I1121 14:11:19.007857 4897 patch_prober.go:28] interesting pod/router-default-5444994796-2kgg2 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 21 14:11:19 crc kubenswrapper[4897]: [-]has-synced failed: reason withheld Nov 21 14:11:19 crc kubenswrapper[4897]: [+]process-running ok Nov 21 14:11:19 crc kubenswrapper[4897]: healthz check failed Nov 21 14:11:19 crc kubenswrapper[4897]: I1121 14:11:19.007921 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-2kgg2" podUID="e095a18e-3a0b-4418-95a8-a2cc3de5bd7b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 21 14:11:19 crc kubenswrapper[4897]: I1121 14:11:19.019320 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 21 14:11:19 crc kubenswrapper[4897]: E1121 14:11:19.019692 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-21 14:11:19.519671049 +0000 UTC m=+156.804264524 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:19 crc kubenswrapper[4897]: I1121 14:11:19.123130 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:19 crc kubenswrapper[4897]: E1121 14:11:19.123460 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-21 14:11:19.623447083 +0000 UTC m=+156.908040558 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5z687" (UID: "9f63b8dd-ffb7-4963-8fcc-0c133a2c045b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:19 crc kubenswrapper[4897]: I1121 14:11:19.223656 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 21 14:11:19 crc kubenswrapper[4897]: E1121 14:11:19.223834 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-21 14:11:19.723805161 +0000 UTC m=+157.008398636 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:19 crc kubenswrapper[4897]: I1121 14:11:19.224269 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:19 crc kubenswrapper[4897]: E1121 14:11:19.224603 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-21 14:11:19.724590462 +0000 UTC m=+157.009183937 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5z687" (UID: "9f63b8dd-ffb7-4963-8fcc-0c133a2c045b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:19 crc kubenswrapper[4897]: I1121 14:11:19.246399 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 21 14:11:19 crc kubenswrapper[4897]: W1121 14:11:19.261939 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod7cd98bfc_77a5_4ef5_86f1_7614b0b2ad2a.slice/crio-eb6f02f2c6ecbb8ee1c30e3fc050f7b4e24b3f424c1396bb90f147004f2196fb WatchSource:0}: Error finding container eb6f02f2c6ecbb8ee1c30e3fc050f7b4e24b3f424c1396bb90f147004f2196fb: Status 404 returned error can't find the container with id eb6f02f2c6ecbb8ee1c30e3fc050f7b4e24b3f424c1396bb90f147004f2196fb Nov 21 14:11:19 crc kubenswrapper[4897]: I1121 14:11:19.317831 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-q6dd2"] Nov 21 14:11:19 crc kubenswrapper[4897]: I1121 14:11:19.318844 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-q6dd2" Nov 21 14:11:19 crc kubenswrapper[4897]: I1121 14:11:19.326956 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 21 14:11:19 crc kubenswrapper[4897]: I1121 14:11:19.327087 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 21 14:11:19 crc kubenswrapper[4897]: E1121 14:11:19.327131 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-21 14:11:19.827099742 +0000 UTC m=+157.111693217 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:19 crc kubenswrapper[4897]: I1121 14:11:19.327179 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:19 crc kubenswrapper[4897]: E1121 14:11:19.327555 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-21 14:11:19.827540044 +0000 UTC m=+157.112133519 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5z687" (UID: "9f63b8dd-ffb7-4963-8fcc-0c133a2c045b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:19 crc kubenswrapper[4897]: I1121 14:11:19.342595 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-q6dd2"] Nov 21 14:11:19 crc kubenswrapper[4897]: I1121 14:11:19.401626 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 21 14:11:19 crc kubenswrapper[4897]: I1121 14:11:19.402224 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 21 14:11:19 crc kubenswrapper[4897]: I1121 14:11:19.405182 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Nov 21 14:11:19 crc kubenswrapper[4897]: I1121 14:11:19.406565 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Nov 21 14:11:19 crc kubenswrapper[4897]: I1121 14:11:19.419180 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 21 14:11:19 crc kubenswrapper[4897]: I1121 14:11:19.428435 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 21 14:11:19 crc kubenswrapper[4897]: I1121 14:11:19.428771 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1dc1b877-a8de-45b0-bab4-4411ecced04d-catalog-content\") pod \"certified-operators-q6dd2\" (UID: \"1dc1b877-a8de-45b0-bab4-4411ecced04d\") " pod="openshift-marketplace/certified-operators-q6dd2" Nov 21 14:11:19 crc kubenswrapper[4897]: I1121 14:11:19.428815 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1dc1b877-a8de-45b0-bab4-4411ecced04d-utilities\") pod \"certified-operators-q6dd2\" (UID: \"1dc1b877-a8de-45b0-bab4-4411ecced04d\") " pod="openshift-marketplace/certified-operators-q6dd2" Nov 21 14:11:19 crc kubenswrapper[4897]: I1121 14:11:19.428883 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l5xws\" (UniqueName: \"kubernetes.io/projected/1dc1b877-a8de-45b0-bab4-4411ecced04d-kube-api-access-l5xws\") pod \"certified-operators-q6dd2\" (UID: \"1dc1b877-a8de-45b0-bab4-4411ecced04d\") " pod="openshift-marketplace/certified-operators-q6dd2" Nov 21 14:11:19 crc kubenswrapper[4897]: E1121 14:11:19.428995 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-21 14:11:19.928976721 +0000 UTC m=+157.213570196 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:19 crc kubenswrapper[4897]: I1121 14:11:19.508425 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"7cd98bfc-77a5-4ef5-86f1-7614b0b2ad2a","Type":"ContainerStarted","Data":"eb6f02f2c6ecbb8ee1c30e3fc050f7b4e24b3f424c1396bb90f147004f2196fb"} Nov 21 14:11:19 crc kubenswrapper[4897]: I1121 14:11:19.515973 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-tqhs9"] Nov 21 14:11:19 crc kubenswrapper[4897]: I1121 14:11:19.517278 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tqhs9" Nov 21 14:11:19 crc kubenswrapper[4897]: I1121 14:11:19.526177 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 21 14:11:19 crc kubenswrapper[4897]: I1121 14:11:19.530058 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l5xws\" (UniqueName: \"kubernetes.io/projected/1dc1b877-a8de-45b0-bab4-4411ecced04d-kube-api-access-l5xws\") pod \"certified-operators-q6dd2\" (UID: \"1dc1b877-a8de-45b0-bab4-4411ecced04d\") " pod="openshift-marketplace/certified-operators-q6dd2" Nov 21 14:11:19 crc kubenswrapper[4897]: I1121 14:11:19.530139 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b5aa94f1-611d-4695-a3c7-1c33c80f1a1a-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"b5aa94f1-611d-4695-a3c7-1c33c80f1a1a\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 21 14:11:19 crc kubenswrapper[4897]: I1121 14:11:19.530215 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:19 crc kubenswrapper[4897]: I1121 14:11:19.530251 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1dc1b877-a8de-45b0-bab4-4411ecced04d-catalog-content\") pod \"certified-operators-q6dd2\" (UID: \"1dc1b877-a8de-45b0-bab4-4411ecced04d\") " pod="openshift-marketplace/certified-operators-q6dd2" Nov 21 14:11:19 crc kubenswrapper[4897]: I1121 14:11:19.530284 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1dc1b877-a8de-45b0-bab4-4411ecced04d-utilities\") pod \"certified-operators-q6dd2\" (UID: \"1dc1b877-a8de-45b0-bab4-4411ecced04d\") " pod="openshift-marketplace/certified-operators-q6dd2" Nov 21 14:11:19 crc kubenswrapper[4897]: I1121 14:11:19.530333 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b5aa94f1-611d-4695-a3c7-1c33c80f1a1a-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"b5aa94f1-611d-4695-a3c7-1c33c80f1a1a\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 21 14:11:19 crc kubenswrapper[4897]: E1121 14:11:19.530910 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-21 14:11:20.030892313 +0000 UTC m=+157.315485788 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5z687" (UID: "9f63b8dd-ffb7-4963-8fcc-0c133a2c045b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:19 crc kubenswrapper[4897]: I1121 14:11:19.531094 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1dc1b877-a8de-45b0-bab4-4411ecced04d-catalog-content\") pod \"certified-operators-q6dd2\" (UID: \"1dc1b877-a8de-45b0-bab4-4411ecced04d\") " pod="openshift-marketplace/certified-operators-q6dd2" Nov 21 14:11:19 crc kubenswrapper[4897]: I1121 14:11:19.531138 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1dc1b877-a8de-45b0-bab4-4411ecced04d-utilities\") pod \"certified-operators-q6dd2\" (UID: \"1dc1b877-a8de-45b0-bab4-4411ecced04d\") " pod="openshift-marketplace/certified-operators-q6dd2" Nov 21 14:11:19 crc kubenswrapper[4897]: I1121 14:11:19.533767 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tqhs9"] Nov 21 14:11:19 crc kubenswrapper[4897]: I1121 14:11:19.590479 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l5xws\" (UniqueName: \"kubernetes.io/projected/1dc1b877-a8de-45b0-bab4-4411ecced04d-kube-api-access-l5xws\") pod \"certified-operators-q6dd2\" (UID: \"1dc1b877-a8de-45b0-bab4-4411ecced04d\") " pod="openshift-marketplace/certified-operators-q6dd2" Nov 21 14:11:19 crc kubenswrapper[4897]: I1121 14:11:19.631010 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 21 14:11:19 crc kubenswrapper[4897]: I1121 14:11:19.631230 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b5aa94f1-611d-4695-a3c7-1c33c80f1a1a-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"b5aa94f1-611d-4695-a3c7-1c33c80f1a1a\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 21 14:11:19 crc kubenswrapper[4897]: I1121 14:11:19.631291 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/767a13d7-0e18-4496-b0bb-5a39292622e2-catalog-content\") pod \"community-operators-tqhs9\" (UID: \"767a13d7-0e18-4496-b0bb-5a39292622e2\") " pod="openshift-marketplace/community-operators-tqhs9" Nov 21 14:11:19 crc kubenswrapper[4897]: I1121 14:11:19.631334 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/767a13d7-0e18-4496-b0bb-5a39292622e2-utilities\") pod \"community-operators-tqhs9\" (UID: \"767a13d7-0e18-4496-b0bb-5a39292622e2\") " pod="openshift-marketplace/community-operators-tqhs9" Nov 21 14:11:19 crc kubenswrapper[4897]: I1121 14:11:19.631372 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b5aa94f1-611d-4695-a3c7-1c33c80f1a1a-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"b5aa94f1-611d-4695-a3c7-1c33c80f1a1a\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 21 14:11:19 crc kubenswrapper[4897]: I1121 14:11:19.631426 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-59rxs\" (UniqueName: \"kubernetes.io/projected/767a13d7-0e18-4496-b0bb-5a39292622e2-kube-api-access-59rxs\") pod \"community-operators-tqhs9\" (UID: \"767a13d7-0e18-4496-b0bb-5a39292622e2\") " pod="openshift-marketplace/community-operators-tqhs9" Nov 21 14:11:19 crc kubenswrapper[4897]: I1121 14:11:19.632535 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b5aa94f1-611d-4695-a3c7-1c33c80f1a1a-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"b5aa94f1-611d-4695-a3c7-1c33c80f1a1a\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 21 14:11:19 crc kubenswrapper[4897]: E1121 14:11:19.632623 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-21 14:11:20.132606389 +0000 UTC m=+157.417199924 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:19 crc kubenswrapper[4897]: I1121 14:11:19.643897 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-q6dd2" Nov 21 14:11:19 crc kubenswrapper[4897]: I1121 14:11:19.658433 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-dfltg" Nov 21 14:11:19 crc kubenswrapper[4897]: I1121 14:11:19.660425 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lclf5" Nov 21 14:11:19 crc kubenswrapper[4897]: I1121 14:11:19.663804 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-d8pkd" Nov 21 14:11:19 crc kubenswrapper[4897]: I1121 14:11:19.668542 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lclf5" Nov 21 14:11:19 crc kubenswrapper[4897]: I1121 14:11:19.674687 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-d8pkd" Nov 21 14:11:19 crc kubenswrapper[4897]: I1121 14:11:19.692312 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b5aa94f1-611d-4695-a3c7-1c33c80f1a1a-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"b5aa94f1-611d-4695-a3c7-1c33c80f1a1a\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 21 14:11:19 crc kubenswrapper[4897]: I1121 14:11:19.722195 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 21 14:11:19 crc kubenswrapper[4897]: I1121 14:11:19.732333 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-59rxs\" (UniqueName: \"kubernetes.io/projected/767a13d7-0e18-4496-b0bb-5a39292622e2-kube-api-access-59rxs\") pod \"community-operators-tqhs9\" (UID: \"767a13d7-0e18-4496-b0bb-5a39292622e2\") " pod="openshift-marketplace/community-operators-tqhs9" Nov 21 14:11:19 crc kubenswrapper[4897]: I1121 14:11:19.732474 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/767a13d7-0e18-4496-b0bb-5a39292622e2-catalog-content\") pod \"community-operators-tqhs9\" (UID: \"767a13d7-0e18-4496-b0bb-5a39292622e2\") " pod="openshift-marketplace/community-operators-tqhs9" Nov 21 14:11:19 crc kubenswrapper[4897]: I1121 14:11:19.732520 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:19 crc kubenswrapper[4897]: I1121 14:11:19.732559 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/767a13d7-0e18-4496-b0bb-5a39292622e2-utilities\") pod \"community-operators-tqhs9\" (UID: \"767a13d7-0e18-4496-b0bb-5a39292622e2\") " pod="openshift-marketplace/community-operators-tqhs9" Nov 21 14:11:19 crc kubenswrapper[4897]: I1121 14:11:19.734632 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-5wf4b"] Nov 21 14:11:19 crc kubenswrapper[4897]: I1121 14:11:19.734855 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/767a13d7-0e18-4496-b0bb-5a39292622e2-catalog-content\") pod \"community-operators-tqhs9\" (UID: \"767a13d7-0e18-4496-b0bb-5a39292622e2\") " pod="openshift-marketplace/community-operators-tqhs9" Nov 21 14:11:19 crc kubenswrapper[4897]: E1121 14:11:19.735298 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-21 14:11:20.235285352 +0000 UTC m=+157.519878827 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5z687" (UID: "9f63b8dd-ffb7-4963-8fcc-0c133a2c045b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:19 crc kubenswrapper[4897]: I1121 14:11:19.735623 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5wf4b" Nov 21 14:11:19 crc kubenswrapper[4897]: I1121 14:11:19.736186 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/767a13d7-0e18-4496-b0bb-5a39292622e2-utilities\") pod \"community-operators-tqhs9\" (UID: \"767a13d7-0e18-4496-b0bb-5a39292622e2\") " pod="openshift-marketplace/community-operators-tqhs9" Nov 21 14:11:19 crc kubenswrapper[4897]: I1121 14:11:19.759877 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5wf4b"] Nov 21 14:11:19 crc kubenswrapper[4897]: I1121 14:11:19.778244 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-59rxs\" (UniqueName: \"kubernetes.io/projected/767a13d7-0e18-4496-b0bb-5a39292622e2-kube-api-access-59rxs\") pod \"community-operators-tqhs9\" (UID: \"767a13d7-0e18-4496-b0bb-5a39292622e2\") " pod="openshift-marketplace/community-operators-tqhs9" Nov 21 14:11:19 crc kubenswrapper[4897]: I1121 14:11:19.844186 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 21 14:11:19 crc kubenswrapper[4897]: I1121 14:11:19.844667 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8ce22b3-b193-4bd0-9a01-9bc260c20430-catalog-content\") pod \"certified-operators-5wf4b\" (UID: \"e8ce22b3-b193-4bd0-9a01-9bc260c20430\") " pod="openshift-marketplace/certified-operators-5wf4b" Nov 21 14:11:19 crc kubenswrapper[4897]: I1121 14:11:19.844750 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-55ctt\" (UniqueName: \"kubernetes.io/projected/e8ce22b3-b193-4bd0-9a01-9bc260c20430-kube-api-access-55ctt\") pod \"certified-operators-5wf4b\" (UID: \"e8ce22b3-b193-4bd0-9a01-9bc260c20430\") " pod="openshift-marketplace/certified-operators-5wf4b" Nov 21 14:11:19 crc kubenswrapper[4897]: I1121 14:11:19.844777 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8ce22b3-b193-4bd0-9a01-9bc260c20430-utilities\") pod \"certified-operators-5wf4b\" (UID: \"e8ce22b3-b193-4bd0-9a01-9bc260c20430\") " pod="openshift-marketplace/certified-operators-5wf4b" Nov 21 14:11:19 crc kubenswrapper[4897]: E1121 14:11:19.844922 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-21 14:11:20.344907762 +0000 UTC m=+157.629501237 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:19 crc kubenswrapper[4897]: I1121 14:11:19.848213 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tqhs9" Nov 21 14:11:19 crc kubenswrapper[4897]: I1121 14:11:19.920576 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-jf4pv"] Nov 21 14:11:19 crc kubenswrapper[4897]: I1121 14:11:19.921999 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jf4pv" Nov 21 14:11:19 crc kubenswrapper[4897]: I1121 14:11:19.928542 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-t5jsz" Nov 21 14:11:19 crc kubenswrapper[4897]: I1121 14:11:19.939786 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jf4pv"] Nov 21 14:11:19 crc kubenswrapper[4897]: I1121 14:11:19.947885 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:19 crc kubenswrapper[4897]: I1121 14:11:19.947928 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-55ctt\" (UniqueName: \"kubernetes.io/projected/e8ce22b3-b193-4bd0-9a01-9bc260c20430-kube-api-access-55ctt\") pod \"certified-operators-5wf4b\" (UID: \"e8ce22b3-b193-4bd0-9a01-9bc260c20430\") " pod="openshift-marketplace/certified-operators-5wf4b" Nov 21 14:11:19 crc kubenswrapper[4897]: I1121 14:11:19.947948 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8ce22b3-b193-4bd0-9a01-9bc260c20430-utilities\") pod \"certified-operators-5wf4b\" (UID: \"e8ce22b3-b193-4bd0-9a01-9bc260c20430\") " pod="openshift-marketplace/certified-operators-5wf4b" Nov 21 14:11:19 crc kubenswrapper[4897]: I1121 14:11:19.947990 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8ce22b3-b193-4bd0-9a01-9bc260c20430-catalog-content\") pod \"certified-operators-5wf4b\" (UID: \"e8ce22b3-b193-4bd0-9a01-9bc260c20430\") " pod="openshift-marketplace/certified-operators-5wf4b" Nov 21 14:11:19 crc kubenswrapper[4897]: I1121 14:11:19.948494 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8ce22b3-b193-4bd0-9a01-9bc260c20430-catalog-content\") pod \"certified-operators-5wf4b\" (UID: \"e8ce22b3-b193-4bd0-9a01-9bc260c20430\") " pod="openshift-marketplace/certified-operators-5wf4b" Nov 21 14:11:19 crc kubenswrapper[4897]: I1121 14:11:19.948768 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8ce22b3-b193-4bd0-9a01-9bc260c20430-utilities\") pod \"certified-operators-5wf4b\" (UID: \"e8ce22b3-b193-4bd0-9a01-9bc260c20430\") " pod="openshift-marketplace/certified-operators-5wf4b" Nov 21 14:11:19 crc kubenswrapper[4897]: E1121 14:11:19.948929 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-21 14:11:20.448914933 +0000 UTC m=+157.733508408 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5z687" (UID: "9f63b8dd-ffb7-4963-8fcc-0c133a2c045b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:19 crc kubenswrapper[4897]: I1121 14:11:19.953582 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-2trmp" Nov 21 14:11:19 crc kubenswrapper[4897]: I1121 14:11:19.973762 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-rnnlh" Nov 21 14:11:19 crc kubenswrapper[4897]: I1121 14:11:19.991293 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-55ctt\" (UniqueName: \"kubernetes.io/projected/e8ce22b3-b193-4bd0-9a01-9bc260c20430-kube-api-access-55ctt\") pod \"certified-operators-5wf4b\" (UID: \"e8ce22b3-b193-4bd0-9a01-9bc260c20430\") " pod="openshift-marketplace/certified-operators-5wf4b" Nov 21 14:11:20 crc kubenswrapper[4897]: I1121 14:11:20.007902 4897 patch_prober.go:28] interesting pod/router-default-5444994796-2kgg2 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 21 14:11:20 crc kubenswrapper[4897]: [-]has-synced failed: reason withheld Nov 21 14:11:20 crc kubenswrapper[4897]: [+]process-running ok Nov 21 14:11:20 crc kubenswrapper[4897]: healthz check failed Nov 21 14:11:20 crc kubenswrapper[4897]: I1121 14:11:20.007962 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-2kgg2" podUID="e095a18e-3a0b-4418-95a8-a2cc3de5bd7b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 21 14:11:20 crc kubenswrapper[4897]: I1121 14:11:20.056005 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 21 14:11:20 crc kubenswrapper[4897]: I1121 14:11:20.056439 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dt6q8\" (UniqueName: \"kubernetes.io/projected/5ea4ecd1-6706-4a89-8034-4f41ee7e0f2d-kube-api-access-dt6q8\") pod \"community-operators-jf4pv\" (UID: \"5ea4ecd1-6706-4a89-8034-4f41ee7e0f2d\") " pod="openshift-marketplace/community-operators-jf4pv" Nov 21 14:11:20 crc kubenswrapper[4897]: I1121 14:11:20.056517 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ea4ecd1-6706-4a89-8034-4f41ee7e0f2d-utilities\") pod \"community-operators-jf4pv\" (UID: \"5ea4ecd1-6706-4a89-8034-4f41ee7e0f2d\") " pod="openshift-marketplace/community-operators-jf4pv" Nov 21 14:11:20 crc kubenswrapper[4897]: I1121 14:11:20.056566 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ea4ecd1-6706-4a89-8034-4f41ee7e0f2d-catalog-content\") pod \"community-operators-jf4pv\" (UID: \"5ea4ecd1-6706-4a89-8034-4f41ee7e0f2d\") " pod="openshift-marketplace/community-operators-jf4pv" Nov 21 14:11:20 crc kubenswrapper[4897]: I1121 14:11:20.056769 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5wf4b" Nov 21 14:11:20 crc kubenswrapper[4897]: E1121 14:11:20.057314 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-21 14:11:20.557297567 +0000 UTC m=+157.841891042 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:20 crc kubenswrapper[4897]: I1121 14:11:20.182800 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:20 crc kubenswrapper[4897]: I1121 14:11:20.183176 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dt6q8\" (UniqueName: \"kubernetes.io/projected/5ea4ecd1-6706-4a89-8034-4f41ee7e0f2d-kube-api-access-dt6q8\") pod \"community-operators-jf4pv\" (UID: \"5ea4ecd1-6706-4a89-8034-4f41ee7e0f2d\") " pod="openshift-marketplace/community-operators-jf4pv" Nov 21 14:11:20 crc kubenswrapper[4897]: I1121 14:11:20.183219 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ea4ecd1-6706-4a89-8034-4f41ee7e0f2d-utilities\") pod \"community-operators-jf4pv\" (UID: \"5ea4ecd1-6706-4a89-8034-4f41ee7e0f2d\") " pod="openshift-marketplace/community-operators-jf4pv" Nov 21 14:11:20 crc kubenswrapper[4897]: I1121 14:11:20.183245 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ea4ecd1-6706-4a89-8034-4f41ee7e0f2d-catalog-content\") pod \"community-operators-jf4pv\" (UID: \"5ea4ecd1-6706-4a89-8034-4f41ee7e0f2d\") " pod="openshift-marketplace/community-operators-jf4pv" Nov 21 14:11:20 crc kubenswrapper[4897]: I1121 14:11:20.183674 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ea4ecd1-6706-4a89-8034-4f41ee7e0f2d-catalog-content\") pod \"community-operators-jf4pv\" (UID: \"5ea4ecd1-6706-4a89-8034-4f41ee7e0f2d\") " pod="openshift-marketplace/community-operators-jf4pv" Nov 21 14:11:20 crc kubenswrapper[4897]: E1121 14:11:20.183940 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-21 14:11:20.683924988 +0000 UTC m=+157.968518463 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5z687" (UID: "9f63b8dd-ffb7-4963-8fcc-0c133a2c045b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:20 crc kubenswrapper[4897]: I1121 14:11:20.184373 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ea4ecd1-6706-4a89-8034-4f41ee7e0f2d-utilities\") pod \"community-operators-jf4pv\" (UID: \"5ea4ecd1-6706-4a89-8034-4f41ee7e0f2d\") " pod="openshift-marketplace/community-operators-jf4pv" Nov 21 14:11:20 crc kubenswrapper[4897]: I1121 14:11:20.211621 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dt6q8\" (UniqueName: \"kubernetes.io/projected/5ea4ecd1-6706-4a89-8034-4f41ee7e0f2d-kube-api-access-dt6q8\") pod \"community-operators-jf4pv\" (UID: \"5ea4ecd1-6706-4a89-8034-4f41ee7e0f2d\") " pod="openshift-marketplace/community-operators-jf4pv" Nov 21 14:11:20 crc kubenswrapper[4897]: I1121 14:11:20.265493 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jf4pv" Nov 21 14:11:20 crc kubenswrapper[4897]: I1121 14:11:20.285234 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 21 14:11:20 crc kubenswrapper[4897]: E1121 14:11:20.285553 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-21 14:11:20.78553435 +0000 UTC m=+158.070127825 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:20 crc kubenswrapper[4897]: I1121 14:11:20.386901 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:20 crc kubenswrapper[4897]: E1121 14:11:20.387323 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-21 14:11:20.887309148 +0000 UTC m=+158.171902633 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5z687" (UID: "9f63b8dd-ffb7-4963-8fcc-0c133a2c045b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:20 crc kubenswrapper[4897]: I1121 14:11:20.488499 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 21 14:11:20 crc kubenswrapper[4897]: E1121 14:11:20.488941 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-21 14:11:20.988926251 +0000 UTC m=+158.273519726 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:20 crc kubenswrapper[4897]: I1121 14:11:20.513487 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"7cd98bfc-77a5-4ef5-86f1-7614b0b2ad2a","Type":"ContainerStarted","Data":"6c758b6efbb5f82c9333dbdc723b09bf1071cf39610e73d8df36a08d014e5542"} Nov 21 14:11:20 crc kubenswrapper[4897]: I1121 14:11:20.590623 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:20 crc kubenswrapper[4897]: E1121 14:11:20.592390 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-21 14:11:21.092377796 +0000 UTC m=+158.376971271 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5z687" (UID: "9f63b8dd-ffb7-4963-8fcc-0c133a2c045b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:20 crc kubenswrapper[4897]: I1121 14:11:20.602687 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=2.602663627 podStartE2EDuration="2.602663627s" podCreationTimestamp="2025-11-21 14:11:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:11:20.599831217 +0000 UTC m=+157.884424702" watchObservedRunningTime="2025-11-21 14:11:20.602663627 +0000 UTC m=+157.887257102" Nov 21 14:11:20 crc kubenswrapper[4897]: I1121 14:11:20.692219 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 21 14:11:20 crc kubenswrapper[4897]: E1121 14:11:20.692796 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-21 14:11:21.192777455 +0000 UTC m=+158.477370930 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:20 crc kubenswrapper[4897]: I1121 14:11:20.782261 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-q6dd2"] Nov 21 14:11:20 crc kubenswrapper[4897]: I1121 14:11:20.803548 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:20 crc kubenswrapper[4897]: E1121 14:11:20.804217 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-21 14:11:21.304202746 +0000 UTC m=+158.588796221 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5z687" (UID: "9f63b8dd-ffb7-4963-8fcc-0c133a2c045b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:20 crc kubenswrapper[4897]: W1121 14:11:20.871968 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1dc1b877_a8de_45b0_bab4_4411ecced04d.slice/crio-0e682cfbb58918fb97e000f0b36d176994a297cf3e898b70d21a249bcb5a1c26 WatchSource:0}: Error finding container 0e682cfbb58918fb97e000f0b36d176994a297cf3e898b70d21a249bcb5a1c26: Status 404 returned error can't find the container with id 0e682cfbb58918fb97e000f0b36d176994a297cf3e898b70d21a249bcb5a1c26 Nov 21 14:11:20 crc kubenswrapper[4897]: I1121 14:11:20.891066 4897 patch_prober.go:28] interesting pod/downloads-7954f5f757-vgwv2 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.26:8080/\": dial tcp 10.217.0.26:8080: connect: connection refused" start-of-body= Nov 21 14:11:20 crc kubenswrapper[4897]: I1121 14:11:20.891130 4897 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-vgwv2" podUID="82f81ac2-d2b2-49de-82f3-a51796dbff03" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.26:8080/\": dial tcp 10.217.0.26:8080: connect: connection refused" Nov 21 14:11:20 crc kubenswrapper[4897]: I1121 14:11:20.891443 4897 patch_prober.go:28] interesting pod/downloads-7954f5f757-vgwv2 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.26:8080/\": dial tcp 10.217.0.26:8080: connect: connection refused" start-of-body= Nov 21 14:11:20 crc kubenswrapper[4897]: I1121 14:11:20.891496 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-vgwv2" podUID="82f81ac2-d2b2-49de-82f3-a51796dbff03" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.26:8080/\": dial tcp 10.217.0.26:8080: connect: connection refused" Nov 21 14:11:20 crc kubenswrapper[4897]: I1121 14:11:20.906066 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-7fsr9" Nov 21 14:11:20 crc kubenswrapper[4897]: I1121 14:11:20.906119 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-7fsr9" Nov 21 14:11:20 crc kubenswrapper[4897]: I1121 14:11:20.906447 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 21 14:11:20 crc kubenswrapper[4897]: E1121 14:11:20.906723 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-21 14:11:21.406702794 +0000 UTC m=+158.691296269 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:20 crc kubenswrapper[4897]: I1121 14:11:20.907098 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:20 crc kubenswrapper[4897]: E1121 14:11:20.907394 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-21 14:11:21.407385233 +0000 UTC m=+158.691978708 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5z687" (UID: "9f63b8dd-ffb7-4963-8fcc-0c133a2c045b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:20 crc kubenswrapper[4897]: I1121 14:11:20.910104 4897 patch_prober.go:28] interesting pod/console-f9d7485db-7fsr9 container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.18:8443/health\": dial tcp 10.217.0.18:8443: connect: connection refused" start-of-body= Nov 21 14:11:20 crc kubenswrapper[4897]: I1121 14:11:20.910151 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-7fsr9" podUID="8342c8a2-87f4-4bf9-a2ca-073c02b277ad" containerName="console" probeResult="failure" output="Get \"https://10.217.0.18:8443/health\": dial tcp 10.217.0.18:8443: connect: connection refused" Nov 21 14:11:20 crc kubenswrapper[4897]: I1121 14:11:20.995731 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-2kgg2" Nov 21 14:11:21 crc kubenswrapper[4897]: I1121 14:11:20.999176 4897 patch_prober.go:28] interesting pod/router-default-5444994796-2kgg2 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 21 14:11:21 crc kubenswrapper[4897]: [-]has-synced failed: reason withheld Nov 21 14:11:21 crc kubenswrapper[4897]: [+]process-running ok Nov 21 14:11:21 crc kubenswrapper[4897]: healthz check failed Nov 21 14:11:21 crc kubenswrapper[4897]: I1121 14:11:20.999211 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-2kgg2" podUID="e095a18e-3a0b-4418-95a8-a2cc3de5bd7b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 21 14:11:21 crc kubenswrapper[4897]: I1121 14:11:21.010309 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 21 14:11:21 crc kubenswrapper[4897]: E1121 14:11:21.011553 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-21 14:11:21.511531218 +0000 UTC m=+158.796124693 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:21 crc kubenswrapper[4897]: I1121 14:11:21.017952 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5wf4b"] Nov 21 14:11:21 crc kubenswrapper[4897]: I1121 14:11:21.078798 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jf4pv"] Nov 21 14:11:21 crc kubenswrapper[4897]: I1121 14:11:21.108212 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 21 14:11:21 crc kubenswrapper[4897]: I1121 14:11:21.112490 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:21 crc kubenswrapper[4897]: E1121 14:11:21.117036 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-21 14:11:21.61702031 +0000 UTC m=+158.901613785 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5z687" (UID: "9f63b8dd-ffb7-4963-8fcc-0c133a2c045b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:21 crc kubenswrapper[4897]: I1121 14:11:21.208790 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tqhs9"] Nov 21 14:11:21 crc kubenswrapper[4897]: I1121 14:11:21.213872 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 21 14:11:21 crc kubenswrapper[4897]: E1121 14:11:21.214471 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-21 14:11:21.714430435 +0000 UTC m=+158.999023910 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:21 crc kubenswrapper[4897]: W1121 14:11:21.233319 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod767a13d7_0e18_4496_b0bb_5a39292622e2.slice/crio-0bc586219d861cea2930e3953fe8cc3621f9fca40536d64da3db17d429927282 WatchSource:0}: Error finding container 0bc586219d861cea2930e3953fe8cc3621f9fca40536d64da3db17d429927282: Status 404 returned error can't find the container with id 0bc586219d861cea2930e3953fe8cc3621f9fca40536d64da3db17d429927282 Nov 21 14:11:21 crc kubenswrapper[4897]: I1121 14:11:21.278212 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4cx2q" Nov 21 14:11:21 crc kubenswrapper[4897]: I1121 14:11:21.302381 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h59mj" Nov 21 14:11:21 crc kubenswrapper[4897]: I1121 14:11:21.316105 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:21 crc kubenswrapper[4897]: E1121 14:11:21.316444 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-21 14:11:21.816433378 +0000 UTC m=+159.101026843 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5z687" (UID: "9f63b8dd-ffb7-4963-8fcc-0c133a2c045b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:21 crc kubenswrapper[4897]: I1121 14:11:21.358370 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6sfv4" Nov 21 14:11:21 crc kubenswrapper[4897]: I1121 14:11:21.359945 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-z9hjh" Nov 21 14:11:21 crc kubenswrapper[4897]: I1121 14:11:21.369665 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-s7tww" Nov 21 14:11:21 crc kubenswrapper[4897]: I1121 14:11:21.369705 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-s7tww" Nov 21 14:11:21 crc kubenswrapper[4897]: I1121 14:11:21.418357 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 21 14:11:21 crc kubenswrapper[4897]: E1121 14:11:21.418541 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-21 14:11:21.918512204 +0000 UTC m=+159.203105679 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:21 crc kubenswrapper[4897]: I1121 14:11:21.418827 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:21 crc kubenswrapper[4897]: E1121 14:11:21.419562 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-21 14:11:21.919549904 +0000 UTC m=+159.204143379 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5z687" (UID: "9f63b8dd-ffb7-4963-8fcc-0c133a2c045b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:21 crc kubenswrapper[4897]: I1121 14:11:21.482408 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-6fqg7"] Nov 21 14:11:21 crc kubenswrapper[4897]: I1121 14:11:21.483456 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6fqg7" Nov 21 14:11:21 crc kubenswrapper[4897]: I1121 14:11:21.485407 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 21 14:11:21 crc kubenswrapper[4897]: I1121 14:11:21.497367 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6fqg7"] Nov 21 14:11:21 crc kubenswrapper[4897]: I1121 14:11:21.519597 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 21 14:11:21 crc kubenswrapper[4897]: E1121 14:11:21.519700 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-21 14:11:22.019676136 +0000 UTC m=+159.304269611 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:21 crc kubenswrapper[4897]: I1121 14:11:21.519893 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:21 crc kubenswrapper[4897]: I1121 14:11:21.521049 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-bjdjf" event={"ID":"78db7e2e-ba45-464e-9b1d-0fc61191955a","Type":"ContainerStarted","Data":"71c4702c0676632917f26d297f4985e42941e6f76c2d6061f2b05f9827edba63"} Nov 21 14:11:21 crc kubenswrapper[4897]: I1121 14:11:21.521095 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-bjdjf" event={"ID":"78db7e2e-ba45-464e-9b1d-0fc61191955a","Type":"ContainerStarted","Data":"4076a7b4695cfd59303187e17d6f17f7a7f1dc9db9b6532b6d52fe4f6ce6faea"} Nov 21 14:11:21 crc kubenswrapper[4897]: E1121 14:11:21.521296 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-21 14:11:22.021283891 +0000 UTC m=+159.305877366 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5z687" (UID: "9f63b8dd-ffb7-4963-8fcc-0c133a2c045b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:21 crc kubenswrapper[4897]: I1121 14:11:21.522602 4897 generic.go:334] "Generic (PLEG): container finished" podID="1dc1b877-a8de-45b0-bab4-4411ecced04d" containerID="b9c27b68540fcadb0a431d5f21d557435c273b2b80b22ddfe779d421a63947ca" exitCode=0 Nov 21 14:11:21 crc kubenswrapper[4897]: I1121 14:11:21.522657 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q6dd2" event={"ID":"1dc1b877-a8de-45b0-bab4-4411ecced04d","Type":"ContainerDied","Data":"b9c27b68540fcadb0a431d5f21d557435c273b2b80b22ddfe779d421a63947ca"} Nov 21 14:11:21 crc kubenswrapper[4897]: I1121 14:11:21.522683 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q6dd2" event={"ID":"1dc1b877-a8de-45b0-bab4-4411ecced04d","Type":"ContainerStarted","Data":"0e682cfbb58918fb97e000f0b36d176994a297cf3e898b70d21a249bcb5a1c26"} Nov 21 14:11:21 crc kubenswrapper[4897]: I1121 14:11:21.524895 4897 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 21 14:11:21 crc kubenswrapper[4897]: I1121 14:11:21.526487 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"b5aa94f1-611d-4695-a3c7-1c33c80f1a1a","Type":"ContainerStarted","Data":"2d5180bb09d2b710821543745c7670fc20f3417633851aca10af552d8bed0042"} Nov 21 14:11:21 crc kubenswrapper[4897]: I1121 14:11:21.530220 4897 generic.go:334] "Generic (PLEG): container finished" podID="7cd98bfc-77a5-4ef5-86f1-7614b0b2ad2a" containerID="6c758b6efbb5f82c9333dbdc723b09bf1071cf39610e73d8df36a08d014e5542" exitCode=0 Nov 21 14:11:21 crc kubenswrapper[4897]: I1121 14:11:21.530301 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"7cd98bfc-77a5-4ef5-86f1-7614b0b2ad2a","Type":"ContainerDied","Data":"6c758b6efbb5f82c9333dbdc723b09bf1071cf39610e73d8df36a08d014e5542"} Nov 21 14:11:21 crc kubenswrapper[4897]: I1121 14:11:21.540667 4897 generic.go:334] "Generic (PLEG): container finished" podID="c14a1c93-8be4-42b7-aae1-9c31d53a6bfc" containerID="7c71e97295b9fe2b2d30a0502ffb2637c6dd199e5c02c6709abf1958f64afa5a" exitCode=0 Nov 21 14:11:21 crc kubenswrapper[4897]: I1121 14:11:21.540794 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29395560-pq7jf" event={"ID":"c14a1c93-8be4-42b7-aae1-9c31d53a6bfc","Type":"ContainerDied","Data":"7c71e97295b9fe2b2d30a0502ffb2637c6dd199e5c02c6709abf1958f64afa5a"} Nov 21 14:11:21 crc kubenswrapper[4897]: I1121 14:11:21.547925 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tqhs9" event={"ID":"767a13d7-0e18-4496-b0bb-5a39292622e2","Type":"ContainerStarted","Data":"0bc586219d861cea2930e3953fe8cc3621f9fca40536d64da3db17d429927282"} Nov 21 14:11:21 crc kubenswrapper[4897]: I1121 14:11:21.553628 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5wf4b" event={"ID":"e8ce22b3-b193-4bd0-9a01-9bc260c20430","Type":"ContainerStarted","Data":"4b959a8e218803f046a5afb07863a1662545b93118ba4d8702a7e6b3aaa2551a"} Nov 21 14:11:21 crc kubenswrapper[4897]: I1121 14:11:21.553674 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5wf4b" event={"ID":"e8ce22b3-b193-4bd0-9a01-9bc260c20430","Type":"ContainerStarted","Data":"6b156833ac602e98d8284ba6e7e11e9fdae647375b750fec64ad922e98f30bde"} Nov 21 14:11:21 crc kubenswrapper[4897]: I1121 14:11:21.564060 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jf4pv" event={"ID":"5ea4ecd1-6706-4a89-8034-4f41ee7e0f2d","Type":"ContainerStarted","Data":"8cbae1e2231d99a0e5a5c8cba032c583df8388b21e53a8a393e3e6089cac6100"} Nov 21 14:11:21 crc kubenswrapper[4897]: I1121 14:11:21.597977 4897 patch_prober.go:28] interesting pod/apiserver-76f77b778f-s7tww container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Nov 21 14:11:21 crc kubenswrapper[4897]: [+]log ok Nov 21 14:11:21 crc kubenswrapper[4897]: [+]etcd ok Nov 21 14:11:21 crc kubenswrapper[4897]: [+]poststarthook/start-apiserver-admission-initializer ok Nov 21 14:11:21 crc kubenswrapper[4897]: [+]poststarthook/generic-apiserver-start-informers ok Nov 21 14:11:21 crc kubenswrapper[4897]: [+]poststarthook/max-in-flight-filter ok Nov 21 14:11:21 crc kubenswrapper[4897]: [+]poststarthook/storage-object-count-tracker-hook ok Nov 21 14:11:21 crc kubenswrapper[4897]: [+]poststarthook/image.openshift.io-apiserver-caches ok Nov 21 14:11:21 crc kubenswrapper[4897]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Nov 21 14:11:21 crc kubenswrapper[4897]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Nov 21 14:11:21 crc kubenswrapper[4897]: [+]poststarthook/project.openshift.io-projectcache ok Nov 21 14:11:21 crc kubenswrapper[4897]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Nov 21 14:11:21 crc kubenswrapper[4897]: [-]poststarthook/openshift.io-startinformers failed: reason withheld Nov 21 14:11:21 crc kubenswrapper[4897]: [+]poststarthook/openshift.io-restmapperupdater ok Nov 21 14:11:21 crc kubenswrapper[4897]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Nov 21 14:11:21 crc kubenswrapper[4897]: livez check failed Nov 21 14:11:21 crc kubenswrapper[4897]: I1121 14:11:21.598054 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-s7tww" podUID="027b031c-1b21-40e8-aa7c-4a342aae265a" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 21 14:11:21 crc kubenswrapper[4897]: I1121 14:11:21.627143 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 21 14:11:21 crc kubenswrapper[4897]: I1121 14:11:21.627407 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4x5qq\" (UniqueName: \"kubernetes.io/projected/8dde9f26-f67d-49d2-98ac-7483796f1212-kube-api-access-4x5qq\") pod \"redhat-marketplace-6fqg7\" (UID: \"8dde9f26-f67d-49d2-98ac-7483796f1212\") " pod="openshift-marketplace/redhat-marketplace-6fqg7" Nov 21 14:11:21 crc kubenswrapper[4897]: I1121 14:11:21.627450 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8dde9f26-f67d-49d2-98ac-7483796f1212-catalog-content\") pod \"redhat-marketplace-6fqg7\" (UID: \"8dde9f26-f67d-49d2-98ac-7483796f1212\") " pod="openshift-marketplace/redhat-marketplace-6fqg7" Nov 21 14:11:21 crc kubenswrapper[4897]: I1121 14:11:21.627517 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8dde9f26-f67d-49d2-98ac-7483796f1212-utilities\") pod \"redhat-marketplace-6fqg7\" (UID: \"8dde9f26-f67d-49d2-98ac-7483796f1212\") " pod="openshift-marketplace/redhat-marketplace-6fqg7" Nov 21 14:11:21 crc kubenswrapper[4897]: E1121 14:11:21.630364 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-21 14:11:22.130340094 +0000 UTC m=+159.414933579 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:21 crc kubenswrapper[4897]: I1121 14:11:21.728217 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4x5qq\" (UniqueName: \"kubernetes.io/projected/8dde9f26-f67d-49d2-98ac-7483796f1212-kube-api-access-4x5qq\") pod \"redhat-marketplace-6fqg7\" (UID: \"8dde9f26-f67d-49d2-98ac-7483796f1212\") " pod="openshift-marketplace/redhat-marketplace-6fqg7" Nov 21 14:11:21 crc kubenswrapper[4897]: I1121 14:11:21.728278 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8dde9f26-f67d-49d2-98ac-7483796f1212-catalog-content\") pod \"redhat-marketplace-6fqg7\" (UID: \"8dde9f26-f67d-49d2-98ac-7483796f1212\") " pod="openshift-marketplace/redhat-marketplace-6fqg7" Nov 21 14:11:21 crc kubenswrapper[4897]: I1121 14:11:21.728319 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8dde9f26-f67d-49d2-98ac-7483796f1212-utilities\") pod \"redhat-marketplace-6fqg7\" (UID: \"8dde9f26-f67d-49d2-98ac-7483796f1212\") " pod="openshift-marketplace/redhat-marketplace-6fqg7" Nov 21 14:11:21 crc kubenswrapper[4897]: I1121 14:11:21.728343 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:21 crc kubenswrapper[4897]: I1121 14:11:21.729213 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8dde9f26-f67d-49d2-98ac-7483796f1212-catalog-content\") pod \"redhat-marketplace-6fqg7\" (UID: \"8dde9f26-f67d-49d2-98ac-7483796f1212\") " pod="openshift-marketplace/redhat-marketplace-6fqg7" Nov 21 14:11:21 crc kubenswrapper[4897]: I1121 14:11:21.729447 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8dde9f26-f67d-49d2-98ac-7483796f1212-utilities\") pod \"redhat-marketplace-6fqg7\" (UID: \"8dde9f26-f67d-49d2-98ac-7483796f1212\") " pod="openshift-marketplace/redhat-marketplace-6fqg7" Nov 21 14:11:21 crc kubenswrapper[4897]: E1121 14:11:21.729715 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-21 14:11:22.229703194 +0000 UTC m=+159.514296669 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5z687" (UID: "9f63b8dd-ffb7-4963-8fcc-0c133a2c045b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:21 crc kubenswrapper[4897]: I1121 14:11:21.750684 4897 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Nov 21 14:11:21 crc kubenswrapper[4897]: I1121 14:11:21.752557 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4x5qq\" (UniqueName: \"kubernetes.io/projected/8dde9f26-f67d-49d2-98ac-7483796f1212-kube-api-access-4x5qq\") pod \"redhat-marketplace-6fqg7\" (UID: \"8dde9f26-f67d-49d2-98ac-7483796f1212\") " pod="openshift-marketplace/redhat-marketplace-6fqg7" Nov 21 14:11:21 crc kubenswrapper[4897]: I1121 14:11:21.798238 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6fqg7" Nov 21 14:11:21 crc kubenswrapper[4897]: I1121 14:11:21.829852 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 21 14:11:21 crc kubenswrapper[4897]: E1121 14:11:21.830103 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-21 14:11:22.330067721 +0000 UTC m=+159.614661196 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:21 crc kubenswrapper[4897]: I1121 14:11:21.830911 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:21 crc kubenswrapper[4897]: E1121 14:11:21.831244 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-21 14:11:22.331231914 +0000 UTC m=+159.615825389 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5z687" (UID: "9f63b8dd-ffb7-4963-8fcc-0c133a2c045b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:21 crc kubenswrapper[4897]: I1121 14:11:21.885247 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-k2ctv"] Nov 21 14:11:21 crc kubenswrapper[4897]: I1121 14:11:21.886327 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k2ctv" Nov 21 14:11:21 crc kubenswrapper[4897]: I1121 14:11:21.900076 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-k2ctv"] Nov 21 14:11:21 crc kubenswrapper[4897]: I1121 14:11:21.932065 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 21 14:11:21 crc kubenswrapper[4897]: E1121 14:11:21.932466 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-21 14:11:22.432435965 +0000 UTC m=+159.717029440 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:22 crc kubenswrapper[4897]: I1121 14:11:22.000155 4897 patch_prober.go:28] interesting pod/router-default-5444994796-2kgg2 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 21 14:11:22 crc kubenswrapper[4897]: [-]has-synced failed: reason withheld Nov 21 14:11:22 crc kubenswrapper[4897]: [+]process-running ok Nov 21 14:11:22 crc kubenswrapper[4897]: healthz check failed Nov 21 14:11:22 crc kubenswrapper[4897]: I1121 14:11:22.000213 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-2kgg2" podUID="e095a18e-3a0b-4418-95a8-a2cc3de5bd7b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 21 14:11:22 crc kubenswrapper[4897]: I1121 14:11:22.034072 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a55d260f-3da6-4b30-b633-6348b31b0b01-catalog-content\") pod \"redhat-marketplace-k2ctv\" (UID: \"a55d260f-3da6-4b30-b633-6348b31b0b01\") " pod="openshift-marketplace/redhat-marketplace-k2ctv" Nov 21 14:11:22 crc kubenswrapper[4897]: I1121 14:11:22.034179 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:22 crc kubenswrapper[4897]: I1121 14:11:22.034219 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a55d260f-3da6-4b30-b633-6348b31b0b01-utilities\") pod \"redhat-marketplace-k2ctv\" (UID: \"a55d260f-3da6-4b30-b633-6348b31b0b01\") " pod="openshift-marketplace/redhat-marketplace-k2ctv" Nov 21 14:11:22 crc kubenswrapper[4897]: I1121 14:11:22.034256 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jc9hp\" (UniqueName: \"kubernetes.io/projected/a55d260f-3da6-4b30-b633-6348b31b0b01-kube-api-access-jc9hp\") pod \"redhat-marketplace-k2ctv\" (UID: \"a55d260f-3da6-4b30-b633-6348b31b0b01\") " pod="openshift-marketplace/redhat-marketplace-k2ctv" Nov 21 14:11:22 crc kubenswrapper[4897]: I1121 14:11:22.034295 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6fqg7"] Nov 21 14:11:22 crc kubenswrapper[4897]: E1121 14:11:22.034685 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-21 14:11:22.534667306 +0000 UTC m=+159.819260781 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5z687" (UID: "9f63b8dd-ffb7-4963-8fcc-0c133a2c045b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:22 crc kubenswrapper[4897]: I1121 14:11:22.134868 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 21 14:11:22 crc kubenswrapper[4897]: E1121 14:11:22.135040 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-21 14:11:22.635011593 +0000 UTC m=+159.919605078 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:22 crc kubenswrapper[4897]: I1121 14:11:22.135209 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:22 crc kubenswrapper[4897]: I1121 14:11:22.135261 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a55d260f-3da6-4b30-b633-6348b31b0b01-utilities\") pod \"redhat-marketplace-k2ctv\" (UID: \"a55d260f-3da6-4b30-b633-6348b31b0b01\") " pod="openshift-marketplace/redhat-marketplace-k2ctv" Nov 21 14:11:22 crc kubenswrapper[4897]: I1121 14:11:22.135309 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jc9hp\" (UniqueName: \"kubernetes.io/projected/a55d260f-3da6-4b30-b633-6348b31b0b01-kube-api-access-jc9hp\") pod \"redhat-marketplace-k2ctv\" (UID: \"a55d260f-3da6-4b30-b633-6348b31b0b01\") " pod="openshift-marketplace/redhat-marketplace-k2ctv" Nov 21 14:11:22 crc kubenswrapper[4897]: I1121 14:11:22.135352 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a55d260f-3da6-4b30-b633-6348b31b0b01-catalog-content\") pod \"redhat-marketplace-k2ctv\" (UID: \"a55d260f-3da6-4b30-b633-6348b31b0b01\") " pod="openshift-marketplace/redhat-marketplace-k2ctv" Nov 21 14:11:22 crc kubenswrapper[4897]: E1121 14:11:22.135620 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-21 14:11:22.635609671 +0000 UTC m=+159.920203216 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5z687" (UID: "9f63b8dd-ffb7-4963-8fcc-0c133a2c045b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:22 crc kubenswrapper[4897]: I1121 14:11:22.135861 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a55d260f-3da6-4b30-b633-6348b31b0b01-catalog-content\") pod \"redhat-marketplace-k2ctv\" (UID: \"a55d260f-3da6-4b30-b633-6348b31b0b01\") " pod="openshift-marketplace/redhat-marketplace-k2ctv" Nov 21 14:11:22 crc kubenswrapper[4897]: I1121 14:11:22.136108 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a55d260f-3da6-4b30-b633-6348b31b0b01-utilities\") pod \"redhat-marketplace-k2ctv\" (UID: \"a55d260f-3da6-4b30-b633-6348b31b0b01\") " pod="openshift-marketplace/redhat-marketplace-k2ctv" Nov 21 14:11:22 crc kubenswrapper[4897]: I1121 14:11:22.159355 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jc9hp\" (UniqueName: \"kubernetes.io/projected/a55d260f-3da6-4b30-b633-6348b31b0b01-kube-api-access-jc9hp\") pod \"redhat-marketplace-k2ctv\" (UID: \"a55d260f-3da6-4b30-b633-6348b31b0b01\") " pod="openshift-marketplace/redhat-marketplace-k2ctv" Nov 21 14:11:22 crc kubenswrapper[4897]: I1121 14:11:22.221057 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k2ctv" Nov 21 14:11:22 crc kubenswrapper[4897]: I1121 14:11:22.237370 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 21 14:11:22 crc kubenswrapper[4897]: E1121 14:11:22.237482 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-21 14:11:22.73746386 +0000 UTC m=+160.022057335 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:22 crc kubenswrapper[4897]: I1121 14:11:22.237799 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:22 crc kubenswrapper[4897]: E1121 14:11:22.238076 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-21 14:11:22.738065087 +0000 UTC m=+160.022658562 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5z687" (UID: "9f63b8dd-ffb7-4963-8fcc-0c133a2c045b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:22 crc kubenswrapper[4897]: I1121 14:11:22.342127 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 21 14:11:22 crc kubenswrapper[4897]: E1121 14:11:22.342337 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-21 14:11:22.842305234 +0000 UTC m=+160.126898709 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:22 crc kubenswrapper[4897]: I1121 14:11:22.342675 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:22 crc kubenswrapper[4897]: E1121 14:11:22.342996 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-21 14:11:22.842982344 +0000 UTC m=+160.127575809 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-5z687" (UID: "9f63b8dd-ffb7-4963-8fcc-0c133a2c045b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 21 14:11:22 crc kubenswrapper[4897]: I1121 14:11:22.358066 4897 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-11-21T14:11:21.750715328Z","Handler":null,"Name":""} Nov 21 14:11:22 crc kubenswrapper[4897]: I1121 14:11:22.382673 4897 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Nov 21 14:11:22 crc kubenswrapper[4897]: I1121 14:11:22.382744 4897 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Nov 21 14:11:22 crc kubenswrapper[4897]: I1121 14:11:22.444106 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 21 14:11:22 crc kubenswrapper[4897]: I1121 14:11:22.449772 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 21 14:11:22 crc kubenswrapper[4897]: I1121 14:11:22.451465 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-k2ctv"] Nov 21 14:11:22 crc kubenswrapper[4897]: I1121 14:11:22.477842 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-sqdvs"] Nov 21 14:11:22 crc kubenswrapper[4897]: I1121 14:11:22.478948 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sqdvs" Nov 21 14:11:22 crc kubenswrapper[4897]: I1121 14:11:22.484352 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 21 14:11:22 crc kubenswrapper[4897]: I1121 14:11:22.485965 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-sqdvs"] Nov 21 14:11:22 crc kubenswrapper[4897]: I1121 14:11:22.545283 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:22 crc kubenswrapper[4897]: I1121 14:11:22.548752 4897 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 21 14:11:22 crc kubenswrapper[4897]: I1121 14:11:22.548783 4897 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:22 crc kubenswrapper[4897]: I1121 14:11:22.572284 4897 generic.go:334] "Generic (PLEG): container finished" podID="5ea4ecd1-6706-4a89-8034-4f41ee7e0f2d" containerID="874794186ab1f82bbcce7b073013f11d92ada6b7264b7da2db77f59a32599d11" exitCode=0 Nov 21 14:11:22 crc kubenswrapper[4897]: I1121 14:11:22.572360 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jf4pv" event={"ID":"5ea4ecd1-6706-4a89-8034-4f41ee7e0f2d","Type":"ContainerDied","Data":"874794186ab1f82bbcce7b073013f11d92ada6b7264b7da2db77f59a32599d11"} Nov 21 14:11:22 crc kubenswrapper[4897]: I1121 14:11:22.576366 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-bjdjf" event={"ID":"78db7e2e-ba45-464e-9b1d-0fc61191955a","Type":"ContainerStarted","Data":"520dd3fbd6a907dcca55f1f4a920869b240eab863e09de9aa763cb69865944e5"} Nov 21 14:11:22 crc kubenswrapper[4897]: I1121 14:11:22.577772 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k2ctv" event={"ID":"a55d260f-3da6-4b30-b633-6348b31b0b01","Type":"ContainerStarted","Data":"a71da170e0a9164c1f62738449a871665e90540e7aa100ae5b2be97288ff7462"} Nov 21 14:11:22 crc kubenswrapper[4897]: I1121 14:11:22.580066 4897 generic.go:334] "Generic (PLEG): container finished" podID="b5aa94f1-611d-4695-a3c7-1c33c80f1a1a" containerID="ca350061118a740fd09291a61bf0414e2e64fbd9d1235805536fa72ed4de5a16" exitCode=0 Nov 21 14:11:22 crc kubenswrapper[4897]: I1121 14:11:22.580135 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"b5aa94f1-611d-4695-a3c7-1c33c80f1a1a","Type":"ContainerDied","Data":"ca350061118a740fd09291a61bf0414e2e64fbd9d1235805536fa72ed4de5a16"} Nov 21 14:11:22 crc kubenswrapper[4897]: I1121 14:11:22.581858 4897 generic.go:334] "Generic (PLEG): container finished" podID="8dde9f26-f67d-49d2-98ac-7483796f1212" containerID="0f2fb4771e7be9d2354d41b4fee2027b5fce078a7d8ebb7f5824f08c597bbf83" exitCode=0 Nov 21 14:11:22 crc kubenswrapper[4897]: I1121 14:11:22.581892 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6fqg7" event={"ID":"8dde9f26-f67d-49d2-98ac-7483796f1212","Type":"ContainerDied","Data":"0f2fb4771e7be9d2354d41b4fee2027b5fce078a7d8ebb7f5824f08c597bbf83"} Nov 21 14:11:22 crc kubenswrapper[4897]: I1121 14:11:22.581917 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6fqg7" event={"ID":"8dde9f26-f67d-49d2-98ac-7483796f1212","Type":"ContainerStarted","Data":"6f0c3aae606f7cac3c0a134f53733b7f6c1d62feb2cbedc2915445c42a62df52"} Nov 21 14:11:22 crc kubenswrapper[4897]: I1121 14:11:22.594527 4897 generic.go:334] "Generic (PLEG): container finished" podID="767a13d7-0e18-4496-b0bb-5a39292622e2" containerID="6be88453f6e893cdb6bf58762103e7d29f4937b2476a9dde9549dd05fc04fad8" exitCode=0 Nov 21 14:11:22 crc kubenswrapper[4897]: I1121 14:11:22.594779 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tqhs9" event={"ID":"767a13d7-0e18-4496-b0bb-5a39292622e2","Type":"ContainerDied","Data":"6be88453f6e893cdb6bf58762103e7d29f4937b2476a9dde9549dd05fc04fad8"} Nov 21 14:11:22 crc kubenswrapper[4897]: I1121 14:11:22.602706 4897 generic.go:334] "Generic (PLEG): container finished" podID="e8ce22b3-b193-4bd0-9a01-9bc260c20430" containerID="4b959a8e218803f046a5afb07863a1662545b93118ba4d8702a7e6b3aaa2551a" exitCode=0 Nov 21 14:11:22 crc kubenswrapper[4897]: I1121 14:11:22.602756 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5wf4b" event={"ID":"e8ce22b3-b193-4bd0-9a01-9bc260c20430","Type":"ContainerDied","Data":"4b959a8e218803f046a5afb07863a1662545b93118ba4d8702a7e6b3aaa2551a"} Nov 21 14:11:22 crc kubenswrapper[4897]: I1121 14:11:22.609427 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-5z687\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:22 crc kubenswrapper[4897]: I1121 14:11:22.623250 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:22 crc kubenswrapper[4897]: I1121 14:11:22.646526 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf875c4e-691b-4d77-ab68-5ac658b6f39b-utilities\") pod \"redhat-operators-sqdvs\" (UID: \"bf875c4e-691b-4d77-ab68-5ac658b6f39b\") " pod="openshift-marketplace/redhat-operators-sqdvs" Nov 21 14:11:22 crc kubenswrapper[4897]: I1121 14:11:22.646574 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf875c4e-691b-4d77-ab68-5ac658b6f39b-catalog-content\") pod \"redhat-operators-sqdvs\" (UID: \"bf875c4e-691b-4d77-ab68-5ac658b6f39b\") " pod="openshift-marketplace/redhat-operators-sqdvs" Nov 21 14:11:22 crc kubenswrapper[4897]: I1121 14:11:22.646621 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f94h4\" (UniqueName: \"kubernetes.io/projected/bf875c4e-691b-4d77-ab68-5ac658b6f39b-kube-api-access-f94h4\") pod \"redhat-operators-sqdvs\" (UID: \"bf875c4e-691b-4d77-ab68-5ac658b6f39b\") " pod="openshift-marketplace/redhat-operators-sqdvs" Nov 21 14:11:22 crc kubenswrapper[4897]: I1121 14:11:22.653560 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-bjdjf" podStartSLOduration=14.653542724 podStartE2EDuration="14.653542724s" podCreationTimestamp="2025-11-21 14:11:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:11:22.65161058 +0000 UTC m=+159.936204055" watchObservedRunningTime="2025-11-21 14:11:22.653542724 +0000 UTC m=+159.938136199" Nov 21 14:11:22 crc kubenswrapper[4897]: I1121 14:11:22.747564 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf875c4e-691b-4d77-ab68-5ac658b6f39b-utilities\") pod \"redhat-operators-sqdvs\" (UID: \"bf875c4e-691b-4d77-ab68-5ac658b6f39b\") " pod="openshift-marketplace/redhat-operators-sqdvs" Nov 21 14:11:22 crc kubenswrapper[4897]: I1121 14:11:22.747621 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf875c4e-691b-4d77-ab68-5ac658b6f39b-catalog-content\") pod \"redhat-operators-sqdvs\" (UID: \"bf875c4e-691b-4d77-ab68-5ac658b6f39b\") " pod="openshift-marketplace/redhat-operators-sqdvs" Nov 21 14:11:22 crc kubenswrapper[4897]: I1121 14:11:22.747715 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f94h4\" (UniqueName: \"kubernetes.io/projected/bf875c4e-691b-4d77-ab68-5ac658b6f39b-kube-api-access-f94h4\") pod \"redhat-operators-sqdvs\" (UID: \"bf875c4e-691b-4d77-ab68-5ac658b6f39b\") " pod="openshift-marketplace/redhat-operators-sqdvs" Nov 21 14:11:22 crc kubenswrapper[4897]: I1121 14:11:22.748806 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf875c4e-691b-4d77-ab68-5ac658b6f39b-utilities\") pod \"redhat-operators-sqdvs\" (UID: \"bf875c4e-691b-4d77-ab68-5ac658b6f39b\") " pod="openshift-marketplace/redhat-operators-sqdvs" Nov 21 14:11:22 crc kubenswrapper[4897]: I1121 14:11:22.751467 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf875c4e-691b-4d77-ab68-5ac658b6f39b-catalog-content\") pod \"redhat-operators-sqdvs\" (UID: \"bf875c4e-691b-4d77-ab68-5ac658b6f39b\") " pod="openshift-marketplace/redhat-operators-sqdvs" Nov 21 14:11:22 crc kubenswrapper[4897]: I1121 14:11:22.770293 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f94h4\" (UniqueName: \"kubernetes.io/projected/bf875c4e-691b-4d77-ab68-5ac658b6f39b-kube-api-access-f94h4\") pod \"redhat-operators-sqdvs\" (UID: \"bf875c4e-691b-4d77-ab68-5ac658b6f39b\") " pod="openshift-marketplace/redhat-operators-sqdvs" Nov 21 14:11:22 crc kubenswrapper[4897]: I1121 14:11:22.811184 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sqdvs" Nov 21 14:11:22 crc kubenswrapper[4897]: I1121 14:11:22.863740 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29395560-pq7jf" Nov 21 14:11:22 crc kubenswrapper[4897]: I1121 14:11:22.896651 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-5v5dq"] Nov 21 14:11:22 crc kubenswrapper[4897]: E1121 14:11:22.896935 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c14a1c93-8be4-42b7-aae1-9c31d53a6bfc" containerName="collect-profiles" Nov 21 14:11:22 crc kubenswrapper[4897]: I1121 14:11:22.896953 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="c14a1c93-8be4-42b7-aae1-9c31d53a6bfc" containerName="collect-profiles" Nov 21 14:11:22 crc kubenswrapper[4897]: I1121 14:11:22.897067 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="c14a1c93-8be4-42b7-aae1-9c31d53a6bfc" containerName="collect-profiles" Nov 21 14:11:22 crc kubenswrapper[4897]: I1121 14:11:22.897804 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5v5dq"] Nov 21 14:11:22 crc kubenswrapper[4897]: I1121 14:11:22.897886 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5v5dq" Nov 21 14:11:22 crc kubenswrapper[4897]: I1121 14:11:22.940461 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 21 14:11:22 crc kubenswrapper[4897]: I1121 14:11:22.949714 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c14a1c93-8be4-42b7-aae1-9c31d53a6bfc-secret-volume\") pod \"c14a1c93-8be4-42b7-aae1-9c31d53a6bfc\" (UID: \"c14a1c93-8be4-42b7-aae1-9c31d53a6bfc\") " Nov 21 14:11:22 crc kubenswrapper[4897]: I1121 14:11:22.950210 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c14a1c93-8be4-42b7-aae1-9c31d53a6bfc-config-volume\") pod \"c14a1c93-8be4-42b7-aae1-9c31d53a6bfc\" (UID: \"c14a1c93-8be4-42b7-aae1-9c31d53a6bfc\") " Nov 21 14:11:22 crc kubenswrapper[4897]: I1121 14:11:22.950285 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-666j9\" (UniqueName: \"kubernetes.io/projected/c14a1c93-8be4-42b7-aae1-9c31d53a6bfc-kube-api-access-666j9\") pod \"c14a1c93-8be4-42b7-aae1-9c31d53a6bfc\" (UID: \"c14a1c93-8be4-42b7-aae1-9c31d53a6bfc\") " Nov 21 14:11:22 crc kubenswrapper[4897]: I1121 14:11:22.953601 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c14a1c93-8be4-42b7-aae1-9c31d53a6bfc-config-volume" (OuterVolumeSpecName: "config-volume") pod "c14a1c93-8be4-42b7-aae1-9c31d53a6bfc" (UID: "c14a1c93-8be4-42b7-aae1-9c31d53a6bfc"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:11:22 crc kubenswrapper[4897]: I1121 14:11:22.965924 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c14a1c93-8be4-42b7-aae1-9c31d53a6bfc-kube-api-access-666j9" (OuterVolumeSpecName: "kube-api-access-666j9") pod "c14a1c93-8be4-42b7-aae1-9c31d53a6bfc" (UID: "c14a1c93-8be4-42b7-aae1-9c31d53a6bfc"). InnerVolumeSpecName "kube-api-access-666j9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:11:22 crc kubenswrapper[4897]: I1121 14:11:22.968746 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c14a1c93-8be4-42b7-aae1-9c31d53a6bfc-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "c14a1c93-8be4-42b7-aae1-9c31d53a6bfc" (UID: "c14a1c93-8be4-42b7-aae1-9c31d53a6bfc"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:11:22 crc kubenswrapper[4897]: I1121 14:11:22.995277 4897 patch_prober.go:28] interesting pod/router-default-5444994796-2kgg2 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 21 14:11:22 crc kubenswrapper[4897]: [-]has-synced failed: reason withheld Nov 21 14:11:22 crc kubenswrapper[4897]: [+]process-running ok Nov 21 14:11:22 crc kubenswrapper[4897]: healthz check failed Nov 21 14:11:22 crc kubenswrapper[4897]: I1121 14:11:22.995356 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-2kgg2" podUID="e095a18e-3a0b-4418-95a8-a2cc3de5bd7b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 21 14:11:23 crc kubenswrapper[4897]: I1121 14:11:23.053917 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7cd98bfc-77a5-4ef5-86f1-7614b0b2ad2a-kube-api-access\") pod \"7cd98bfc-77a5-4ef5-86f1-7614b0b2ad2a\" (UID: \"7cd98bfc-77a5-4ef5-86f1-7614b0b2ad2a\") " Nov 21 14:11:23 crc kubenswrapper[4897]: I1121 14:11:23.054051 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7cd98bfc-77a5-4ef5-86f1-7614b0b2ad2a-kubelet-dir\") pod \"7cd98bfc-77a5-4ef5-86f1-7614b0b2ad2a\" (UID: \"7cd98bfc-77a5-4ef5-86f1-7614b0b2ad2a\") " Nov 21 14:11:23 crc kubenswrapper[4897]: I1121 14:11:23.054204 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7cd98bfc-77a5-4ef5-86f1-7614b0b2ad2a-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "7cd98bfc-77a5-4ef5-86f1-7614b0b2ad2a" (UID: "7cd98bfc-77a5-4ef5-86f1-7614b0b2ad2a"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 21 14:11:23 crc kubenswrapper[4897]: I1121 14:11:23.054332 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/afe9da5b-bdfb-47c5-bd31-abc7c0d53b63-utilities\") pod \"redhat-operators-5v5dq\" (UID: \"afe9da5b-bdfb-47c5-bd31-abc7c0d53b63\") " pod="openshift-marketplace/redhat-operators-5v5dq" Nov 21 14:11:23 crc kubenswrapper[4897]: I1121 14:11:23.054378 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/afe9da5b-bdfb-47c5-bd31-abc7c0d53b63-catalog-content\") pod \"redhat-operators-5v5dq\" (UID: \"afe9da5b-bdfb-47c5-bd31-abc7c0d53b63\") " pod="openshift-marketplace/redhat-operators-5v5dq" Nov 21 14:11:23 crc kubenswrapper[4897]: I1121 14:11:23.054415 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7pnxh\" (UniqueName: \"kubernetes.io/projected/afe9da5b-bdfb-47c5-bd31-abc7c0d53b63-kube-api-access-7pnxh\") pod \"redhat-operators-5v5dq\" (UID: \"afe9da5b-bdfb-47c5-bd31-abc7c0d53b63\") " pod="openshift-marketplace/redhat-operators-5v5dq" Nov 21 14:11:23 crc kubenswrapper[4897]: I1121 14:11:23.054540 4897 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c14a1c93-8be4-42b7-aae1-9c31d53a6bfc-config-volume\") on node \"crc\" DevicePath \"\"" Nov 21 14:11:23 crc kubenswrapper[4897]: I1121 14:11:23.054558 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-666j9\" (UniqueName: \"kubernetes.io/projected/c14a1c93-8be4-42b7-aae1-9c31d53a6bfc-kube-api-access-666j9\") on node \"crc\" DevicePath \"\"" Nov 21 14:11:23 crc kubenswrapper[4897]: I1121 14:11:23.054574 4897 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c14a1c93-8be4-42b7-aae1-9c31d53a6bfc-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 21 14:11:23 crc kubenswrapper[4897]: I1121 14:11:23.054583 4897 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7cd98bfc-77a5-4ef5-86f1-7614b0b2ad2a-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 21 14:11:23 crc kubenswrapper[4897]: I1121 14:11:23.059760 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7cd98bfc-77a5-4ef5-86f1-7614b0b2ad2a-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "7cd98bfc-77a5-4ef5-86f1-7614b0b2ad2a" (UID: "7cd98bfc-77a5-4ef5-86f1-7614b0b2ad2a"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:11:23 crc kubenswrapper[4897]: I1121 14:11:23.091073 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-sqdvs"] Nov 21 14:11:23 crc kubenswrapper[4897]: W1121 14:11:23.110132 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbf875c4e_691b_4d77_ab68_5ac658b6f39b.slice/crio-fddb7ffc183f994cf60ee875192cf65a4000ea02b86a4615f1a69859ad8eef58 WatchSource:0}: Error finding container fddb7ffc183f994cf60ee875192cf65a4000ea02b86a4615f1a69859ad8eef58: Status 404 returned error can't find the container with id fddb7ffc183f994cf60ee875192cf65a4000ea02b86a4615f1a69859ad8eef58 Nov 21 14:11:23 crc kubenswrapper[4897]: I1121 14:11:23.147574 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-5z687"] Nov 21 14:11:23 crc kubenswrapper[4897]: I1121 14:11:23.155617 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/afe9da5b-bdfb-47c5-bd31-abc7c0d53b63-utilities\") pod \"redhat-operators-5v5dq\" (UID: \"afe9da5b-bdfb-47c5-bd31-abc7c0d53b63\") " pod="openshift-marketplace/redhat-operators-5v5dq" Nov 21 14:11:23 crc kubenswrapper[4897]: I1121 14:11:23.155696 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/afe9da5b-bdfb-47c5-bd31-abc7c0d53b63-catalog-content\") pod \"redhat-operators-5v5dq\" (UID: \"afe9da5b-bdfb-47c5-bd31-abc7c0d53b63\") " pod="openshift-marketplace/redhat-operators-5v5dq" Nov 21 14:11:23 crc kubenswrapper[4897]: I1121 14:11:23.155782 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7pnxh\" (UniqueName: \"kubernetes.io/projected/afe9da5b-bdfb-47c5-bd31-abc7c0d53b63-kube-api-access-7pnxh\") pod \"redhat-operators-5v5dq\" (UID: \"afe9da5b-bdfb-47c5-bd31-abc7c0d53b63\") " pod="openshift-marketplace/redhat-operators-5v5dq" Nov 21 14:11:23 crc kubenswrapper[4897]: I1121 14:11:23.155868 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7cd98bfc-77a5-4ef5-86f1-7614b0b2ad2a-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 21 14:11:23 crc kubenswrapper[4897]: I1121 14:11:23.156657 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/afe9da5b-bdfb-47c5-bd31-abc7c0d53b63-utilities\") pod \"redhat-operators-5v5dq\" (UID: \"afe9da5b-bdfb-47c5-bd31-abc7c0d53b63\") " pod="openshift-marketplace/redhat-operators-5v5dq" Nov 21 14:11:23 crc kubenswrapper[4897]: I1121 14:11:23.157077 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/afe9da5b-bdfb-47c5-bd31-abc7c0d53b63-catalog-content\") pod \"redhat-operators-5v5dq\" (UID: \"afe9da5b-bdfb-47c5-bd31-abc7c0d53b63\") " pod="openshift-marketplace/redhat-operators-5v5dq" Nov 21 14:11:23 crc kubenswrapper[4897]: I1121 14:11:23.176719 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7pnxh\" (UniqueName: \"kubernetes.io/projected/afe9da5b-bdfb-47c5-bd31-abc7c0d53b63-kube-api-access-7pnxh\") pod \"redhat-operators-5v5dq\" (UID: \"afe9da5b-bdfb-47c5-bd31-abc7c0d53b63\") " pod="openshift-marketplace/redhat-operators-5v5dq" Nov 21 14:11:23 crc kubenswrapper[4897]: I1121 14:11:23.238615 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5v5dq" Nov 21 14:11:23 crc kubenswrapper[4897]: I1121 14:11:23.387639 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-lq8bp" Nov 21 14:11:23 crc kubenswrapper[4897]: I1121 14:11:23.486623 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5v5dq"] Nov 21 14:11:23 crc kubenswrapper[4897]: W1121 14:11:23.506070 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podafe9da5b_bdfb_47c5_bd31_abc7c0d53b63.slice/crio-eb3afaa57ab38165909faf494ca5d1fd2a75ce94bfe1a4e90fbfe0eb1dff17f0 WatchSource:0}: Error finding container eb3afaa57ab38165909faf494ca5d1fd2a75ce94bfe1a4e90fbfe0eb1dff17f0: Status 404 returned error can't find the container with id eb3afaa57ab38165909faf494ca5d1fd2a75ce94bfe1a4e90fbfe0eb1dff17f0 Nov 21 14:11:23 crc kubenswrapper[4897]: I1121 14:11:23.612914 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29395560-pq7jf" event={"ID":"c14a1c93-8be4-42b7-aae1-9c31d53a6bfc","Type":"ContainerDied","Data":"f11286e8fc2e30ab9359b216a9923028c1bfc07c1ca7b16681bcd9b4acdcefae"} Nov 21 14:11:23 crc kubenswrapper[4897]: I1121 14:11:23.612978 4897 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f11286e8fc2e30ab9359b216a9923028c1bfc07c1ca7b16681bcd9b4acdcefae" Nov 21 14:11:23 crc kubenswrapper[4897]: I1121 14:11:23.612943 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29395560-pq7jf" Nov 21 14:11:23 crc kubenswrapper[4897]: I1121 14:11:23.617236 4897 generic.go:334] "Generic (PLEG): container finished" podID="bf875c4e-691b-4d77-ab68-5ac658b6f39b" containerID="423decd312997271961e228a25768e0f740f3fb7a02fcf8559757f5655bbf1d1" exitCode=0 Nov 21 14:11:23 crc kubenswrapper[4897]: I1121 14:11:23.617347 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sqdvs" event={"ID":"bf875c4e-691b-4d77-ab68-5ac658b6f39b","Type":"ContainerDied","Data":"423decd312997271961e228a25768e0f740f3fb7a02fcf8559757f5655bbf1d1"} Nov 21 14:11:23 crc kubenswrapper[4897]: I1121 14:11:23.617402 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sqdvs" event={"ID":"bf875c4e-691b-4d77-ab68-5ac658b6f39b","Type":"ContainerStarted","Data":"fddb7ffc183f994cf60ee875192cf65a4000ea02b86a4615f1a69859ad8eef58"} Nov 21 14:11:23 crc kubenswrapper[4897]: I1121 14:11:23.622942 4897 generic.go:334] "Generic (PLEG): container finished" podID="a55d260f-3da6-4b30-b633-6348b31b0b01" containerID="ffedeae3a60a10cd4bb4fa9046a20001bbb9798139976bc047843ed79798324a" exitCode=0 Nov 21 14:11:23 crc kubenswrapper[4897]: I1121 14:11:23.623118 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k2ctv" event={"ID":"a55d260f-3da6-4b30-b633-6348b31b0b01","Type":"ContainerDied","Data":"ffedeae3a60a10cd4bb4fa9046a20001bbb9798139976bc047843ed79798324a"} Nov 21 14:11:23 crc kubenswrapper[4897]: I1121 14:11:23.624812 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5v5dq" event={"ID":"afe9da5b-bdfb-47c5-bd31-abc7c0d53b63","Type":"ContainerStarted","Data":"eb3afaa57ab38165909faf494ca5d1fd2a75ce94bfe1a4e90fbfe0eb1dff17f0"} Nov 21 14:11:23 crc kubenswrapper[4897]: I1121 14:11:23.629330 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"7cd98bfc-77a5-4ef5-86f1-7614b0b2ad2a","Type":"ContainerDied","Data":"eb6f02f2c6ecbb8ee1c30e3fc050f7b4e24b3f424c1396bb90f147004f2196fb"} Nov 21 14:11:23 crc kubenswrapper[4897]: I1121 14:11:23.629367 4897 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eb6f02f2c6ecbb8ee1c30e3fc050f7b4e24b3f424c1396bb90f147004f2196fb" Nov 21 14:11:23 crc kubenswrapper[4897]: I1121 14:11:23.629440 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 21 14:11:23 crc kubenswrapper[4897]: I1121 14:11:23.651007 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-5z687" event={"ID":"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b","Type":"ContainerStarted","Data":"79ade4cd758c4fb4a6d2ad5e6f0ed42b487b6acd04c14586c067c870f5026815"} Nov 21 14:11:23 crc kubenswrapper[4897]: I1121 14:11:23.651087 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-5z687" event={"ID":"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b","Type":"ContainerStarted","Data":"d3abd9e080efefb4cc6168f2aa1c73f860bf0a037e32d66c5ff618ed90540b37"} Nov 21 14:11:23 crc kubenswrapper[4897]: I1121 14:11:23.686223 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-5z687" podStartSLOduration=134.686197212 podStartE2EDuration="2m14.686197212s" podCreationTimestamp="2025-11-21 14:09:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:11:23.679498653 +0000 UTC m=+160.964106678" watchObservedRunningTime="2025-11-21 14:11:23.686197212 +0000 UTC m=+160.970790707" Nov 21 14:11:23 crc kubenswrapper[4897]: I1121 14:11:23.938048 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 21 14:11:23 crc kubenswrapper[4897]: I1121 14:11:23.992902 4897 patch_prober.go:28] interesting pod/router-default-5444994796-2kgg2 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 21 14:11:23 crc kubenswrapper[4897]: [-]has-synced failed: reason withheld Nov 21 14:11:23 crc kubenswrapper[4897]: [+]process-running ok Nov 21 14:11:23 crc kubenswrapper[4897]: healthz check failed Nov 21 14:11:23 crc kubenswrapper[4897]: I1121 14:11:23.992969 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-2kgg2" podUID="e095a18e-3a0b-4418-95a8-a2cc3de5bd7b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 21 14:11:24 crc kubenswrapper[4897]: I1121 14:11:24.069113 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b5aa94f1-611d-4695-a3c7-1c33c80f1a1a-kubelet-dir\") pod \"b5aa94f1-611d-4695-a3c7-1c33c80f1a1a\" (UID: \"b5aa94f1-611d-4695-a3c7-1c33c80f1a1a\") " Nov 21 14:11:24 crc kubenswrapper[4897]: I1121 14:11:24.069195 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b5aa94f1-611d-4695-a3c7-1c33c80f1a1a-kube-api-access\") pod \"b5aa94f1-611d-4695-a3c7-1c33c80f1a1a\" (UID: \"b5aa94f1-611d-4695-a3c7-1c33c80f1a1a\") " Nov 21 14:11:24 crc kubenswrapper[4897]: I1121 14:11:24.069258 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b5aa94f1-611d-4695-a3c7-1c33c80f1a1a-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "b5aa94f1-611d-4695-a3c7-1c33c80f1a1a" (UID: "b5aa94f1-611d-4695-a3c7-1c33c80f1a1a"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 21 14:11:24 crc kubenswrapper[4897]: I1121 14:11:24.069737 4897 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b5aa94f1-611d-4695-a3c7-1c33c80f1a1a-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 21 14:11:24 crc kubenswrapper[4897]: I1121 14:11:24.083251 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5aa94f1-611d-4695-a3c7-1c33c80f1a1a-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "b5aa94f1-611d-4695-a3c7-1c33c80f1a1a" (UID: "b5aa94f1-611d-4695-a3c7-1c33c80f1a1a"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:11:24 crc kubenswrapper[4897]: I1121 14:11:24.130164 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Nov 21 14:11:24 crc kubenswrapper[4897]: I1121 14:11:24.171545 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b5aa94f1-611d-4695-a3c7-1c33c80f1a1a-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 21 14:11:24 crc kubenswrapper[4897]: I1121 14:11:24.663377 4897 generic.go:334] "Generic (PLEG): container finished" podID="afe9da5b-bdfb-47c5-bd31-abc7c0d53b63" containerID="8b359f86d7e327a550d9c0e7965014564aafab42158f162dfa81b4054c08b683" exitCode=0 Nov 21 14:11:24 crc kubenswrapper[4897]: I1121 14:11:24.663471 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5v5dq" event={"ID":"afe9da5b-bdfb-47c5-bd31-abc7c0d53b63","Type":"ContainerDied","Data":"8b359f86d7e327a550d9c0e7965014564aafab42158f162dfa81b4054c08b683"} Nov 21 14:11:24 crc kubenswrapper[4897]: I1121 14:11:24.669912 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 21 14:11:24 crc kubenswrapper[4897]: I1121 14:11:24.670204 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"b5aa94f1-611d-4695-a3c7-1c33c80f1a1a","Type":"ContainerDied","Data":"2d5180bb09d2b710821543745c7670fc20f3417633851aca10af552d8bed0042"} Nov 21 14:11:24 crc kubenswrapper[4897]: I1121 14:11:24.670299 4897 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2d5180bb09d2b710821543745c7670fc20f3417633851aca10af552d8bed0042" Nov 21 14:11:24 crc kubenswrapper[4897]: I1121 14:11:24.670408 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:24 crc kubenswrapper[4897]: I1121 14:11:24.992371 4897 patch_prober.go:28] interesting pod/router-default-5444994796-2kgg2 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 21 14:11:24 crc kubenswrapper[4897]: [-]has-synced failed: reason withheld Nov 21 14:11:24 crc kubenswrapper[4897]: [+]process-running ok Nov 21 14:11:24 crc kubenswrapper[4897]: healthz check failed Nov 21 14:11:24 crc kubenswrapper[4897]: I1121 14:11:24.992460 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-2kgg2" podUID="e095a18e-3a0b-4418-95a8-a2cc3de5bd7b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 21 14:11:25 crc kubenswrapper[4897]: I1121 14:11:25.992257 4897 patch_prober.go:28] interesting pod/router-default-5444994796-2kgg2 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 21 14:11:25 crc kubenswrapper[4897]: [-]has-synced failed: reason withheld Nov 21 14:11:25 crc kubenswrapper[4897]: [+]process-running ok Nov 21 14:11:25 crc kubenswrapper[4897]: healthz check failed Nov 21 14:11:25 crc kubenswrapper[4897]: I1121 14:11:25.992921 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-2kgg2" podUID="e095a18e-3a0b-4418-95a8-a2cc3de5bd7b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 21 14:11:26 crc kubenswrapper[4897]: I1121 14:11:26.376629 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-s7tww" Nov 21 14:11:26 crc kubenswrapper[4897]: I1121 14:11:26.389063 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-s7tww" Nov 21 14:11:26 crc kubenswrapper[4897]: I1121 14:11:26.995399 4897 patch_prober.go:28] interesting pod/router-default-5444994796-2kgg2 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 21 14:11:26 crc kubenswrapper[4897]: [-]has-synced failed: reason withheld Nov 21 14:11:26 crc kubenswrapper[4897]: [+]process-running ok Nov 21 14:11:26 crc kubenswrapper[4897]: healthz check failed Nov 21 14:11:26 crc kubenswrapper[4897]: I1121 14:11:26.995473 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-2kgg2" podUID="e095a18e-3a0b-4418-95a8-a2cc3de5bd7b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 21 14:11:27 crc kubenswrapper[4897]: I1121 14:11:27.991195 4897 patch_prober.go:28] interesting pod/router-default-5444994796-2kgg2 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 21 14:11:27 crc kubenswrapper[4897]: [-]has-synced failed: reason withheld Nov 21 14:11:27 crc kubenswrapper[4897]: [+]process-running ok Nov 21 14:11:27 crc kubenswrapper[4897]: healthz check failed Nov 21 14:11:27 crc kubenswrapper[4897]: I1121 14:11:27.991380 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-2kgg2" podUID="e095a18e-3a0b-4418-95a8-a2cc3de5bd7b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 21 14:11:28 crc kubenswrapper[4897]: I1121 14:11:28.991720 4897 patch_prober.go:28] interesting pod/router-default-5444994796-2kgg2 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 21 14:11:28 crc kubenswrapper[4897]: [-]has-synced failed: reason withheld Nov 21 14:11:28 crc kubenswrapper[4897]: [+]process-running ok Nov 21 14:11:28 crc kubenswrapper[4897]: healthz check failed Nov 21 14:11:28 crc kubenswrapper[4897]: I1121 14:11:28.992269 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-2kgg2" podUID="e095a18e-3a0b-4418-95a8-a2cc3de5bd7b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 21 14:11:29 crc kubenswrapper[4897]: I1121 14:11:29.990353 4897 patch_prober.go:28] interesting pod/router-default-5444994796-2kgg2 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 21 14:11:29 crc kubenswrapper[4897]: [-]has-synced failed: reason withheld Nov 21 14:11:29 crc kubenswrapper[4897]: [+]process-running ok Nov 21 14:11:29 crc kubenswrapper[4897]: healthz check failed Nov 21 14:11:29 crc kubenswrapper[4897]: I1121 14:11:29.990436 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-2kgg2" podUID="e095a18e-3a0b-4418-95a8-a2cc3de5bd7b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 21 14:11:30 crc kubenswrapper[4897]: I1121 14:11:30.889082 4897 patch_prober.go:28] interesting pod/downloads-7954f5f757-vgwv2 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.26:8080/\": dial tcp 10.217.0.26:8080: connect: connection refused" start-of-body= Nov 21 14:11:30 crc kubenswrapper[4897]: I1121 14:11:30.889146 4897 patch_prober.go:28] interesting pod/downloads-7954f5f757-vgwv2 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.26:8080/\": dial tcp 10.217.0.26:8080: connect: connection refused" start-of-body= Nov 21 14:11:30 crc kubenswrapper[4897]: I1121 14:11:30.889157 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-vgwv2" podUID="82f81ac2-d2b2-49de-82f3-a51796dbff03" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.26:8080/\": dial tcp 10.217.0.26:8080: connect: connection refused" Nov 21 14:11:30 crc kubenswrapper[4897]: I1121 14:11:30.889213 4897 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-vgwv2" podUID="82f81ac2-d2b2-49de-82f3-a51796dbff03" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.26:8080/\": dial tcp 10.217.0.26:8080: connect: connection refused" Nov 21 14:11:30 crc kubenswrapper[4897]: I1121 14:11:30.903283 4897 patch_prober.go:28] interesting pod/console-f9d7485db-7fsr9 container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.18:8443/health\": dial tcp 10.217.0.18:8443: connect: connection refused" start-of-body= Nov 21 14:11:30 crc kubenswrapper[4897]: I1121 14:11:30.903347 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-7fsr9" podUID="8342c8a2-87f4-4bf9-a2ca-073c02b277ad" containerName="console" probeResult="failure" output="Get \"https://10.217.0.18:8443/health\": dial tcp 10.217.0.18:8443: connect: connection refused" Nov 21 14:11:30 crc kubenswrapper[4897]: I1121 14:11:30.993420 4897 patch_prober.go:28] interesting pod/router-default-5444994796-2kgg2 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 21 14:11:30 crc kubenswrapper[4897]: [-]has-synced failed: reason withheld Nov 21 14:11:30 crc kubenswrapper[4897]: [+]process-running ok Nov 21 14:11:30 crc kubenswrapper[4897]: healthz check failed Nov 21 14:11:30 crc kubenswrapper[4897]: I1121 14:11:30.993484 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-2kgg2" podUID="e095a18e-3a0b-4418-95a8-a2cc3de5bd7b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 21 14:11:31 crc kubenswrapper[4897]: I1121 14:11:31.787972 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd-metrics-certs\") pod \"network-metrics-daemon-rs4c8\" (UID: \"01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd\") " pod="openshift-multus/network-metrics-daemon-rs4c8" Nov 21 14:11:31 crc kubenswrapper[4897]: I1121 14:11:31.793259 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd-metrics-certs\") pod \"network-metrics-daemon-rs4c8\" (UID: \"01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd\") " pod="openshift-multus/network-metrics-daemon-rs4c8" Nov 21 14:11:31 crc kubenswrapper[4897]: I1121 14:11:31.991265 4897 patch_prober.go:28] interesting pod/router-default-5444994796-2kgg2 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 21 14:11:31 crc kubenswrapper[4897]: [-]has-synced failed: reason withheld Nov 21 14:11:31 crc kubenswrapper[4897]: [+]process-running ok Nov 21 14:11:31 crc kubenswrapper[4897]: healthz check failed Nov 21 14:11:31 crc kubenswrapper[4897]: I1121 14:11:31.991319 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-2kgg2" podUID="e095a18e-3a0b-4418-95a8-a2cc3de5bd7b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 21 14:11:32 crc kubenswrapper[4897]: I1121 14:11:32.015792 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-rs4c8" Nov 21 14:11:32 crc kubenswrapper[4897]: I1121 14:11:32.992827 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-2kgg2" Nov 21 14:11:32 crc kubenswrapper[4897]: I1121 14:11:32.995543 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-2kgg2" Nov 21 14:11:34 crc kubenswrapper[4897]: I1121 14:11:34.371215 4897 patch_prober.go:28] interesting pod/machine-config-daemon-krv5b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 21 14:11:34 crc kubenswrapper[4897]: I1121 14:11:34.371349 4897 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 21 14:11:40 crc kubenswrapper[4897]: I1121 14:11:40.904373 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-vgwv2" Nov 21 14:11:40 crc kubenswrapper[4897]: I1121 14:11:40.953057 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-7fsr9" Nov 21 14:11:40 crc kubenswrapper[4897]: I1121 14:11:40.957961 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-7fsr9" Nov 21 14:11:42 crc kubenswrapper[4897]: I1121 14:11:42.630746 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:11:51 crc kubenswrapper[4897]: I1121 14:11:51.318606 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-cc48c" Nov 21 14:11:53 crc kubenswrapper[4897]: E1121 14:11:53.706844 4897 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Nov 21 14:11:53 crc kubenswrapper[4897]: E1121 14:11:53.707051 4897 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-59rxs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-tqhs9_openshift-marketplace(767a13d7-0e18-4496-b0bb-5a39292622e2): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 21 14:11:53 crc kubenswrapper[4897]: E1121 14:11:53.708316 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-tqhs9" podUID="767a13d7-0e18-4496-b0bb-5a39292622e2" Nov 21 14:11:55 crc kubenswrapper[4897]: E1121 14:11:55.548229 4897 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Nov 21 14:11:55 crc kubenswrapper[4897]: E1121 14:11:55.549672 4897 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4x5qq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-6fqg7_openshift-marketplace(8dde9f26-f67d-49d2-98ac-7483796f1212): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 21 14:11:55 crc kubenswrapper[4897]: E1121 14:11:55.551355 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-6fqg7" podUID="8dde9f26-f67d-49d2-98ac-7483796f1212" Nov 21 14:11:57 crc kubenswrapper[4897]: I1121 14:11:57.804027 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 21 14:12:01 crc kubenswrapper[4897]: E1121 14:12:01.113954 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-6fqg7" podUID="8dde9f26-f67d-49d2-98ac-7483796f1212" Nov 21 14:12:01 crc kubenswrapper[4897]: E1121 14:12:01.113901 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-tqhs9" podUID="767a13d7-0e18-4496-b0bb-5a39292622e2" Nov 21 14:12:04 crc kubenswrapper[4897]: I1121 14:12:04.370925 4897 patch_prober.go:28] interesting pod/machine-config-daemon-krv5b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 21 14:12:04 crc kubenswrapper[4897]: I1121 14:12:04.371297 4897 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 21 14:12:07 crc kubenswrapper[4897]: E1121 14:12:07.561721 4897 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Nov 21 14:12:07 crc kubenswrapper[4897]: E1121 14:12:07.562739 4897 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-l5xws,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-q6dd2_openshift-marketplace(1dc1b877-a8de-45b0-bab4-4411ecced04d): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 21 14:12:07 crc kubenswrapper[4897]: E1121 14:12:07.563950 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-q6dd2" podUID="1dc1b877-a8de-45b0-bab4-4411ecced04d" Nov 21 14:12:16 crc kubenswrapper[4897]: E1121 14:12:16.009194 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-q6dd2" podUID="1dc1b877-a8de-45b0-bab4-4411ecced04d" Nov 21 14:12:16 crc kubenswrapper[4897]: I1121 14:12:16.428390 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-rs4c8"] Nov 21 14:12:16 crc kubenswrapper[4897]: W1121 14:12:16.443730 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod01c4dbd9_8e9d_441a_bfdb_7e0cb5ffdfbd.slice/crio-e7b58cf9fbb2cd87842f18d226cd75ed75b381719dada26f875e68e41eef375f WatchSource:0}: Error finding container e7b58cf9fbb2cd87842f18d226cd75ed75b381719dada26f875e68e41eef375f: Status 404 returned error can't find the container with id e7b58cf9fbb2cd87842f18d226cd75ed75b381719dada26f875e68e41eef375f Nov 21 14:12:16 crc kubenswrapper[4897]: E1121 14:12:16.544075 4897 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Nov 21 14:12:16 crc kubenswrapper[4897]: E1121 14:12:16.544474 4897 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7pnxh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-5v5dq_openshift-marketplace(afe9da5b-bdfb-47c5-bd31-abc7c0d53b63): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 21 14:12:16 crc kubenswrapper[4897]: E1121 14:12:16.545740 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-5v5dq" podUID="afe9da5b-bdfb-47c5-bd31-abc7c0d53b63" Nov 21 14:12:16 crc kubenswrapper[4897]: E1121 14:12:16.547133 4897 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Nov 21 14:12:16 crc kubenswrapper[4897]: E1121 14:12:16.547298 4897 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jc9hp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-k2ctv_openshift-marketplace(a55d260f-3da6-4b30-b633-6348b31b0b01): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 21 14:12:16 crc kubenswrapper[4897]: E1121 14:12:16.548737 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-k2ctv" podUID="a55d260f-3da6-4b30-b633-6348b31b0b01" Nov 21 14:12:16 crc kubenswrapper[4897]: E1121 14:12:16.622644 4897 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Nov 21 14:12:16 crc kubenswrapper[4897]: E1121 14:12:16.623077 4897 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-55ctt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-5wf4b_openshift-marketplace(e8ce22b3-b193-4bd0-9a01-9bc260c20430): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 21 14:12:16 crc kubenswrapper[4897]: E1121 14:12:16.624466 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-5wf4b" podUID="e8ce22b3-b193-4bd0-9a01-9bc260c20430" Nov 21 14:12:16 crc kubenswrapper[4897]: I1121 14:12:16.989429 4897 generic.go:334] "Generic (PLEG): container finished" podID="5ea4ecd1-6706-4a89-8034-4f41ee7e0f2d" containerID="e34bf3228c81cca708e41bf3fc3e1b5ed56fa5fd3e214da58ecf2aa67ec8661f" exitCode=0 Nov 21 14:12:16 crc kubenswrapper[4897]: I1121 14:12:16.989545 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jf4pv" event={"ID":"5ea4ecd1-6706-4a89-8034-4f41ee7e0f2d","Type":"ContainerDied","Data":"e34bf3228c81cca708e41bf3fc3e1b5ed56fa5fd3e214da58ecf2aa67ec8661f"} Nov 21 14:12:16 crc kubenswrapper[4897]: I1121 14:12:16.991442 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sqdvs" event={"ID":"bf875c4e-691b-4d77-ab68-5ac658b6f39b","Type":"ContainerStarted","Data":"52dc206c72235c5655018ca60c89bc6e0406c05e6ccad554b5c6930b171b6d52"} Nov 21 14:12:16 crc kubenswrapper[4897]: I1121 14:12:16.993584 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-rs4c8" event={"ID":"01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd","Type":"ContainerStarted","Data":"e04460c9ef855ffa4ec3c6e943363296d1d61c36a0de0642d1f4fd26e0258fa5"} Nov 21 14:12:16 crc kubenswrapper[4897]: I1121 14:12:16.993652 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-rs4c8" event={"ID":"01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd","Type":"ContainerStarted","Data":"6face73773694a9f802b961942cf82bfb7cd8385dcc79b19178522a630d8166b"} Nov 21 14:12:16 crc kubenswrapper[4897]: I1121 14:12:16.993665 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-rs4c8" event={"ID":"01c4dbd9-8e9d-441a-bfdb-7e0cb5ffdfbd","Type":"ContainerStarted","Data":"e7b58cf9fbb2cd87842f18d226cd75ed75b381719dada26f875e68e41eef375f"} Nov 21 14:12:17 crc kubenswrapper[4897]: I1121 14:12:17.049948 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-rs4c8" podStartSLOduration=188.04992925 podStartE2EDuration="3m8.04992925s" podCreationTimestamp="2025-11-21 14:09:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:12:17.046621174 +0000 UTC m=+214.331214649" watchObservedRunningTime="2025-11-21 14:12:17.04992925 +0000 UTC m=+214.334522715" Nov 21 14:12:17 crc kubenswrapper[4897]: E1121 14:12:17.062809 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-5wf4b" podUID="e8ce22b3-b193-4bd0-9a01-9bc260c20430" Nov 21 14:12:17 crc kubenswrapper[4897]: E1121 14:12:17.063175 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-k2ctv" podUID="a55d260f-3da6-4b30-b633-6348b31b0b01" Nov 21 14:12:17 crc kubenswrapper[4897]: E1121 14:12:17.075207 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-5v5dq" podUID="afe9da5b-bdfb-47c5-bd31-abc7c0d53b63" Nov 21 14:12:18 crc kubenswrapper[4897]: I1121 14:12:18.001611 4897 generic.go:334] "Generic (PLEG): container finished" podID="767a13d7-0e18-4496-b0bb-5a39292622e2" containerID="404afe9d100084f246838f84effde9c75118dddb3d6f048061e3292125b4ca03" exitCode=0 Nov 21 14:12:18 crc kubenswrapper[4897]: I1121 14:12:18.001679 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tqhs9" event={"ID":"767a13d7-0e18-4496-b0bb-5a39292622e2","Type":"ContainerDied","Data":"404afe9d100084f246838f84effde9c75118dddb3d6f048061e3292125b4ca03"} Nov 21 14:12:18 crc kubenswrapper[4897]: I1121 14:12:18.005779 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jf4pv" event={"ID":"5ea4ecd1-6706-4a89-8034-4f41ee7e0f2d","Type":"ContainerStarted","Data":"0222fc2020cd78e6f8fa1b4c2a4848439fcc1dc12c9df77757977aa632c1321c"} Nov 21 14:12:18 crc kubenswrapper[4897]: I1121 14:12:18.008353 4897 generic.go:334] "Generic (PLEG): container finished" podID="bf875c4e-691b-4d77-ab68-5ac658b6f39b" containerID="52dc206c72235c5655018ca60c89bc6e0406c05e6ccad554b5c6930b171b6d52" exitCode=0 Nov 21 14:12:18 crc kubenswrapper[4897]: I1121 14:12:18.008451 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sqdvs" event={"ID":"bf875c4e-691b-4d77-ab68-5ac658b6f39b","Type":"ContainerDied","Data":"52dc206c72235c5655018ca60c89bc6e0406c05e6ccad554b5c6930b171b6d52"} Nov 21 14:12:18 crc kubenswrapper[4897]: I1121 14:12:18.015194 4897 generic.go:334] "Generic (PLEG): container finished" podID="8dde9f26-f67d-49d2-98ac-7483796f1212" containerID="46112836ee5444839fd76ed891b605d0f9a89cc0b735c1e6aefcc0e23623a60d" exitCode=0 Nov 21 14:12:18 crc kubenswrapper[4897]: I1121 14:12:18.015270 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6fqg7" event={"ID":"8dde9f26-f67d-49d2-98ac-7483796f1212","Type":"ContainerDied","Data":"46112836ee5444839fd76ed891b605d0f9a89cc0b735c1e6aefcc0e23623a60d"} Nov 21 14:12:18 crc kubenswrapper[4897]: I1121 14:12:18.051374 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-jf4pv" podStartSLOduration=4.131973318 podStartE2EDuration="59.051353778s" podCreationTimestamp="2025-11-21 14:11:19 +0000 UTC" firstStartedPulling="2025-11-21 14:11:22.574057468 +0000 UTC m=+159.858650943" lastFinishedPulling="2025-11-21 14:12:17.493437928 +0000 UTC m=+214.778031403" observedRunningTime="2025-11-21 14:12:18.050593647 +0000 UTC m=+215.335187122" watchObservedRunningTime="2025-11-21 14:12:18.051353778 +0000 UTC m=+215.335947253" Nov 21 14:12:19 crc kubenswrapper[4897]: I1121 14:12:19.021917 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sqdvs" event={"ID":"bf875c4e-691b-4d77-ab68-5ac658b6f39b","Type":"ContainerStarted","Data":"6936da99ee2db8e16a1af98a86c4cc9c32ec07a2c11853dff43fb6c255021576"} Nov 21 14:12:19 crc kubenswrapper[4897]: I1121 14:12:19.023173 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6fqg7" event={"ID":"8dde9f26-f67d-49d2-98ac-7483796f1212","Type":"ContainerStarted","Data":"2ca36527ca91ac00bf01ae7ada651924905110730aba5170b1d3d43a5aa97ab7"} Nov 21 14:12:19 crc kubenswrapper[4897]: I1121 14:12:19.026336 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tqhs9" event={"ID":"767a13d7-0e18-4496-b0bb-5a39292622e2","Type":"ContainerStarted","Data":"01d955c394e426a56d3b10050ddce048da4f9940dbdf4902063201002c6aad75"} Nov 21 14:12:19 crc kubenswrapper[4897]: I1121 14:12:19.041018 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-sqdvs" podStartSLOduration=2.169515814 podStartE2EDuration="57.040998402s" podCreationTimestamp="2025-11-21 14:11:22 +0000 UTC" firstStartedPulling="2025-11-21 14:11:23.620948627 +0000 UTC m=+160.905542102" lastFinishedPulling="2025-11-21 14:12:18.492431215 +0000 UTC m=+215.777024690" observedRunningTime="2025-11-21 14:12:19.040078675 +0000 UTC m=+216.324672150" watchObservedRunningTime="2025-11-21 14:12:19.040998402 +0000 UTC m=+216.325591877" Nov 21 14:12:19 crc kubenswrapper[4897]: I1121 14:12:19.057556 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-6fqg7" podStartSLOduration=2.04746837 podStartE2EDuration="58.057539196s" podCreationTimestamp="2025-11-21 14:11:21 +0000 UTC" firstStartedPulling="2025-11-21 14:11:22.583299088 +0000 UTC m=+159.867892563" lastFinishedPulling="2025-11-21 14:12:18.593369914 +0000 UTC m=+215.877963389" observedRunningTime="2025-11-21 14:12:19.054296851 +0000 UTC m=+216.338890326" watchObservedRunningTime="2025-11-21 14:12:19.057539196 +0000 UTC m=+216.342132671" Nov 21 14:12:19 crc kubenswrapper[4897]: I1121 14:12:19.073198 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-tqhs9" podStartSLOduration=4.196941053 podStartE2EDuration="1m0.073181863s" podCreationTimestamp="2025-11-21 14:11:19 +0000 UTC" firstStartedPulling="2025-11-21 14:11:22.598694024 +0000 UTC m=+159.883287499" lastFinishedPulling="2025-11-21 14:12:18.474934834 +0000 UTC m=+215.759528309" observedRunningTime="2025-11-21 14:12:19.070573196 +0000 UTC m=+216.355166671" watchObservedRunningTime="2025-11-21 14:12:19.073181863 +0000 UTC m=+216.357775338" Nov 21 14:12:19 crc kubenswrapper[4897]: I1121 14:12:19.849376 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-tqhs9" Nov 21 14:12:19 crc kubenswrapper[4897]: I1121 14:12:19.849477 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-tqhs9" Nov 21 14:12:20 crc kubenswrapper[4897]: I1121 14:12:20.266379 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-jf4pv" Nov 21 14:12:20 crc kubenswrapper[4897]: I1121 14:12:20.266444 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-jf4pv" Nov 21 14:12:20 crc kubenswrapper[4897]: I1121 14:12:20.315021 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-jf4pv" Nov 21 14:12:20 crc kubenswrapper[4897]: I1121 14:12:20.994281 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-tqhs9" podUID="767a13d7-0e18-4496-b0bb-5a39292622e2" containerName="registry-server" probeResult="failure" output=< Nov 21 14:12:20 crc kubenswrapper[4897]: timeout: failed to connect service ":50051" within 1s Nov 21 14:12:20 crc kubenswrapper[4897]: > Nov 21 14:12:21 crc kubenswrapper[4897]: I1121 14:12:21.798558 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-6fqg7" Nov 21 14:12:21 crc kubenswrapper[4897]: I1121 14:12:21.798980 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-6fqg7" Nov 21 14:12:21 crc kubenswrapper[4897]: I1121 14:12:21.847458 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-6fqg7" Nov 21 14:12:22 crc kubenswrapper[4897]: I1121 14:12:22.811728 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-sqdvs" Nov 21 14:12:22 crc kubenswrapper[4897]: I1121 14:12:22.812068 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-sqdvs" Nov 21 14:12:23 crc kubenswrapper[4897]: I1121 14:12:23.856871 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-sqdvs" podUID="bf875c4e-691b-4d77-ab68-5ac658b6f39b" containerName="registry-server" probeResult="failure" output=< Nov 21 14:12:23 crc kubenswrapper[4897]: timeout: failed to connect service ":50051" within 1s Nov 21 14:12:23 crc kubenswrapper[4897]: > Nov 21 14:12:26 crc kubenswrapper[4897]: I1121 14:12:26.525596 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-2trmp"] Nov 21 14:12:29 crc kubenswrapper[4897]: I1121 14:12:29.890388 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-tqhs9" Nov 21 14:12:29 crc kubenswrapper[4897]: I1121 14:12:29.928242 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-tqhs9" Nov 21 14:12:30 crc kubenswrapper[4897]: I1121 14:12:30.324100 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-jf4pv" Nov 21 14:12:31 crc kubenswrapper[4897]: I1121 14:12:31.845079 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-6fqg7" Nov 21 14:12:32 crc kubenswrapper[4897]: I1121 14:12:32.219839 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jf4pv"] Nov 21 14:12:32 crc kubenswrapper[4897]: I1121 14:12:32.220133 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-jf4pv" podUID="5ea4ecd1-6706-4a89-8034-4f41ee7e0f2d" containerName="registry-server" containerID="cri-o://0222fc2020cd78e6f8fa1b4c2a4848439fcc1dc12c9df77757977aa632c1321c" gracePeriod=2 Nov 21 14:12:32 crc kubenswrapper[4897]: I1121 14:12:32.922222 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-sqdvs" Nov 21 14:12:32 crc kubenswrapper[4897]: I1121 14:12:32.962141 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-sqdvs" Nov 21 14:12:33 crc kubenswrapper[4897]: I1121 14:12:33.100524 4897 generic.go:334] "Generic (PLEG): container finished" podID="5ea4ecd1-6706-4a89-8034-4f41ee7e0f2d" containerID="0222fc2020cd78e6f8fa1b4c2a4848439fcc1dc12c9df77757977aa632c1321c" exitCode=0 Nov 21 14:12:33 crc kubenswrapper[4897]: I1121 14:12:33.100774 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jf4pv" event={"ID":"5ea4ecd1-6706-4a89-8034-4f41ee7e0f2d","Type":"ContainerDied","Data":"0222fc2020cd78e6f8fa1b4c2a4848439fcc1dc12c9df77757977aa632c1321c"} Nov 21 14:12:34 crc kubenswrapper[4897]: I1121 14:12:34.371301 4897 patch_prober.go:28] interesting pod/machine-config-daemon-krv5b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 21 14:12:34 crc kubenswrapper[4897]: I1121 14:12:34.371359 4897 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 21 14:12:34 crc kubenswrapper[4897]: I1121 14:12:34.371407 4897 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" Nov 21 14:12:34 crc kubenswrapper[4897]: I1121 14:12:34.372008 4897 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5854028821e6a360a259427b68e7bb4c2a7220e0c68f4cc91639b2ece0540a02"} pod="openshift-machine-config-operator/machine-config-daemon-krv5b" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 21 14:12:34 crc kubenswrapper[4897]: I1121 14:12:34.372121 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" containerName="machine-config-daemon" containerID="cri-o://5854028821e6a360a259427b68e7bb4c2a7220e0c68f4cc91639b2ece0540a02" gracePeriod=600 Nov 21 14:12:35 crc kubenswrapper[4897]: I1121 14:12:35.112125 4897 generic.go:334] "Generic (PLEG): container finished" podID="e7670227-d280-4847-b882-754429f56b0f" containerID="5854028821e6a360a259427b68e7bb4c2a7220e0c68f4cc91639b2ece0540a02" exitCode=0 Nov 21 14:12:35 crc kubenswrapper[4897]: I1121 14:12:35.112188 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" event={"ID":"e7670227-d280-4847-b882-754429f56b0f","Type":"ContainerDied","Data":"5854028821e6a360a259427b68e7bb4c2a7220e0c68f4cc91639b2ece0540a02"} Nov 21 14:12:36 crc kubenswrapper[4897]: I1121 14:12:36.463393 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jf4pv" Nov 21 14:12:36 crc kubenswrapper[4897]: I1121 14:12:36.628476 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ea4ecd1-6706-4a89-8034-4f41ee7e0f2d-utilities\") pod \"5ea4ecd1-6706-4a89-8034-4f41ee7e0f2d\" (UID: \"5ea4ecd1-6706-4a89-8034-4f41ee7e0f2d\") " Nov 21 14:12:36 crc kubenswrapper[4897]: I1121 14:12:36.628579 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ea4ecd1-6706-4a89-8034-4f41ee7e0f2d-catalog-content\") pod \"5ea4ecd1-6706-4a89-8034-4f41ee7e0f2d\" (UID: \"5ea4ecd1-6706-4a89-8034-4f41ee7e0f2d\") " Nov 21 14:12:36 crc kubenswrapper[4897]: I1121 14:12:36.628630 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dt6q8\" (UniqueName: \"kubernetes.io/projected/5ea4ecd1-6706-4a89-8034-4f41ee7e0f2d-kube-api-access-dt6q8\") pod \"5ea4ecd1-6706-4a89-8034-4f41ee7e0f2d\" (UID: \"5ea4ecd1-6706-4a89-8034-4f41ee7e0f2d\") " Nov 21 14:12:36 crc kubenswrapper[4897]: I1121 14:12:36.629569 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ea4ecd1-6706-4a89-8034-4f41ee7e0f2d-utilities" (OuterVolumeSpecName: "utilities") pod "5ea4ecd1-6706-4a89-8034-4f41ee7e0f2d" (UID: "5ea4ecd1-6706-4a89-8034-4f41ee7e0f2d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:12:36 crc kubenswrapper[4897]: I1121 14:12:36.635615 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ea4ecd1-6706-4a89-8034-4f41ee7e0f2d-kube-api-access-dt6q8" (OuterVolumeSpecName: "kube-api-access-dt6q8") pod "5ea4ecd1-6706-4a89-8034-4f41ee7e0f2d" (UID: "5ea4ecd1-6706-4a89-8034-4f41ee7e0f2d"). InnerVolumeSpecName "kube-api-access-dt6q8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:12:36 crc kubenswrapper[4897]: I1121 14:12:36.677578 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ea4ecd1-6706-4a89-8034-4f41ee7e0f2d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5ea4ecd1-6706-4a89-8034-4f41ee7e0f2d" (UID: "5ea4ecd1-6706-4a89-8034-4f41ee7e0f2d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:12:36 crc kubenswrapper[4897]: I1121 14:12:36.729829 4897 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ea4ecd1-6706-4a89-8034-4f41ee7e0f2d-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 21 14:12:36 crc kubenswrapper[4897]: I1121 14:12:36.729872 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dt6q8\" (UniqueName: \"kubernetes.io/projected/5ea4ecd1-6706-4a89-8034-4f41ee7e0f2d-kube-api-access-dt6q8\") on node \"crc\" DevicePath \"\"" Nov 21 14:12:36 crc kubenswrapper[4897]: I1121 14:12:36.729885 4897 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ea4ecd1-6706-4a89-8034-4f41ee7e0f2d-utilities\") on node \"crc\" DevicePath \"\"" Nov 21 14:12:37 crc kubenswrapper[4897]: I1121 14:12:37.127094 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jf4pv" event={"ID":"5ea4ecd1-6706-4a89-8034-4f41ee7e0f2d","Type":"ContainerDied","Data":"8cbae1e2231d99a0e5a5c8cba032c583df8388b21e53a8a393e3e6089cac6100"} Nov 21 14:12:37 crc kubenswrapper[4897]: I1121 14:12:37.127119 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jf4pv" Nov 21 14:12:37 crc kubenswrapper[4897]: I1121 14:12:37.127167 4897 scope.go:117] "RemoveContainer" containerID="0222fc2020cd78e6f8fa1b4c2a4848439fcc1dc12c9df77757977aa632c1321c" Nov 21 14:12:37 crc kubenswrapper[4897]: I1121 14:12:37.136392 4897 generic.go:334] "Generic (PLEG): container finished" podID="a55d260f-3da6-4b30-b633-6348b31b0b01" containerID="3c0ec7156094b8fb5cef911ec29a478935c769b5835bc31296352806da3a80da" exitCode=0 Nov 21 14:12:37 crc kubenswrapper[4897]: I1121 14:12:37.136424 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k2ctv" event={"ID":"a55d260f-3da6-4b30-b633-6348b31b0b01","Type":"ContainerDied","Data":"3c0ec7156094b8fb5cef911ec29a478935c769b5835bc31296352806da3a80da"} Nov 21 14:12:37 crc kubenswrapper[4897]: I1121 14:12:37.144066 4897 generic.go:334] "Generic (PLEG): container finished" podID="afe9da5b-bdfb-47c5-bd31-abc7c0d53b63" containerID="257c4d05215d7d7933bbf99580b0eed8328a3e0139545682a4ff31ae300e004d" exitCode=0 Nov 21 14:12:37 crc kubenswrapper[4897]: I1121 14:12:37.144137 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5v5dq" event={"ID":"afe9da5b-bdfb-47c5-bd31-abc7c0d53b63","Type":"ContainerDied","Data":"257c4d05215d7d7933bbf99580b0eed8328a3e0139545682a4ff31ae300e004d"} Nov 21 14:12:37 crc kubenswrapper[4897]: I1121 14:12:37.147530 4897 generic.go:334] "Generic (PLEG): container finished" podID="1dc1b877-a8de-45b0-bab4-4411ecced04d" containerID="3c4b45b28a007a59596ee574c1cced8c1fe7d9eb0b178c0ca7e9f28ee9f3b830" exitCode=0 Nov 21 14:12:37 crc kubenswrapper[4897]: I1121 14:12:37.147566 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q6dd2" event={"ID":"1dc1b877-a8de-45b0-bab4-4411ecced04d","Type":"ContainerDied","Data":"3c4b45b28a007a59596ee574c1cced8c1fe7d9eb0b178c0ca7e9f28ee9f3b830"} Nov 21 14:12:37 crc kubenswrapper[4897]: I1121 14:12:37.156569 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" event={"ID":"e7670227-d280-4847-b882-754429f56b0f","Type":"ContainerStarted","Data":"d1709941d05f61c782c0e1fa4c44808bcba5b532f6ab33b1f154eb915fe3841b"} Nov 21 14:12:37 crc kubenswrapper[4897]: I1121 14:12:37.158420 4897 generic.go:334] "Generic (PLEG): container finished" podID="e8ce22b3-b193-4bd0-9a01-9bc260c20430" containerID="f861b14eee0005415bb479efc83429a96c4b9ddb8e266a2228c3576f5b5738b8" exitCode=0 Nov 21 14:12:37 crc kubenswrapper[4897]: I1121 14:12:37.158446 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5wf4b" event={"ID":"e8ce22b3-b193-4bd0-9a01-9bc260c20430","Type":"ContainerDied","Data":"f861b14eee0005415bb479efc83429a96c4b9ddb8e266a2228c3576f5b5738b8"} Nov 21 14:12:37 crc kubenswrapper[4897]: I1121 14:12:37.161789 4897 scope.go:117] "RemoveContainer" containerID="e34bf3228c81cca708e41bf3fc3e1b5ed56fa5fd3e214da58ecf2aa67ec8661f" Nov 21 14:12:37 crc kubenswrapper[4897]: I1121 14:12:37.196560 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jf4pv"] Nov 21 14:12:37 crc kubenswrapper[4897]: I1121 14:12:37.199052 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-jf4pv"] Nov 21 14:12:37 crc kubenswrapper[4897]: I1121 14:12:37.201983 4897 scope.go:117] "RemoveContainer" containerID="874794186ab1f82bbcce7b073013f11d92ada6b7264b7da2db77f59a32599d11" Nov 21 14:12:38 crc kubenswrapper[4897]: I1121 14:12:38.095385 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ea4ecd1-6706-4a89-8034-4f41ee7e0f2d" path="/var/lib/kubelet/pods/5ea4ecd1-6706-4a89-8034-4f41ee7e0f2d/volumes" Nov 21 14:12:38 crc kubenswrapper[4897]: I1121 14:12:38.165113 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5wf4b" event={"ID":"e8ce22b3-b193-4bd0-9a01-9bc260c20430","Type":"ContainerStarted","Data":"099164c17647b70cf212a713bbcf752919b2e93b2c91337f970a23e38ba6480c"} Nov 21 14:12:38 crc kubenswrapper[4897]: I1121 14:12:38.168215 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k2ctv" event={"ID":"a55d260f-3da6-4b30-b633-6348b31b0b01","Type":"ContainerStarted","Data":"74540c8980823b2e6ea29264362d89096ea4d3c71393c4cd0328abfd83026532"} Nov 21 14:12:38 crc kubenswrapper[4897]: I1121 14:12:38.170761 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5v5dq" event={"ID":"afe9da5b-bdfb-47c5-bd31-abc7c0d53b63","Type":"ContainerStarted","Data":"f22c412ec4b894794b89970d21fadd6ead0a12daa8aa2d97e31d51a044f1d382"} Nov 21 14:12:38 crc kubenswrapper[4897]: I1121 14:12:38.173050 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q6dd2" event={"ID":"1dc1b877-a8de-45b0-bab4-4411ecced04d","Type":"ContainerStarted","Data":"57daea7e609a9d7bca2ddc01887c447f5a0488528509239c6a88833905c3e474"} Nov 21 14:12:38 crc kubenswrapper[4897]: I1121 14:12:38.209453 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-5wf4b" podStartSLOduration=4.194867277 podStartE2EDuration="1m19.209434348s" podCreationTimestamp="2025-11-21 14:11:19 +0000 UTC" firstStartedPulling="2025-11-21 14:11:22.604832998 +0000 UTC m=+159.889426473" lastFinishedPulling="2025-11-21 14:12:37.619400069 +0000 UTC m=+234.903993544" observedRunningTime="2025-11-21 14:12:38.190032461 +0000 UTC m=+235.474625936" watchObservedRunningTime="2025-11-21 14:12:38.209434348 +0000 UTC m=+235.494027823" Nov 21 14:12:38 crc kubenswrapper[4897]: I1121 14:12:38.209696 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-5v5dq" podStartSLOduration=3.190229651 podStartE2EDuration="1m16.209692695s" podCreationTimestamp="2025-11-21 14:11:22 +0000 UTC" firstStartedPulling="2025-11-21 14:11:24.666364236 +0000 UTC m=+161.950957701" lastFinishedPulling="2025-11-21 14:12:37.68582727 +0000 UTC m=+234.970420745" observedRunningTime="2025-11-21 14:12:38.207533193 +0000 UTC m=+235.492126678" watchObservedRunningTime="2025-11-21 14:12:38.209692695 +0000 UTC m=+235.494286170" Nov 21 14:12:38 crc kubenswrapper[4897]: I1121 14:12:38.228351 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-k2ctv" podStartSLOduration=3.3070484479999998 podStartE2EDuration="1m17.22833624s" podCreationTimestamp="2025-11-21 14:11:21 +0000 UTC" firstStartedPulling="2025-11-21 14:11:23.625073944 +0000 UTC m=+160.909667419" lastFinishedPulling="2025-11-21 14:12:37.546361736 +0000 UTC m=+234.830955211" observedRunningTime="2025-11-21 14:12:38.226407204 +0000 UTC m=+235.511000699" watchObservedRunningTime="2025-11-21 14:12:38.22833624 +0000 UTC m=+235.512929715" Nov 21 14:12:39 crc kubenswrapper[4897]: I1121 14:12:39.645340 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-q6dd2" Nov 21 14:12:39 crc kubenswrapper[4897]: I1121 14:12:39.645380 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-q6dd2" Nov 21 14:12:39 crc kubenswrapper[4897]: I1121 14:12:39.686807 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-q6dd2" Nov 21 14:12:39 crc kubenswrapper[4897]: I1121 14:12:39.702392 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-q6dd2" podStartSLOduration=4.631915034 podStartE2EDuration="1m20.702375007s" podCreationTimestamp="2025-11-21 14:11:19 +0000 UTC" firstStartedPulling="2025-11-21 14:11:21.524588434 +0000 UTC m=+158.809181909" lastFinishedPulling="2025-11-21 14:12:37.595048407 +0000 UTC m=+234.879641882" observedRunningTime="2025-11-21 14:12:38.249342464 +0000 UTC m=+235.533935949" watchObservedRunningTime="2025-11-21 14:12:39.702375007 +0000 UTC m=+236.986968472" Nov 21 14:12:40 crc kubenswrapper[4897]: I1121 14:12:40.059533 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-5wf4b" Nov 21 14:12:40 crc kubenswrapper[4897]: I1121 14:12:40.059587 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-5wf4b" Nov 21 14:12:40 crc kubenswrapper[4897]: I1121 14:12:40.108558 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-5wf4b" Nov 21 14:12:42 crc kubenswrapper[4897]: I1121 14:12:42.222078 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-k2ctv" Nov 21 14:12:42 crc kubenswrapper[4897]: I1121 14:12:42.222994 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-k2ctv" Nov 21 14:12:42 crc kubenswrapper[4897]: I1121 14:12:42.259387 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-k2ctv" Nov 21 14:12:43 crc kubenswrapper[4897]: I1121 14:12:43.239206 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-5v5dq" Nov 21 14:12:43 crc kubenswrapper[4897]: I1121 14:12:43.239332 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-5v5dq" Nov 21 14:12:43 crc kubenswrapper[4897]: I1121 14:12:43.247986 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-k2ctv" Nov 21 14:12:44 crc kubenswrapper[4897]: I1121 14:12:44.274544 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-5v5dq" podUID="afe9da5b-bdfb-47c5-bd31-abc7c0d53b63" containerName="registry-server" probeResult="failure" output=< Nov 21 14:12:44 crc kubenswrapper[4897]: timeout: failed to connect service ":50051" within 1s Nov 21 14:12:44 crc kubenswrapper[4897]: > Nov 21 14:12:44 crc kubenswrapper[4897]: I1121 14:12:44.425997 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-k2ctv"] Nov 21 14:12:45 crc kubenswrapper[4897]: I1121 14:12:45.223191 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-k2ctv" podUID="a55d260f-3da6-4b30-b633-6348b31b0b01" containerName="registry-server" containerID="cri-o://74540c8980823b2e6ea29264362d89096ea4d3c71393c4cd0328abfd83026532" gracePeriod=2 Nov 21 14:12:45 crc kubenswrapper[4897]: I1121 14:12:45.613227 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k2ctv" Nov 21 14:12:45 crc kubenswrapper[4897]: I1121 14:12:45.741931 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jc9hp\" (UniqueName: \"kubernetes.io/projected/a55d260f-3da6-4b30-b633-6348b31b0b01-kube-api-access-jc9hp\") pod \"a55d260f-3da6-4b30-b633-6348b31b0b01\" (UID: \"a55d260f-3da6-4b30-b633-6348b31b0b01\") " Nov 21 14:12:45 crc kubenswrapper[4897]: I1121 14:12:45.742072 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a55d260f-3da6-4b30-b633-6348b31b0b01-catalog-content\") pod \"a55d260f-3da6-4b30-b633-6348b31b0b01\" (UID: \"a55d260f-3da6-4b30-b633-6348b31b0b01\") " Nov 21 14:12:45 crc kubenswrapper[4897]: I1121 14:12:45.742096 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a55d260f-3da6-4b30-b633-6348b31b0b01-utilities\") pod \"a55d260f-3da6-4b30-b633-6348b31b0b01\" (UID: \"a55d260f-3da6-4b30-b633-6348b31b0b01\") " Nov 21 14:12:45 crc kubenswrapper[4897]: I1121 14:12:45.743136 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a55d260f-3da6-4b30-b633-6348b31b0b01-utilities" (OuterVolumeSpecName: "utilities") pod "a55d260f-3da6-4b30-b633-6348b31b0b01" (UID: "a55d260f-3da6-4b30-b633-6348b31b0b01"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:12:45 crc kubenswrapper[4897]: I1121 14:12:45.748336 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a55d260f-3da6-4b30-b633-6348b31b0b01-kube-api-access-jc9hp" (OuterVolumeSpecName: "kube-api-access-jc9hp") pod "a55d260f-3da6-4b30-b633-6348b31b0b01" (UID: "a55d260f-3da6-4b30-b633-6348b31b0b01"). InnerVolumeSpecName "kube-api-access-jc9hp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:12:45 crc kubenswrapper[4897]: I1121 14:12:45.764717 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a55d260f-3da6-4b30-b633-6348b31b0b01-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a55d260f-3da6-4b30-b633-6348b31b0b01" (UID: "a55d260f-3da6-4b30-b633-6348b31b0b01"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:12:45 crc kubenswrapper[4897]: I1121 14:12:45.843215 4897 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a55d260f-3da6-4b30-b633-6348b31b0b01-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 21 14:12:45 crc kubenswrapper[4897]: I1121 14:12:45.843263 4897 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a55d260f-3da6-4b30-b633-6348b31b0b01-utilities\") on node \"crc\" DevicePath \"\"" Nov 21 14:12:45 crc kubenswrapper[4897]: I1121 14:12:45.843273 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jc9hp\" (UniqueName: \"kubernetes.io/projected/a55d260f-3da6-4b30-b633-6348b31b0b01-kube-api-access-jc9hp\") on node \"crc\" DevicePath \"\"" Nov 21 14:12:46 crc kubenswrapper[4897]: I1121 14:12:46.231703 4897 generic.go:334] "Generic (PLEG): container finished" podID="a55d260f-3da6-4b30-b633-6348b31b0b01" containerID="74540c8980823b2e6ea29264362d89096ea4d3c71393c4cd0328abfd83026532" exitCode=0 Nov 21 14:12:46 crc kubenswrapper[4897]: I1121 14:12:46.231770 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k2ctv" Nov 21 14:12:46 crc kubenswrapper[4897]: I1121 14:12:46.231776 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k2ctv" event={"ID":"a55d260f-3da6-4b30-b633-6348b31b0b01","Type":"ContainerDied","Data":"74540c8980823b2e6ea29264362d89096ea4d3c71393c4cd0328abfd83026532"} Nov 21 14:12:46 crc kubenswrapper[4897]: I1121 14:12:46.231839 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k2ctv" event={"ID":"a55d260f-3da6-4b30-b633-6348b31b0b01","Type":"ContainerDied","Data":"a71da170e0a9164c1f62738449a871665e90540e7aa100ae5b2be97288ff7462"} Nov 21 14:12:46 crc kubenswrapper[4897]: I1121 14:12:46.231861 4897 scope.go:117] "RemoveContainer" containerID="74540c8980823b2e6ea29264362d89096ea4d3c71393c4cd0328abfd83026532" Nov 21 14:12:46 crc kubenswrapper[4897]: I1121 14:12:46.251315 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-k2ctv"] Nov 21 14:12:46 crc kubenswrapper[4897]: I1121 14:12:46.256690 4897 scope.go:117] "RemoveContainer" containerID="3c0ec7156094b8fb5cef911ec29a478935c769b5835bc31296352806da3a80da" Nov 21 14:12:46 crc kubenswrapper[4897]: I1121 14:12:46.258435 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-k2ctv"] Nov 21 14:12:46 crc kubenswrapper[4897]: I1121 14:12:46.275319 4897 scope.go:117] "RemoveContainer" containerID="ffedeae3a60a10cd4bb4fa9046a20001bbb9798139976bc047843ed79798324a" Nov 21 14:12:46 crc kubenswrapper[4897]: I1121 14:12:46.290753 4897 scope.go:117] "RemoveContainer" containerID="74540c8980823b2e6ea29264362d89096ea4d3c71393c4cd0328abfd83026532" Nov 21 14:12:46 crc kubenswrapper[4897]: E1121 14:12:46.291361 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"74540c8980823b2e6ea29264362d89096ea4d3c71393c4cd0328abfd83026532\": container with ID starting with 74540c8980823b2e6ea29264362d89096ea4d3c71393c4cd0328abfd83026532 not found: ID does not exist" containerID="74540c8980823b2e6ea29264362d89096ea4d3c71393c4cd0328abfd83026532" Nov 21 14:12:46 crc kubenswrapper[4897]: I1121 14:12:46.291402 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"74540c8980823b2e6ea29264362d89096ea4d3c71393c4cd0328abfd83026532"} err="failed to get container status \"74540c8980823b2e6ea29264362d89096ea4d3c71393c4cd0328abfd83026532\": rpc error: code = NotFound desc = could not find container \"74540c8980823b2e6ea29264362d89096ea4d3c71393c4cd0328abfd83026532\": container with ID starting with 74540c8980823b2e6ea29264362d89096ea4d3c71393c4cd0328abfd83026532 not found: ID does not exist" Nov 21 14:12:46 crc kubenswrapper[4897]: I1121 14:12:46.291468 4897 scope.go:117] "RemoveContainer" containerID="3c0ec7156094b8fb5cef911ec29a478935c769b5835bc31296352806da3a80da" Nov 21 14:12:46 crc kubenswrapper[4897]: E1121 14:12:46.291970 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3c0ec7156094b8fb5cef911ec29a478935c769b5835bc31296352806da3a80da\": container with ID starting with 3c0ec7156094b8fb5cef911ec29a478935c769b5835bc31296352806da3a80da not found: ID does not exist" containerID="3c0ec7156094b8fb5cef911ec29a478935c769b5835bc31296352806da3a80da" Nov 21 14:12:46 crc kubenswrapper[4897]: I1121 14:12:46.292030 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c0ec7156094b8fb5cef911ec29a478935c769b5835bc31296352806da3a80da"} err="failed to get container status \"3c0ec7156094b8fb5cef911ec29a478935c769b5835bc31296352806da3a80da\": rpc error: code = NotFound desc = could not find container \"3c0ec7156094b8fb5cef911ec29a478935c769b5835bc31296352806da3a80da\": container with ID starting with 3c0ec7156094b8fb5cef911ec29a478935c769b5835bc31296352806da3a80da not found: ID does not exist" Nov 21 14:12:46 crc kubenswrapper[4897]: I1121 14:12:46.292047 4897 scope.go:117] "RemoveContainer" containerID="ffedeae3a60a10cd4bb4fa9046a20001bbb9798139976bc047843ed79798324a" Nov 21 14:12:46 crc kubenswrapper[4897]: E1121 14:12:46.292379 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ffedeae3a60a10cd4bb4fa9046a20001bbb9798139976bc047843ed79798324a\": container with ID starting with ffedeae3a60a10cd4bb4fa9046a20001bbb9798139976bc047843ed79798324a not found: ID does not exist" containerID="ffedeae3a60a10cd4bb4fa9046a20001bbb9798139976bc047843ed79798324a" Nov 21 14:12:46 crc kubenswrapper[4897]: I1121 14:12:46.292436 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ffedeae3a60a10cd4bb4fa9046a20001bbb9798139976bc047843ed79798324a"} err="failed to get container status \"ffedeae3a60a10cd4bb4fa9046a20001bbb9798139976bc047843ed79798324a\": rpc error: code = NotFound desc = could not find container \"ffedeae3a60a10cd4bb4fa9046a20001bbb9798139976bc047843ed79798324a\": container with ID starting with ffedeae3a60a10cd4bb4fa9046a20001bbb9798139976bc047843ed79798324a not found: ID does not exist" Nov 21 14:12:48 crc kubenswrapper[4897]: I1121 14:12:48.094543 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a55d260f-3da6-4b30-b633-6348b31b0b01" path="/var/lib/kubelet/pods/a55d260f-3da6-4b30-b633-6348b31b0b01/volumes" Nov 21 14:12:49 crc kubenswrapper[4897]: I1121 14:12:49.690741 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-q6dd2" Nov 21 14:12:50 crc kubenswrapper[4897]: I1121 14:12:50.096382 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-5wf4b" Nov 21 14:12:50 crc kubenswrapper[4897]: I1121 14:12:50.818064 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5wf4b"] Nov 21 14:12:50 crc kubenswrapper[4897]: I1121 14:12:50.818596 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-5wf4b" podUID="e8ce22b3-b193-4bd0-9a01-9bc260c20430" containerName="registry-server" containerID="cri-o://099164c17647b70cf212a713bbcf752919b2e93b2c91337f970a23e38ba6480c" gracePeriod=2 Nov 21 14:12:51 crc kubenswrapper[4897]: I1121 14:12:51.565240 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-2trmp" podUID="0e4b5f84-4954-4d85-8bae-db18ed352464" containerName="oauth-openshift" containerID="cri-o://bb86b29391e447e7a95b1e90e43d0d41b3f73430bf10f101de634d819340dac0" gracePeriod=15 Nov 21 14:12:52 crc kubenswrapper[4897]: I1121 14:12:52.263405 4897 generic.go:334] "Generic (PLEG): container finished" podID="e8ce22b3-b193-4bd0-9a01-9bc260c20430" containerID="099164c17647b70cf212a713bbcf752919b2e93b2c91337f970a23e38ba6480c" exitCode=0 Nov 21 14:12:52 crc kubenswrapper[4897]: I1121 14:12:52.263482 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5wf4b" event={"ID":"e8ce22b3-b193-4bd0-9a01-9bc260c20430","Type":"ContainerDied","Data":"099164c17647b70cf212a713bbcf752919b2e93b2c91337f970a23e38ba6480c"} Nov 21 14:12:52 crc kubenswrapper[4897]: I1121 14:12:52.265198 4897 generic.go:334] "Generic (PLEG): container finished" podID="0e4b5f84-4954-4d85-8bae-db18ed352464" containerID="bb86b29391e447e7a95b1e90e43d0d41b3f73430bf10f101de634d819340dac0" exitCode=0 Nov 21 14:12:52 crc kubenswrapper[4897]: I1121 14:12:52.265251 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-2trmp" event={"ID":"0e4b5f84-4954-4d85-8bae-db18ed352464","Type":"ContainerDied","Data":"bb86b29391e447e7a95b1e90e43d0d41b3f73430bf10f101de634d819340dac0"} Nov 21 14:12:53 crc kubenswrapper[4897]: I1121 14:12:53.256180 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5wf4b" Nov 21 14:12:53 crc kubenswrapper[4897]: I1121 14:12:53.285998 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5wf4b" event={"ID":"e8ce22b3-b193-4bd0-9a01-9bc260c20430","Type":"ContainerDied","Data":"6b156833ac602e98d8284ba6e7e11e9fdae647375b750fec64ad922e98f30bde"} Nov 21 14:12:53 crc kubenswrapper[4897]: I1121 14:12:53.286050 4897 scope.go:117] "RemoveContainer" containerID="099164c17647b70cf212a713bbcf752919b2e93b2c91337f970a23e38ba6480c" Nov 21 14:12:53 crc kubenswrapper[4897]: I1121 14:12:53.286112 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5wf4b" Nov 21 14:12:53 crc kubenswrapper[4897]: I1121 14:12:53.305604 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-5v5dq" Nov 21 14:12:53 crc kubenswrapper[4897]: I1121 14:12:53.322571 4897 scope.go:117] "RemoveContainer" containerID="f861b14eee0005415bb479efc83429a96c4b9ddb8e266a2228c3576f5b5738b8" Nov 21 14:12:53 crc kubenswrapper[4897]: I1121 14:12:53.338901 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-55ctt\" (UniqueName: \"kubernetes.io/projected/e8ce22b3-b193-4bd0-9a01-9bc260c20430-kube-api-access-55ctt\") pod \"e8ce22b3-b193-4bd0-9a01-9bc260c20430\" (UID: \"e8ce22b3-b193-4bd0-9a01-9bc260c20430\") " Nov 21 14:12:53 crc kubenswrapper[4897]: I1121 14:12:53.338992 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8ce22b3-b193-4bd0-9a01-9bc260c20430-catalog-content\") pod \"e8ce22b3-b193-4bd0-9a01-9bc260c20430\" (UID: \"e8ce22b3-b193-4bd0-9a01-9bc260c20430\") " Nov 21 14:12:53 crc kubenswrapper[4897]: I1121 14:12:53.346709 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8ce22b3-b193-4bd0-9a01-9bc260c20430-kube-api-access-55ctt" (OuterVolumeSpecName: "kube-api-access-55ctt") pod "e8ce22b3-b193-4bd0-9a01-9bc260c20430" (UID: "e8ce22b3-b193-4bd0-9a01-9bc260c20430"). InnerVolumeSpecName "kube-api-access-55ctt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:12:53 crc kubenswrapper[4897]: I1121 14:12:53.358234 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-5v5dq" Nov 21 14:12:53 crc kubenswrapper[4897]: I1121 14:12:53.361283 4897 scope.go:117] "RemoveContainer" containerID="4b959a8e218803f046a5afb07863a1662545b93118ba4d8702a7e6b3aaa2551a" Nov 21 14:12:53 crc kubenswrapper[4897]: I1121 14:12:53.390826 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e8ce22b3-b193-4bd0-9a01-9bc260c20430-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e8ce22b3-b193-4bd0-9a01-9bc260c20430" (UID: "e8ce22b3-b193-4bd0-9a01-9bc260c20430"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:12:53 crc kubenswrapper[4897]: I1121 14:12:53.396257 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-2trmp" Nov 21 14:12:53 crc kubenswrapper[4897]: I1121 14:12:53.440652 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8ce22b3-b193-4bd0-9a01-9bc260c20430-utilities\") pod \"e8ce22b3-b193-4bd0-9a01-9bc260c20430\" (UID: \"e8ce22b3-b193-4bd0-9a01-9bc260c20430\") " Nov 21 14:12:53 crc kubenswrapper[4897]: I1121 14:12:53.441383 4897 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8ce22b3-b193-4bd0-9a01-9bc260c20430-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 21 14:12:53 crc kubenswrapper[4897]: I1121 14:12:53.441395 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e8ce22b3-b193-4bd0-9a01-9bc260c20430-utilities" (OuterVolumeSpecName: "utilities") pod "e8ce22b3-b193-4bd0-9a01-9bc260c20430" (UID: "e8ce22b3-b193-4bd0-9a01-9bc260c20430"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:12:53 crc kubenswrapper[4897]: I1121 14:12:53.441407 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-55ctt\" (UniqueName: \"kubernetes.io/projected/e8ce22b3-b193-4bd0-9a01-9bc260c20430-kube-api-access-55ctt\") on node \"crc\" DevicePath \"\"" Nov 21 14:12:53 crc kubenswrapper[4897]: I1121 14:12:53.542845 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/0e4b5f84-4954-4d85-8bae-db18ed352464-v4-0-config-user-idp-0-file-data\") pod \"0e4b5f84-4954-4d85-8bae-db18ed352464\" (UID: \"0e4b5f84-4954-4d85-8bae-db18ed352464\") " Nov 21 14:12:53 crc kubenswrapper[4897]: I1121 14:12:53.542968 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/0e4b5f84-4954-4d85-8bae-db18ed352464-v4-0-config-system-ocp-branding-template\") pod \"0e4b5f84-4954-4d85-8bae-db18ed352464\" (UID: \"0e4b5f84-4954-4d85-8bae-db18ed352464\") " Nov 21 14:12:53 crc kubenswrapper[4897]: I1121 14:12:53.543019 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/0e4b5f84-4954-4d85-8bae-db18ed352464-v4-0-config-user-template-login\") pod \"0e4b5f84-4954-4d85-8bae-db18ed352464\" (UID: \"0e4b5f84-4954-4d85-8bae-db18ed352464\") " Nov 21 14:12:53 crc kubenswrapper[4897]: I1121 14:12:53.543053 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/0e4b5f84-4954-4d85-8bae-db18ed352464-v4-0-config-user-template-provider-selection\") pod \"0e4b5f84-4954-4d85-8bae-db18ed352464\" (UID: \"0e4b5f84-4954-4d85-8bae-db18ed352464\") " Nov 21 14:12:53 crc kubenswrapper[4897]: I1121 14:12:53.543116 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/0e4b5f84-4954-4d85-8bae-db18ed352464-v4-0-config-system-router-certs\") pod \"0e4b5f84-4954-4d85-8bae-db18ed352464\" (UID: \"0e4b5f84-4954-4d85-8bae-db18ed352464\") " Nov 21 14:12:53 crc kubenswrapper[4897]: I1121 14:12:53.543355 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/0e4b5f84-4954-4d85-8bae-db18ed352464-v4-0-config-system-session\") pod \"0e4b5f84-4954-4d85-8bae-db18ed352464\" (UID: \"0e4b5f84-4954-4d85-8bae-db18ed352464\") " Nov 21 14:12:53 crc kubenswrapper[4897]: I1121 14:12:53.543409 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/0e4b5f84-4954-4d85-8bae-db18ed352464-v4-0-config-user-template-error\") pod \"0e4b5f84-4954-4d85-8bae-db18ed352464\" (UID: \"0e4b5f84-4954-4d85-8bae-db18ed352464\") " Nov 21 14:12:53 crc kubenswrapper[4897]: I1121 14:12:53.543445 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0e4b5f84-4954-4d85-8bae-db18ed352464-v4-0-config-system-trusted-ca-bundle\") pod \"0e4b5f84-4954-4d85-8bae-db18ed352464\" (UID: \"0e4b5f84-4954-4d85-8bae-db18ed352464\") " Nov 21 14:12:53 crc kubenswrapper[4897]: I1121 14:12:53.543655 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/0e4b5f84-4954-4d85-8bae-db18ed352464-v4-0-config-system-service-ca\") pod \"0e4b5f84-4954-4d85-8bae-db18ed352464\" (UID: \"0e4b5f84-4954-4d85-8bae-db18ed352464\") " Nov 21 14:12:53 crc kubenswrapper[4897]: I1121 14:12:53.543727 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/0e4b5f84-4954-4d85-8bae-db18ed352464-v4-0-config-system-serving-cert\") pod \"0e4b5f84-4954-4d85-8bae-db18ed352464\" (UID: \"0e4b5f84-4954-4d85-8bae-db18ed352464\") " Nov 21 14:12:53 crc kubenswrapper[4897]: I1121 14:12:53.543809 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0e4b5f84-4954-4d85-8bae-db18ed352464-audit-dir\") pod \"0e4b5f84-4954-4d85-8bae-db18ed352464\" (UID: \"0e4b5f84-4954-4d85-8bae-db18ed352464\") " Nov 21 14:12:53 crc kubenswrapper[4897]: I1121 14:12:53.543863 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/0e4b5f84-4954-4d85-8bae-db18ed352464-v4-0-config-system-cliconfig\") pod \"0e4b5f84-4954-4d85-8bae-db18ed352464\" (UID: \"0e4b5f84-4954-4d85-8bae-db18ed352464\") " Nov 21 14:12:53 crc kubenswrapper[4897]: I1121 14:12:53.543910 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0e4b5f84-4954-4d85-8bae-db18ed352464-audit-policies\") pod \"0e4b5f84-4954-4d85-8bae-db18ed352464\" (UID: \"0e4b5f84-4954-4d85-8bae-db18ed352464\") " Nov 21 14:12:53 crc kubenswrapper[4897]: I1121 14:12:53.543942 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0e4b5f84-4954-4d85-8bae-db18ed352464-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "0e4b5f84-4954-4d85-8bae-db18ed352464" (UID: "0e4b5f84-4954-4d85-8bae-db18ed352464"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 21 14:12:53 crc kubenswrapper[4897]: I1121 14:12:53.543966 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vmhkr\" (UniqueName: \"kubernetes.io/projected/0e4b5f84-4954-4d85-8bae-db18ed352464-kube-api-access-vmhkr\") pod \"0e4b5f84-4954-4d85-8bae-db18ed352464\" (UID: \"0e4b5f84-4954-4d85-8bae-db18ed352464\") " Nov 21 14:12:53 crc kubenswrapper[4897]: I1121 14:12:53.544738 4897 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0e4b5f84-4954-4d85-8bae-db18ed352464-audit-dir\") on node \"crc\" DevicePath \"\"" Nov 21 14:12:53 crc kubenswrapper[4897]: I1121 14:12:53.544767 4897 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8ce22b3-b193-4bd0-9a01-9bc260c20430-utilities\") on node \"crc\" DevicePath \"\"" Nov 21 14:12:53 crc kubenswrapper[4897]: I1121 14:12:53.544857 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0e4b5f84-4954-4d85-8bae-db18ed352464-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "0e4b5f84-4954-4d85-8bae-db18ed352464" (UID: "0e4b5f84-4954-4d85-8bae-db18ed352464"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:12:53 crc kubenswrapper[4897]: I1121 14:12:53.545292 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0e4b5f84-4954-4d85-8bae-db18ed352464-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "0e4b5f84-4954-4d85-8bae-db18ed352464" (UID: "0e4b5f84-4954-4d85-8bae-db18ed352464"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:12:53 crc kubenswrapper[4897]: I1121 14:12:53.545324 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0e4b5f84-4954-4d85-8bae-db18ed352464-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "0e4b5f84-4954-4d85-8bae-db18ed352464" (UID: "0e4b5f84-4954-4d85-8bae-db18ed352464"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:12:53 crc kubenswrapper[4897]: I1121 14:12:53.545469 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0e4b5f84-4954-4d85-8bae-db18ed352464-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "0e4b5f84-4954-4d85-8bae-db18ed352464" (UID: "0e4b5f84-4954-4d85-8bae-db18ed352464"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:12:53 crc kubenswrapper[4897]: I1121 14:12:53.547356 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e4b5f84-4954-4d85-8bae-db18ed352464-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "0e4b5f84-4954-4d85-8bae-db18ed352464" (UID: "0e4b5f84-4954-4d85-8bae-db18ed352464"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:12:53 crc kubenswrapper[4897]: I1121 14:12:53.547671 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e4b5f84-4954-4d85-8bae-db18ed352464-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "0e4b5f84-4954-4d85-8bae-db18ed352464" (UID: "0e4b5f84-4954-4d85-8bae-db18ed352464"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:12:53 crc kubenswrapper[4897]: I1121 14:12:53.547973 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e4b5f84-4954-4d85-8bae-db18ed352464-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "0e4b5f84-4954-4d85-8bae-db18ed352464" (UID: "0e4b5f84-4954-4d85-8bae-db18ed352464"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:12:53 crc kubenswrapper[4897]: I1121 14:12:53.547975 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e4b5f84-4954-4d85-8bae-db18ed352464-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "0e4b5f84-4954-4d85-8bae-db18ed352464" (UID: "0e4b5f84-4954-4d85-8bae-db18ed352464"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:12:53 crc kubenswrapper[4897]: I1121 14:12:53.548088 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e4b5f84-4954-4d85-8bae-db18ed352464-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "0e4b5f84-4954-4d85-8bae-db18ed352464" (UID: "0e4b5f84-4954-4d85-8bae-db18ed352464"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:12:53 crc kubenswrapper[4897]: I1121 14:12:53.548355 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e4b5f84-4954-4d85-8bae-db18ed352464-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "0e4b5f84-4954-4d85-8bae-db18ed352464" (UID: "0e4b5f84-4954-4d85-8bae-db18ed352464"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:12:53 crc kubenswrapper[4897]: I1121 14:12:53.548455 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e4b5f84-4954-4d85-8bae-db18ed352464-kube-api-access-vmhkr" (OuterVolumeSpecName: "kube-api-access-vmhkr") pod "0e4b5f84-4954-4d85-8bae-db18ed352464" (UID: "0e4b5f84-4954-4d85-8bae-db18ed352464"). InnerVolumeSpecName "kube-api-access-vmhkr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:12:53 crc kubenswrapper[4897]: I1121 14:12:53.548471 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e4b5f84-4954-4d85-8bae-db18ed352464-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "0e4b5f84-4954-4d85-8bae-db18ed352464" (UID: "0e4b5f84-4954-4d85-8bae-db18ed352464"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:12:53 crc kubenswrapper[4897]: I1121 14:12:53.549606 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e4b5f84-4954-4d85-8bae-db18ed352464-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "0e4b5f84-4954-4d85-8bae-db18ed352464" (UID: "0e4b5f84-4954-4d85-8bae-db18ed352464"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:12:53 crc kubenswrapper[4897]: I1121 14:12:53.616474 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5wf4b"] Nov 21 14:12:53 crc kubenswrapper[4897]: I1121 14:12:53.622029 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-5wf4b"] Nov 21 14:12:53 crc kubenswrapper[4897]: I1121 14:12:53.646198 4897 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0e4b5f84-4954-4d85-8bae-db18ed352464-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 21 14:12:53 crc kubenswrapper[4897]: I1121 14:12:53.646238 4897 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/0e4b5f84-4954-4d85-8bae-db18ed352464-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Nov 21 14:12:53 crc kubenswrapper[4897]: I1121 14:12:53.646253 4897 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/0e4b5f84-4954-4d85-8bae-db18ed352464-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Nov 21 14:12:53 crc kubenswrapper[4897]: I1121 14:12:53.646265 4897 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/0e4b5f84-4954-4d85-8bae-db18ed352464-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 21 14:12:53 crc kubenswrapper[4897]: I1121 14:12:53.646280 4897 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/0e4b5f84-4954-4d85-8bae-db18ed352464-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Nov 21 14:12:53 crc kubenswrapper[4897]: I1121 14:12:53.646292 4897 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0e4b5f84-4954-4d85-8bae-db18ed352464-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 21 14:12:53 crc kubenswrapper[4897]: I1121 14:12:53.646305 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vmhkr\" (UniqueName: \"kubernetes.io/projected/0e4b5f84-4954-4d85-8bae-db18ed352464-kube-api-access-vmhkr\") on node \"crc\" DevicePath \"\"" Nov 21 14:12:53 crc kubenswrapper[4897]: I1121 14:12:53.646317 4897 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/0e4b5f84-4954-4d85-8bae-db18ed352464-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Nov 21 14:12:53 crc kubenswrapper[4897]: I1121 14:12:53.646332 4897 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/0e4b5f84-4954-4d85-8bae-db18ed352464-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Nov 21 14:12:53 crc kubenswrapper[4897]: I1121 14:12:53.646346 4897 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/0e4b5f84-4954-4d85-8bae-db18ed352464-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Nov 21 14:12:53 crc kubenswrapper[4897]: I1121 14:12:53.646359 4897 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/0e4b5f84-4954-4d85-8bae-db18ed352464-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Nov 21 14:12:53 crc kubenswrapper[4897]: I1121 14:12:53.646372 4897 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/0e4b5f84-4954-4d85-8bae-db18ed352464-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Nov 21 14:12:53 crc kubenswrapper[4897]: I1121 14:12:53.646388 4897 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/0e4b5f84-4954-4d85-8bae-db18ed352464-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Nov 21 14:12:54 crc kubenswrapper[4897]: I1121 14:12:54.094691 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e8ce22b3-b193-4bd0-9a01-9bc260c20430" path="/var/lib/kubelet/pods/e8ce22b3-b193-4bd0-9a01-9bc260c20430/volumes" Nov 21 14:12:54 crc kubenswrapper[4897]: I1121 14:12:54.291444 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-2trmp" Nov 21 14:12:54 crc kubenswrapper[4897]: I1121 14:12:54.291530 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-2trmp" event={"ID":"0e4b5f84-4954-4d85-8bae-db18ed352464","Type":"ContainerDied","Data":"6476da7699e9ba7191eaa620c6d3b896c8ad8bc1907d7c323f28d02aa9d040ae"} Nov 21 14:12:54 crc kubenswrapper[4897]: I1121 14:12:54.291566 4897 scope.go:117] "RemoveContainer" containerID="bb86b29391e447e7a95b1e90e43d0d41b3f73430bf10f101de634d819340dac0" Nov 21 14:12:54 crc kubenswrapper[4897]: I1121 14:12:54.317067 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-2trmp"] Nov 21 14:12:54 crc kubenswrapper[4897]: I1121 14:12:54.320026 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-2trmp"] Nov 21 14:12:55 crc kubenswrapper[4897]: I1121 14:12:55.623442 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5v5dq"] Nov 21 14:12:55 crc kubenswrapper[4897]: I1121 14:12:55.623809 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-5v5dq" podUID="afe9da5b-bdfb-47c5-bd31-abc7c0d53b63" containerName="registry-server" containerID="cri-o://f22c412ec4b894794b89970d21fadd6ead0a12daa8aa2d97e31d51a044f1d382" gracePeriod=2 Nov 21 14:12:56 crc kubenswrapper[4897]: I1121 14:12:56.104925 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0e4b5f84-4954-4d85-8bae-db18ed352464" path="/var/lib/kubelet/pods/0e4b5f84-4954-4d85-8bae-db18ed352464/volumes" Nov 21 14:12:56 crc kubenswrapper[4897]: I1121 14:12:56.310126 4897 generic.go:334] "Generic (PLEG): container finished" podID="afe9da5b-bdfb-47c5-bd31-abc7c0d53b63" containerID="f22c412ec4b894794b89970d21fadd6ead0a12daa8aa2d97e31d51a044f1d382" exitCode=0 Nov 21 14:12:56 crc kubenswrapper[4897]: I1121 14:12:56.310185 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5v5dq" event={"ID":"afe9da5b-bdfb-47c5-bd31-abc7c0d53b63","Type":"ContainerDied","Data":"f22c412ec4b894794b89970d21fadd6ead0a12daa8aa2d97e31d51a044f1d382"} Nov 21 14:12:56 crc kubenswrapper[4897]: I1121 14:12:56.630822 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5v5dq" Nov 21 14:12:56 crc kubenswrapper[4897]: I1121 14:12:56.782985 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/afe9da5b-bdfb-47c5-bd31-abc7c0d53b63-utilities\") pod \"afe9da5b-bdfb-47c5-bd31-abc7c0d53b63\" (UID: \"afe9da5b-bdfb-47c5-bd31-abc7c0d53b63\") " Nov 21 14:12:56 crc kubenswrapper[4897]: I1121 14:12:56.783153 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/afe9da5b-bdfb-47c5-bd31-abc7c0d53b63-catalog-content\") pod \"afe9da5b-bdfb-47c5-bd31-abc7c0d53b63\" (UID: \"afe9da5b-bdfb-47c5-bd31-abc7c0d53b63\") " Nov 21 14:12:56 crc kubenswrapper[4897]: I1121 14:12:56.783248 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7pnxh\" (UniqueName: \"kubernetes.io/projected/afe9da5b-bdfb-47c5-bd31-abc7c0d53b63-kube-api-access-7pnxh\") pod \"afe9da5b-bdfb-47c5-bd31-abc7c0d53b63\" (UID: \"afe9da5b-bdfb-47c5-bd31-abc7c0d53b63\") " Nov 21 14:12:56 crc kubenswrapper[4897]: I1121 14:12:56.784186 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/afe9da5b-bdfb-47c5-bd31-abc7c0d53b63-utilities" (OuterVolumeSpecName: "utilities") pod "afe9da5b-bdfb-47c5-bd31-abc7c0d53b63" (UID: "afe9da5b-bdfb-47c5-bd31-abc7c0d53b63"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:12:56 crc kubenswrapper[4897]: I1121 14:12:56.789150 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/afe9da5b-bdfb-47c5-bd31-abc7c0d53b63-kube-api-access-7pnxh" (OuterVolumeSpecName: "kube-api-access-7pnxh") pod "afe9da5b-bdfb-47c5-bd31-abc7c0d53b63" (UID: "afe9da5b-bdfb-47c5-bd31-abc7c0d53b63"). InnerVolumeSpecName "kube-api-access-7pnxh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:12:56 crc kubenswrapper[4897]: I1121 14:12:56.874135 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/afe9da5b-bdfb-47c5-bd31-abc7c0d53b63-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "afe9da5b-bdfb-47c5-bd31-abc7c0d53b63" (UID: "afe9da5b-bdfb-47c5-bd31-abc7c0d53b63"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:12:56 crc kubenswrapper[4897]: I1121 14:12:56.885919 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7pnxh\" (UniqueName: \"kubernetes.io/projected/afe9da5b-bdfb-47c5-bd31-abc7c0d53b63-kube-api-access-7pnxh\") on node \"crc\" DevicePath \"\"" Nov 21 14:12:56 crc kubenswrapper[4897]: I1121 14:12:56.886004 4897 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/afe9da5b-bdfb-47c5-bd31-abc7c0d53b63-utilities\") on node \"crc\" DevicePath \"\"" Nov 21 14:12:56 crc kubenswrapper[4897]: I1121 14:12:56.886027 4897 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/afe9da5b-bdfb-47c5-bd31-abc7c0d53b63-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 21 14:12:57 crc kubenswrapper[4897]: I1121 14:12:57.321549 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5v5dq" event={"ID":"afe9da5b-bdfb-47c5-bd31-abc7c0d53b63","Type":"ContainerDied","Data":"eb3afaa57ab38165909faf494ca5d1fd2a75ce94bfe1a4e90fbfe0eb1dff17f0"} Nov 21 14:12:57 crc kubenswrapper[4897]: I1121 14:12:57.321647 4897 scope.go:117] "RemoveContainer" containerID="f22c412ec4b894794b89970d21fadd6ead0a12daa8aa2d97e31d51a044f1d382" Nov 21 14:12:57 crc kubenswrapper[4897]: I1121 14:12:57.321579 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5v5dq" Nov 21 14:12:57 crc kubenswrapper[4897]: I1121 14:12:57.357652 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5v5dq"] Nov 21 14:12:57 crc kubenswrapper[4897]: I1121 14:12:57.358591 4897 scope.go:117] "RemoveContainer" containerID="257c4d05215d7d7933bbf99580b0eed8328a3e0139545682a4ff31ae300e004d" Nov 21 14:12:57 crc kubenswrapper[4897]: I1121 14:12:57.362547 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-5v5dq"] Nov 21 14:12:57 crc kubenswrapper[4897]: I1121 14:12:57.393348 4897 scope.go:117] "RemoveContainer" containerID="8b359f86d7e327a550d9c0e7965014564aafab42158f162dfa81b4054c08b683" Nov 21 14:12:58 crc kubenswrapper[4897]: I1121 14:12:58.098836 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="afe9da5b-bdfb-47c5-bd31-abc7c0d53b63" path="/var/lib/kubelet/pods/afe9da5b-bdfb-47c5-bd31-abc7c0d53b63/volumes" Nov 21 14:13:01 crc kubenswrapper[4897]: I1121 14:13:01.653647 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-57bcd9fbb-mlnrl"] Nov 21 14:13:01 crc kubenswrapper[4897]: E1121 14:13:01.654617 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afe9da5b-bdfb-47c5-bd31-abc7c0d53b63" containerName="extract-utilities" Nov 21 14:13:01 crc kubenswrapper[4897]: I1121 14:13:01.654645 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="afe9da5b-bdfb-47c5-bd31-abc7c0d53b63" containerName="extract-utilities" Nov 21 14:13:01 crc kubenswrapper[4897]: E1121 14:13:01.654661 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8ce22b3-b193-4bd0-9a01-9bc260c20430" containerName="registry-server" Nov 21 14:13:01 crc kubenswrapper[4897]: I1121 14:13:01.654673 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8ce22b3-b193-4bd0-9a01-9bc260c20430" containerName="registry-server" Nov 21 14:13:01 crc kubenswrapper[4897]: E1121 14:13:01.654689 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ea4ecd1-6706-4a89-8034-4f41ee7e0f2d" containerName="extract-utilities" Nov 21 14:13:01 crc kubenswrapper[4897]: I1121 14:13:01.654700 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ea4ecd1-6706-4a89-8034-4f41ee7e0f2d" containerName="extract-utilities" Nov 21 14:13:01 crc kubenswrapper[4897]: E1121 14:13:01.654734 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afe9da5b-bdfb-47c5-bd31-abc7c0d53b63" containerName="extract-content" Nov 21 14:13:01 crc kubenswrapper[4897]: I1121 14:13:01.654745 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="afe9da5b-bdfb-47c5-bd31-abc7c0d53b63" containerName="extract-content" Nov 21 14:13:01 crc kubenswrapper[4897]: E1121 14:13:01.654760 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a55d260f-3da6-4b30-b633-6348b31b0b01" containerName="extract-content" Nov 21 14:13:01 crc kubenswrapper[4897]: I1121 14:13:01.654770 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="a55d260f-3da6-4b30-b633-6348b31b0b01" containerName="extract-content" Nov 21 14:13:01 crc kubenswrapper[4897]: E1121 14:13:01.654786 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afe9da5b-bdfb-47c5-bd31-abc7c0d53b63" containerName="registry-server" Nov 21 14:13:01 crc kubenswrapper[4897]: I1121 14:13:01.654801 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="afe9da5b-bdfb-47c5-bd31-abc7c0d53b63" containerName="registry-server" Nov 21 14:13:01 crc kubenswrapper[4897]: E1121 14:13:01.654814 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8ce22b3-b193-4bd0-9a01-9bc260c20430" containerName="extract-content" Nov 21 14:13:01 crc kubenswrapper[4897]: I1121 14:13:01.654825 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8ce22b3-b193-4bd0-9a01-9bc260c20430" containerName="extract-content" Nov 21 14:13:01 crc kubenswrapper[4897]: E1121 14:13:01.654836 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e4b5f84-4954-4d85-8bae-db18ed352464" containerName="oauth-openshift" Nov 21 14:13:01 crc kubenswrapper[4897]: I1121 14:13:01.654846 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e4b5f84-4954-4d85-8bae-db18ed352464" containerName="oauth-openshift" Nov 21 14:13:01 crc kubenswrapper[4897]: E1121 14:13:01.654861 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7cd98bfc-77a5-4ef5-86f1-7614b0b2ad2a" containerName="pruner" Nov 21 14:13:01 crc kubenswrapper[4897]: I1121 14:13:01.654871 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="7cd98bfc-77a5-4ef5-86f1-7614b0b2ad2a" containerName="pruner" Nov 21 14:13:01 crc kubenswrapper[4897]: E1121 14:13:01.654885 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5aa94f1-611d-4695-a3c7-1c33c80f1a1a" containerName="pruner" Nov 21 14:13:01 crc kubenswrapper[4897]: I1121 14:13:01.654893 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5aa94f1-611d-4695-a3c7-1c33c80f1a1a" containerName="pruner" Nov 21 14:13:01 crc kubenswrapper[4897]: E1121 14:13:01.654903 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a55d260f-3da6-4b30-b633-6348b31b0b01" containerName="registry-server" Nov 21 14:13:01 crc kubenswrapper[4897]: I1121 14:13:01.654911 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="a55d260f-3da6-4b30-b633-6348b31b0b01" containerName="registry-server" Nov 21 14:13:01 crc kubenswrapper[4897]: E1121 14:13:01.654921 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ea4ecd1-6706-4a89-8034-4f41ee7e0f2d" containerName="extract-content" Nov 21 14:13:01 crc kubenswrapper[4897]: I1121 14:13:01.654929 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ea4ecd1-6706-4a89-8034-4f41ee7e0f2d" containerName="extract-content" Nov 21 14:13:01 crc kubenswrapper[4897]: E1121 14:13:01.654940 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8ce22b3-b193-4bd0-9a01-9bc260c20430" containerName="extract-utilities" Nov 21 14:13:01 crc kubenswrapper[4897]: I1121 14:13:01.654948 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8ce22b3-b193-4bd0-9a01-9bc260c20430" containerName="extract-utilities" Nov 21 14:13:01 crc kubenswrapper[4897]: E1121 14:13:01.654960 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a55d260f-3da6-4b30-b633-6348b31b0b01" containerName="extract-utilities" Nov 21 14:13:01 crc kubenswrapper[4897]: I1121 14:13:01.654968 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="a55d260f-3da6-4b30-b633-6348b31b0b01" containerName="extract-utilities" Nov 21 14:13:01 crc kubenswrapper[4897]: E1121 14:13:01.654983 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ea4ecd1-6706-4a89-8034-4f41ee7e0f2d" containerName="registry-server" Nov 21 14:13:01 crc kubenswrapper[4897]: I1121 14:13:01.654992 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ea4ecd1-6706-4a89-8034-4f41ee7e0f2d" containerName="registry-server" Nov 21 14:13:01 crc kubenswrapper[4897]: I1121 14:13:01.655122 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="a55d260f-3da6-4b30-b633-6348b31b0b01" containerName="registry-server" Nov 21 14:13:01 crc kubenswrapper[4897]: I1121 14:13:01.655150 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="7cd98bfc-77a5-4ef5-86f1-7614b0b2ad2a" containerName="pruner" Nov 21 14:13:01 crc kubenswrapper[4897]: I1121 14:13:01.655162 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="afe9da5b-bdfb-47c5-bd31-abc7c0d53b63" containerName="registry-server" Nov 21 14:13:01 crc kubenswrapper[4897]: I1121 14:13:01.655172 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8ce22b3-b193-4bd0-9a01-9bc260c20430" containerName="registry-server" Nov 21 14:13:01 crc kubenswrapper[4897]: I1121 14:13:01.655181 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ea4ecd1-6706-4a89-8034-4f41ee7e0f2d" containerName="registry-server" Nov 21 14:13:01 crc kubenswrapper[4897]: I1121 14:13:01.655192 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5aa94f1-611d-4695-a3c7-1c33c80f1a1a" containerName="pruner" Nov 21 14:13:01 crc kubenswrapper[4897]: I1121 14:13:01.655201 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e4b5f84-4954-4d85-8bae-db18ed352464" containerName="oauth-openshift" Nov 21 14:13:01 crc kubenswrapper[4897]: I1121 14:13:01.655835 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-57bcd9fbb-mlnrl" Nov 21 14:13:01 crc kubenswrapper[4897]: I1121 14:13:01.660196 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Nov 21 14:13:01 crc kubenswrapper[4897]: I1121 14:13:01.662343 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Nov 21 14:13:01 crc kubenswrapper[4897]: I1121 14:13:01.663414 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Nov 21 14:13:01 crc kubenswrapper[4897]: I1121 14:13:01.663617 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Nov 21 14:13:01 crc kubenswrapper[4897]: I1121 14:13:01.663795 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Nov 21 14:13:01 crc kubenswrapper[4897]: I1121 14:13:01.664005 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Nov 21 14:13:01 crc kubenswrapper[4897]: I1121 14:13:01.664189 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Nov 21 14:13:01 crc kubenswrapper[4897]: I1121 14:13:01.664367 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Nov 21 14:13:01 crc kubenswrapper[4897]: I1121 14:13:01.664576 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Nov 21 14:13:01 crc kubenswrapper[4897]: I1121 14:13:01.664754 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Nov 21 14:13:01 crc kubenswrapper[4897]: I1121 14:13:01.665128 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Nov 21 14:13:01 crc kubenswrapper[4897]: I1121 14:13:01.668966 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Nov 21 14:13:01 crc kubenswrapper[4897]: I1121 14:13:01.679258 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-57bcd9fbb-mlnrl"] Nov 21 14:13:01 crc kubenswrapper[4897]: I1121 14:13:01.679688 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Nov 21 14:13:01 crc kubenswrapper[4897]: I1121 14:13:01.685706 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Nov 21 14:13:01 crc kubenswrapper[4897]: I1121 14:13:01.688476 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Nov 21 14:13:01 crc kubenswrapper[4897]: I1121 14:13:01.754445 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/61aec289-e4e9-439e-9d27-b36498177683-v4-0-config-system-serving-cert\") pod \"oauth-openshift-57bcd9fbb-mlnrl\" (UID: \"61aec289-e4e9-439e-9d27-b36498177683\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-mlnrl" Nov 21 14:13:01 crc kubenswrapper[4897]: I1121 14:13:01.754868 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/61aec289-e4e9-439e-9d27-b36498177683-v4-0-config-user-template-login\") pod \"oauth-openshift-57bcd9fbb-mlnrl\" (UID: \"61aec289-e4e9-439e-9d27-b36498177683\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-mlnrl" Nov 21 14:13:01 crc kubenswrapper[4897]: I1121 14:13:01.755043 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/61aec289-e4e9-439e-9d27-b36498177683-v4-0-config-system-service-ca\") pod \"oauth-openshift-57bcd9fbb-mlnrl\" (UID: \"61aec289-e4e9-439e-9d27-b36498177683\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-mlnrl" Nov 21 14:13:01 crc kubenswrapper[4897]: I1121 14:13:01.755143 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/61aec289-e4e9-439e-9d27-b36498177683-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-57bcd9fbb-mlnrl\" (UID: \"61aec289-e4e9-439e-9d27-b36498177683\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-mlnrl" Nov 21 14:13:01 crc kubenswrapper[4897]: I1121 14:13:01.755248 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/61aec289-e4e9-439e-9d27-b36498177683-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-57bcd9fbb-mlnrl\" (UID: \"61aec289-e4e9-439e-9d27-b36498177683\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-mlnrl" Nov 21 14:13:01 crc kubenswrapper[4897]: I1121 14:13:01.755328 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/61aec289-e4e9-439e-9d27-b36498177683-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-57bcd9fbb-mlnrl\" (UID: \"61aec289-e4e9-439e-9d27-b36498177683\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-mlnrl" Nov 21 14:13:01 crc kubenswrapper[4897]: I1121 14:13:01.755590 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/61aec289-e4e9-439e-9d27-b36498177683-v4-0-config-system-cliconfig\") pod \"oauth-openshift-57bcd9fbb-mlnrl\" (UID: \"61aec289-e4e9-439e-9d27-b36498177683\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-mlnrl" Nov 21 14:13:01 crc kubenswrapper[4897]: I1121 14:13:01.755680 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/61aec289-e4e9-439e-9d27-b36498177683-v4-0-config-system-router-certs\") pod \"oauth-openshift-57bcd9fbb-mlnrl\" (UID: \"61aec289-e4e9-439e-9d27-b36498177683\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-mlnrl" Nov 21 14:13:01 crc kubenswrapper[4897]: I1121 14:13:01.755717 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/61aec289-e4e9-439e-9d27-b36498177683-audit-policies\") pod \"oauth-openshift-57bcd9fbb-mlnrl\" (UID: \"61aec289-e4e9-439e-9d27-b36498177683\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-mlnrl" Nov 21 14:13:01 crc kubenswrapper[4897]: I1121 14:13:01.755746 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jxl72\" (UniqueName: \"kubernetes.io/projected/61aec289-e4e9-439e-9d27-b36498177683-kube-api-access-jxl72\") pod \"oauth-openshift-57bcd9fbb-mlnrl\" (UID: \"61aec289-e4e9-439e-9d27-b36498177683\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-mlnrl" Nov 21 14:13:01 crc kubenswrapper[4897]: I1121 14:13:01.755766 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/61aec289-e4e9-439e-9d27-b36498177683-v4-0-config-user-template-error\") pod \"oauth-openshift-57bcd9fbb-mlnrl\" (UID: \"61aec289-e4e9-439e-9d27-b36498177683\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-mlnrl" Nov 21 14:13:01 crc kubenswrapper[4897]: I1121 14:13:01.755829 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/61aec289-e4e9-439e-9d27-b36498177683-audit-dir\") pod \"oauth-openshift-57bcd9fbb-mlnrl\" (UID: \"61aec289-e4e9-439e-9d27-b36498177683\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-mlnrl" Nov 21 14:13:01 crc kubenswrapper[4897]: I1121 14:13:01.755891 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/61aec289-e4e9-439e-9d27-b36498177683-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-57bcd9fbb-mlnrl\" (UID: \"61aec289-e4e9-439e-9d27-b36498177683\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-mlnrl" Nov 21 14:13:01 crc kubenswrapper[4897]: I1121 14:13:01.755929 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/61aec289-e4e9-439e-9d27-b36498177683-v4-0-config-system-session\") pod \"oauth-openshift-57bcd9fbb-mlnrl\" (UID: \"61aec289-e4e9-439e-9d27-b36498177683\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-mlnrl" Nov 21 14:13:01 crc kubenswrapper[4897]: I1121 14:13:01.856896 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/61aec289-e4e9-439e-9d27-b36498177683-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-57bcd9fbb-mlnrl\" (UID: \"61aec289-e4e9-439e-9d27-b36498177683\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-mlnrl" Nov 21 14:13:01 crc kubenswrapper[4897]: I1121 14:13:01.856942 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/61aec289-e4e9-439e-9d27-b36498177683-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-57bcd9fbb-mlnrl\" (UID: \"61aec289-e4e9-439e-9d27-b36498177683\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-mlnrl" Nov 21 14:13:01 crc kubenswrapper[4897]: I1121 14:13:01.856982 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/61aec289-e4e9-439e-9d27-b36498177683-v4-0-config-system-cliconfig\") pod \"oauth-openshift-57bcd9fbb-mlnrl\" (UID: \"61aec289-e4e9-439e-9d27-b36498177683\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-mlnrl" Nov 21 14:13:01 crc kubenswrapper[4897]: I1121 14:13:01.857004 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/61aec289-e4e9-439e-9d27-b36498177683-audit-policies\") pod \"oauth-openshift-57bcd9fbb-mlnrl\" (UID: \"61aec289-e4e9-439e-9d27-b36498177683\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-mlnrl" Nov 21 14:13:01 crc kubenswrapper[4897]: I1121 14:13:01.857026 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/61aec289-e4e9-439e-9d27-b36498177683-v4-0-config-system-router-certs\") pod \"oauth-openshift-57bcd9fbb-mlnrl\" (UID: \"61aec289-e4e9-439e-9d27-b36498177683\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-mlnrl" Nov 21 14:13:01 crc kubenswrapper[4897]: I1121 14:13:01.857042 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/61aec289-e4e9-439e-9d27-b36498177683-v4-0-config-user-template-error\") pod \"oauth-openshift-57bcd9fbb-mlnrl\" (UID: \"61aec289-e4e9-439e-9d27-b36498177683\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-mlnrl" Nov 21 14:13:01 crc kubenswrapper[4897]: I1121 14:13:01.857057 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jxl72\" (UniqueName: \"kubernetes.io/projected/61aec289-e4e9-439e-9d27-b36498177683-kube-api-access-jxl72\") pod \"oauth-openshift-57bcd9fbb-mlnrl\" (UID: \"61aec289-e4e9-439e-9d27-b36498177683\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-mlnrl" Nov 21 14:13:01 crc kubenswrapper[4897]: I1121 14:13:01.857080 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/61aec289-e4e9-439e-9d27-b36498177683-audit-dir\") pod \"oauth-openshift-57bcd9fbb-mlnrl\" (UID: \"61aec289-e4e9-439e-9d27-b36498177683\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-mlnrl" Nov 21 14:13:01 crc kubenswrapper[4897]: I1121 14:13:01.857101 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/61aec289-e4e9-439e-9d27-b36498177683-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-57bcd9fbb-mlnrl\" (UID: \"61aec289-e4e9-439e-9d27-b36498177683\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-mlnrl" Nov 21 14:13:01 crc kubenswrapper[4897]: I1121 14:13:01.857121 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/61aec289-e4e9-439e-9d27-b36498177683-v4-0-config-system-session\") pod \"oauth-openshift-57bcd9fbb-mlnrl\" (UID: \"61aec289-e4e9-439e-9d27-b36498177683\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-mlnrl" Nov 21 14:13:01 crc kubenswrapper[4897]: I1121 14:13:01.857148 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/61aec289-e4e9-439e-9d27-b36498177683-v4-0-config-system-serving-cert\") pod \"oauth-openshift-57bcd9fbb-mlnrl\" (UID: \"61aec289-e4e9-439e-9d27-b36498177683\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-mlnrl" Nov 21 14:13:01 crc kubenswrapper[4897]: I1121 14:13:01.857169 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/61aec289-e4e9-439e-9d27-b36498177683-v4-0-config-user-template-login\") pod \"oauth-openshift-57bcd9fbb-mlnrl\" (UID: \"61aec289-e4e9-439e-9d27-b36498177683\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-mlnrl" Nov 21 14:13:01 crc kubenswrapper[4897]: I1121 14:13:01.857190 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/61aec289-e4e9-439e-9d27-b36498177683-v4-0-config-system-service-ca\") pod \"oauth-openshift-57bcd9fbb-mlnrl\" (UID: \"61aec289-e4e9-439e-9d27-b36498177683\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-mlnrl" Nov 21 14:13:01 crc kubenswrapper[4897]: I1121 14:13:01.857211 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/61aec289-e4e9-439e-9d27-b36498177683-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-57bcd9fbb-mlnrl\" (UID: \"61aec289-e4e9-439e-9d27-b36498177683\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-mlnrl" Nov 21 14:13:01 crc kubenswrapper[4897]: I1121 14:13:01.858181 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/61aec289-e4e9-439e-9d27-b36498177683-audit-dir\") pod \"oauth-openshift-57bcd9fbb-mlnrl\" (UID: \"61aec289-e4e9-439e-9d27-b36498177683\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-mlnrl" Nov 21 14:13:01 crc kubenswrapper[4897]: I1121 14:13:01.858718 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/61aec289-e4e9-439e-9d27-b36498177683-audit-policies\") pod \"oauth-openshift-57bcd9fbb-mlnrl\" (UID: \"61aec289-e4e9-439e-9d27-b36498177683\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-mlnrl" Nov 21 14:13:01 crc kubenswrapper[4897]: I1121 14:13:01.859112 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/61aec289-e4e9-439e-9d27-b36498177683-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-57bcd9fbb-mlnrl\" (UID: \"61aec289-e4e9-439e-9d27-b36498177683\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-mlnrl" Nov 21 14:13:01 crc kubenswrapper[4897]: I1121 14:13:01.859234 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/61aec289-e4e9-439e-9d27-b36498177683-v4-0-config-system-service-ca\") pod \"oauth-openshift-57bcd9fbb-mlnrl\" (UID: \"61aec289-e4e9-439e-9d27-b36498177683\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-mlnrl" Nov 21 14:13:01 crc kubenswrapper[4897]: I1121 14:13:01.859419 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/61aec289-e4e9-439e-9d27-b36498177683-v4-0-config-system-cliconfig\") pod \"oauth-openshift-57bcd9fbb-mlnrl\" (UID: \"61aec289-e4e9-439e-9d27-b36498177683\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-mlnrl" Nov 21 14:13:01 crc kubenswrapper[4897]: I1121 14:13:01.864569 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/61aec289-e4e9-439e-9d27-b36498177683-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-57bcd9fbb-mlnrl\" (UID: \"61aec289-e4e9-439e-9d27-b36498177683\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-mlnrl" Nov 21 14:13:01 crc kubenswrapper[4897]: I1121 14:13:01.864601 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/61aec289-e4e9-439e-9d27-b36498177683-v4-0-config-user-template-login\") pod \"oauth-openshift-57bcd9fbb-mlnrl\" (UID: \"61aec289-e4e9-439e-9d27-b36498177683\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-mlnrl" Nov 21 14:13:01 crc kubenswrapper[4897]: I1121 14:13:01.864645 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/61aec289-e4e9-439e-9d27-b36498177683-v4-0-config-system-session\") pod \"oauth-openshift-57bcd9fbb-mlnrl\" (UID: \"61aec289-e4e9-439e-9d27-b36498177683\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-mlnrl" Nov 21 14:13:01 crc kubenswrapper[4897]: I1121 14:13:01.864804 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/61aec289-e4e9-439e-9d27-b36498177683-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-57bcd9fbb-mlnrl\" (UID: \"61aec289-e4e9-439e-9d27-b36498177683\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-mlnrl" Nov 21 14:13:01 crc kubenswrapper[4897]: I1121 14:13:01.865446 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/61aec289-e4e9-439e-9d27-b36498177683-v4-0-config-system-serving-cert\") pod \"oauth-openshift-57bcd9fbb-mlnrl\" (UID: \"61aec289-e4e9-439e-9d27-b36498177683\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-mlnrl" Nov 21 14:13:01 crc kubenswrapper[4897]: I1121 14:13:01.865618 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/61aec289-e4e9-439e-9d27-b36498177683-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-57bcd9fbb-mlnrl\" (UID: \"61aec289-e4e9-439e-9d27-b36498177683\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-mlnrl" Nov 21 14:13:01 crc kubenswrapper[4897]: I1121 14:13:01.865833 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/61aec289-e4e9-439e-9d27-b36498177683-v4-0-config-system-router-certs\") pod \"oauth-openshift-57bcd9fbb-mlnrl\" (UID: \"61aec289-e4e9-439e-9d27-b36498177683\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-mlnrl" Nov 21 14:13:01 crc kubenswrapper[4897]: I1121 14:13:01.866081 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/61aec289-e4e9-439e-9d27-b36498177683-v4-0-config-user-template-error\") pod \"oauth-openshift-57bcd9fbb-mlnrl\" (UID: \"61aec289-e4e9-439e-9d27-b36498177683\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-mlnrl" Nov 21 14:13:01 crc kubenswrapper[4897]: I1121 14:13:01.878415 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jxl72\" (UniqueName: \"kubernetes.io/projected/61aec289-e4e9-439e-9d27-b36498177683-kube-api-access-jxl72\") pod \"oauth-openshift-57bcd9fbb-mlnrl\" (UID: \"61aec289-e4e9-439e-9d27-b36498177683\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-mlnrl" Nov 21 14:13:01 crc kubenswrapper[4897]: I1121 14:13:01.987294 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-57bcd9fbb-mlnrl" Nov 21 14:13:02 crc kubenswrapper[4897]: I1121 14:13:02.414262 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-57bcd9fbb-mlnrl"] Nov 21 14:13:03 crc kubenswrapper[4897]: I1121 14:13:03.360547 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-57bcd9fbb-mlnrl" event={"ID":"61aec289-e4e9-439e-9d27-b36498177683","Type":"ContainerStarted","Data":"2755e3986d71fa3751da01b1af3b7b20a17bbd39a32dea74b9cb1d4f5882c437"} Nov 21 14:13:03 crc kubenswrapper[4897]: I1121 14:13:03.361109 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-57bcd9fbb-mlnrl" event={"ID":"61aec289-e4e9-439e-9d27-b36498177683","Type":"ContainerStarted","Data":"a0e7e4b31b8fc711d564f8d9eb08d1ca4e6f7dcb5e730a9696d7d6d988acf515"} Nov 21 14:13:03 crc kubenswrapper[4897]: I1121 14:13:03.361127 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-57bcd9fbb-mlnrl" Nov 21 14:13:03 crc kubenswrapper[4897]: I1121 14:13:03.366243 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-57bcd9fbb-mlnrl" Nov 21 14:13:03 crc kubenswrapper[4897]: I1121 14:13:03.380629 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-57bcd9fbb-mlnrl" podStartSLOduration=37.380612008 podStartE2EDuration="37.380612008s" podCreationTimestamp="2025-11-21 14:12:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:13:03.377426623 +0000 UTC m=+260.662020118" watchObservedRunningTime="2025-11-21 14:13:03.380612008 +0000 UTC m=+260.665205483" Nov 21 14:13:16 crc kubenswrapper[4897]: I1121 14:13:16.070032 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-q6dd2"] Nov 21 14:13:16 crc kubenswrapper[4897]: I1121 14:13:16.071182 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-q6dd2" podUID="1dc1b877-a8de-45b0-bab4-4411ecced04d" containerName="registry-server" containerID="cri-o://57daea7e609a9d7bca2ddc01887c447f5a0488528509239c6a88833905c3e474" gracePeriod=30 Nov 21 14:13:16 crc kubenswrapper[4897]: I1121 14:13:16.077964 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tqhs9"] Nov 21 14:13:16 crc kubenswrapper[4897]: I1121 14:13:16.078535 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-tqhs9" podUID="767a13d7-0e18-4496-b0bb-5a39292622e2" containerName="registry-server" containerID="cri-o://01d955c394e426a56d3b10050ddce048da4f9940dbdf4902063201002c6aad75" gracePeriod=30 Nov 21 14:13:16 crc kubenswrapper[4897]: I1121 14:13:16.118191 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-z9hjh"] Nov 21 14:13:16 crc kubenswrapper[4897]: I1121 14:13:16.118569 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6fqg7"] Nov 21 14:13:16 crc kubenswrapper[4897]: I1121 14:13:16.118587 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-5jpnn"] Nov 21 14:13:16 crc kubenswrapper[4897]: I1121 14:13:16.123902 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-z9hjh" podUID="e128526a-8489-47e1-a711-4977f2b0d46d" containerName="marketplace-operator" containerID="cri-o://fd8a88f7278af250e4cea7de45d4fc4a9196d7593856d54edf70fd466c2f0ea0" gracePeriod=30 Nov 21 14:13:16 crc kubenswrapper[4897]: I1121 14:13:16.124085 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-6fqg7" podUID="8dde9f26-f67d-49d2-98ac-7483796f1212" containerName="registry-server" containerID="cri-o://2ca36527ca91ac00bf01ae7ada651924905110730aba5170b1d3d43a5aa97ab7" gracePeriod=30 Nov 21 14:13:16 crc kubenswrapper[4897]: I1121 14:13:16.124727 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-sqdvs"] Nov 21 14:13:16 crc kubenswrapper[4897]: I1121 14:13:16.124941 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-sqdvs" podUID="bf875c4e-691b-4d77-ab68-5ac658b6f39b" containerName="registry-server" containerID="cri-o://6936da99ee2db8e16a1af98a86c4cc9c32ec07a2c11853dff43fb6c255021576" gracePeriod=30 Nov 21 14:13:16 crc kubenswrapper[4897]: I1121 14:13:16.125091 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-5jpnn" Nov 21 14:13:16 crc kubenswrapper[4897]: I1121 14:13:16.132517 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-5jpnn"] Nov 21 14:13:16 crc kubenswrapper[4897]: I1121 14:13:16.262852 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cdgth\" (UniqueName: \"kubernetes.io/projected/c6d71827-f5bb-4d61-bb8f-25addd3be667-kube-api-access-cdgth\") pod \"marketplace-operator-79b997595-5jpnn\" (UID: \"c6d71827-f5bb-4d61-bb8f-25addd3be667\") " pod="openshift-marketplace/marketplace-operator-79b997595-5jpnn" Nov 21 14:13:16 crc kubenswrapper[4897]: I1121 14:13:16.263100 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c6d71827-f5bb-4d61-bb8f-25addd3be667-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-5jpnn\" (UID: \"c6d71827-f5bb-4d61-bb8f-25addd3be667\") " pod="openshift-marketplace/marketplace-operator-79b997595-5jpnn" Nov 21 14:13:16 crc kubenswrapper[4897]: I1121 14:13:16.263132 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/c6d71827-f5bb-4d61-bb8f-25addd3be667-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-5jpnn\" (UID: \"c6d71827-f5bb-4d61-bb8f-25addd3be667\") " pod="openshift-marketplace/marketplace-operator-79b997595-5jpnn" Nov 21 14:13:16 crc kubenswrapper[4897]: I1121 14:13:16.363889 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cdgth\" (UniqueName: \"kubernetes.io/projected/c6d71827-f5bb-4d61-bb8f-25addd3be667-kube-api-access-cdgth\") pod \"marketplace-operator-79b997595-5jpnn\" (UID: \"c6d71827-f5bb-4d61-bb8f-25addd3be667\") " pod="openshift-marketplace/marketplace-operator-79b997595-5jpnn" Nov 21 14:13:16 crc kubenswrapper[4897]: I1121 14:13:16.364288 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c6d71827-f5bb-4d61-bb8f-25addd3be667-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-5jpnn\" (UID: \"c6d71827-f5bb-4d61-bb8f-25addd3be667\") " pod="openshift-marketplace/marketplace-operator-79b997595-5jpnn" Nov 21 14:13:16 crc kubenswrapper[4897]: I1121 14:13:16.364310 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/c6d71827-f5bb-4d61-bb8f-25addd3be667-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-5jpnn\" (UID: \"c6d71827-f5bb-4d61-bb8f-25addd3be667\") " pod="openshift-marketplace/marketplace-operator-79b997595-5jpnn" Nov 21 14:13:16 crc kubenswrapper[4897]: I1121 14:13:16.365615 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c6d71827-f5bb-4d61-bb8f-25addd3be667-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-5jpnn\" (UID: \"c6d71827-f5bb-4d61-bb8f-25addd3be667\") " pod="openshift-marketplace/marketplace-operator-79b997595-5jpnn" Nov 21 14:13:16 crc kubenswrapper[4897]: I1121 14:13:16.373333 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/c6d71827-f5bb-4d61-bb8f-25addd3be667-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-5jpnn\" (UID: \"c6d71827-f5bb-4d61-bb8f-25addd3be667\") " pod="openshift-marketplace/marketplace-operator-79b997595-5jpnn" Nov 21 14:13:16 crc kubenswrapper[4897]: I1121 14:13:16.381972 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cdgth\" (UniqueName: \"kubernetes.io/projected/c6d71827-f5bb-4d61-bb8f-25addd3be667-kube-api-access-cdgth\") pod \"marketplace-operator-79b997595-5jpnn\" (UID: \"c6d71827-f5bb-4d61-bb8f-25addd3be667\") " pod="openshift-marketplace/marketplace-operator-79b997595-5jpnn" Nov 21 14:13:16 crc kubenswrapper[4897]: I1121 14:13:16.425963 4897 generic.go:334] "Generic (PLEG): container finished" podID="bf875c4e-691b-4d77-ab68-5ac658b6f39b" containerID="6936da99ee2db8e16a1af98a86c4cc9c32ec07a2c11853dff43fb6c255021576" exitCode=0 Nov 21 14:13:16 crc kubenswrapper[4897]: I1121 14:13:16.426046 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sqdvs" event={"ID":"bf875c4e-691b-4d77-ab68-5ac658b6f39b","Type":"ContainerDied","Data":"6936da99ee2db8e16a1af98a86c4cc9c32ec07a2c11853dff43fb6c255021576"} Nov 21 14:13:16 crc kubenswrapper[4897]: I1121 14:13:16.428379 4897 generic.go:334] "Generic (PLEG): container finished" podID="1dc1b877-a8de-45b0-bab4-4411ecced04d" containerID="57daea7e609a9d7bca2ddc01887c447f5a0488528509239c6a88833905c3e474" exitCode=0 Nov 21 14:13:16 crc kubenswrapper[4897]: I1121 14:13:16.428472 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q6dd2" event={"ID":"1dc1b877-a8de-45b0-bab4-4411ecced04d","Type":"ContainerDied","Data":"57daea7e609a9d7bca2ddc01887c447f5a0488528509239c6a88833905c3e474"} Nov 21 14:13:16 crc kubenswrapper[4897]: I1121 14:13:16.428520 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q6dd2" event={"ID":"1dc1b877-a8de-45b0-bab4-4411ecced04d","Type":"ContainerDied","Data":"0e682cfbb58918fb97e000f0b36d176994a297cf3e898b70d21a249bcb5a1c26"} Nov 21 14:13:16 crc kubenswrapper[4897]: I1121 14:13:16.428538 4897 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0e682cfbb58918fb97e000f0b36d176994a297cf3e898b70d21a249bcb5a1c26" Nov 21 14:13:16 crc kubenswrapper[4897]: E1121 14:13:16.437168 4897 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8dde9f26_f67d_49d2_98ac_7483796f1212.slice/crio-2ca36527ca91ac00bf01ae7ada651924905110730aba5170b1d3d43a5aa97ab7.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode128526a_8489_47e1_a711_4977f2b0d46d.slice/crio-conmon-fd8a88f7278af250e4cea7de45d4fc4a9196d7593856d54edf70fd466c2f0ea0.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8dde9f26_f67d_49d2_98ac_7483796f1212.slice/crio-conmon-2ca36527ca91ac00bf01ae7ada651924905110730aba5170b1d3d43a5aa97ab7.scope\": RecentStats: unable to find data in memory cache]" Nov 21 14:13:16 crc kubenswrapper[4897]: I1121 14:13:16.445538 4897 generic.go:334] "Generic (PLEG): container finished" podID="8dde9f26-f67d-49d2-98ac-7483796f1212" containerID="2ca36527ca91ac00bf01ae7ada651924905110730aba5170b1d3d43a5aa97ab7" exitCode=0 Nov 21 14:13:16 crc kubenswrapper[4897]: I1121 14:13:16.445581 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6fqg7" event={"ID":"8dde9f26-f67d-49d2-98ac-7483796f1212","Type":"ContainerDied","Data":"2ca36527ca91ac00bf01ae7ada651924905110730aba5170b1d3d43a5aa97ab7"} Nov 21 14:13:16 crc kubenswrapper[4897]: I1121 14:13:16.448252 4897 generic.go:334] "Generic (PLEG): container finished" podID="767a13d7-0e18-4496-b0bb-5a39292622e2" containerID="01d955c394e426a56d3b10050ddce048da4f9940dbdf4902063201002c6aad75" exitCode=0 Nov 21 14:13:16 crc kubenswrapper[4897]: I1121 14:13:16.448330 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tqhs9" event={"ID":"767a13d7-0e18-4496-b0bb-5a39292622e2","Type":"ContainerDied","Data":"01d955c394e426a56d3b10050ddce048da4f9940dbdf4902063201002c6aad75"} Nov 21 14:13:16 crc kubenswrapper[4897]: I1121 14:13:16.450345 4897 generic.go:334] "Generic (PLEG): container finished" podID="e128526a-8489-47e1-a711-4977f2b0d46d" containerID="fd8a88f7278af250e4cea7de45d4fc4a9196d7593856d54edf70fd466c2f0ea0" exitCode=0 Nov 21 14:13:16 crc kubenswrapper[4897]: I1121 14:13:16.450441 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-z9hjh" event={"ID":"e128526a-8489-47e1-a711-4977f2b0d46d","Type":"ContainerDied","Data":"fd8a88f7278af250e4cea7de45d4fc4a9196d7593856d54edf70fd466c2f0ea0"} Nov 21 14:13:16 crc kubenswrapper[4897]: I1121 14:13:16.464045 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-5jpnn" Nov 21 14:13:16 crc kubenswrapper[4897]: I1121 14:13:16.473551 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-q6dd2" Nov 21 14:13:16 crc kubenswrapper[4897]: I1121 14:13:16.510316 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tqhs9" Nov 21 14:13:16 crc kubenswrapper[4897]: I1121 14:13:16.568467 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1dc1b877-a8de-45b0-bab4-4411ecced04d-utilities\") pod \"1dc1b877-a8de-45b0-bab4-4411ecced04d\" (UID: \"1dc1b877-a8de-45b0-bab4-4411ecced04d\") " Nov 21 14:13:16 crc kubenswrapper[4897]: I1121 14:13:16.568566 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/767a13d7-0e18-4496-b0bb-5a39292622e2-utilities\") pod \"767a13d7-0e18-4496-b0bb-5a39292622e2\" (UID: \"767a13d7-0e18-4496-b0bb-5a39292622e2\") " Nov 21 14:13:16 crc kubenswrapper[4897]: I1121 14:13:16.568599 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/767a13d7-0e18-4496-b0bb-5a39292622e2-catalog-content\") pod \"767a13d7-0e18-4496-b0bb-5a39292622e2\" (UID: \"767a13d7-0e18-4496-b0bb-5a39292622e2\") " Nov 21 14:13:16 crc kubenswrapper[4897]: I1121 14:13:16.568662 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-59rxs\" (UniqueName: \"kubernetes.io/projected/767a13d7-0e18-4496-b0bb-5a39292622e2-kube-api-access-59rxs\") pod \"767a13d7-0e18-4496-b0bb-5a39292622e2\" (UID: \"767a13d7-0e18-4496-b0bb-5a39292622e2\") " Nov 21 14:13:16 crc kubenswrapper[4897]: I1121 14:13:16.568698 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l5xws\" (UniqueName: \"kubernetes.io/projected/1dc1b877-a8de-45b0-bab4-4411ecced04d-kube-api-access-l5xws\") pod \"1dc1b877-a8de-45b0-bab4-4411ecced04d\" (UID: \"1dc1b877-a8de-45b0-bab4-4411ecced04d\") " Nov 21 14:13:16 crc kubenswrapper[4897]: I1121 14:13:16.568725 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1dc1b877-a8de-45b0-bab4-4411ecced04d-catalog-content\") pod \"1dc1b877-a8de-45b0-bab4-4411ecced04d\" (UID: \"1dc1b877-a8de-45b0-bab4-4411ecced04d\") " Nov 21 14:13:16 crc kubenswrapper[4897]: I1121 14:13:16.569949 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/767a13d7-0e18-4496-b0bb-5a39292622e2-utilities" (OuterVolumeSpecName: "utilities") pod "767a13d7-0e18-4496-b0bb-5a39292622e2" (UID: "767a13d7-0e18-4496-b0bb-5a39292622e2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:13:16 crc kubenswrapper[4897]: I1121 14:13:16.570664 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1dc1b877-a8de-45b0-bab4-4411ecced04d-utilities" (OuterVolumeSpecName: "utilities") pod "1dc1b877-a8de-45b0-bab4-4411ecced04d" (UID: "1dc1b877-a8de-45b0-bab4-4411ecced04d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:13:16 crc kubenswrapper[4897]: I1121 14:13:16.579691 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/767a13d7-0e18-4496-b0bb-5a39292622e2-kube-api-access-59rxs" (OuterVolumeSpecName: "kube-api-access-59rxs") pod "767a13d7-0e18-4496-b0bb-5a39292622e2" (UID: "767a13d7-0e18-4496-b0bb-5a39292622e2"). InnerVolumeSpecName "kube-api-access-59rxs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:13:16 crc kubenswrapper[4897]: I1121 14:13:16.580436 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1dc1b877-a8de-45b0-bab4-4411ecced04d-kube-api-access-l5xws" (OuterVolumeSpecName: "kube-api-access-l5xws") pod "1dc1b877-a8de-45b0-bab4-4411ecced04d" (UID: "1dc1b877-a8de-45b0-bab4-4411ecced04d"). InnerVolumeSpecName "kube-api-access-l5xws". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:13:16 crc kubenswrapper[4897]: I1121 14:13:16.633786 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1dc1b877-a8de-45b0-bab4-4411ecced04d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1dc1b877-a8de-45b0-bab4-4411ecced04d" (UID: "1dc1b877-a8de-45b0-bab4-4411ecced04d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:13:16 crc kubenswrapper[4897]: I1121 14:13:16.638860 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/767a13d7-0e18-4496-b0bb-5a39292622e2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "767a13d7-0e18-4496-b0bb-5a39292622e2" (UID: "767a13d7-0e18-4496-b0bb-5a39292622e2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:13:16 crc kubenswrapper[4897]: I1121 14:13:16.646249 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6fqg7" Nov 21 14:13:16 crc kubenswrapper[4897]: I1121 14:13:16.652678 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-z9hjh" Nov 21 14:13:16 crc kubenswrapper[4897]: I1121 14:13:16.670546 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l5xws\" (UniqueName: \"kubernetes.io/projected/1dc1b877-a8de-45b0-bab4-4411ecced04d-kube-api-access-l5xws\") on node \"crc\" DevicePath \"\"" Nov 21 14:13:16 crc kubenswrapper[4897]: I1121 14:13:16.670583 4897 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1dc1b877-a8de-45b0-bab4-4411ecced04d-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 21 14:13:16 crc kubenswrapper[4897]: I1121 14:13:16.670596 4897 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1dc1b877-a8de-45b0-bab4-4411ecced04d-utilities\") on node \"crc\" DevicePath \"\"" Nov 21 14:13:16 crc kubenswrapper[4897]: I1121 14:13:16.670607 4897 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/767a13d7-0e18-4496-b0bb-5a39292622e2-utilities\") on node \"crc\" DevicePath \"\"" Nov 21 14:13:16 crc kubenswrapper[4897]: I1121 14:13:16.670618 4897 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/767a13d7-0e18-4496-b0bb-5a39292622e2-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 21 14:13:16 crc kubenswrapper[4897]: I1121 14:13:16.670630 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-59rxs\" (UniqueName: \"kubernetes.io/projected/767a13d7-0e18-4496-b0bb-5a39292622e2-kube-api-access-59rxs\") on node \"crc\" DevicePath \"\"" Nov 21 14:13:16 crc kubenswrapper[4897]: I1121 14:13:16.692585 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sqdvs" Nov 21 14:13:16 crc kubenswrapper[4897]: I1121 14:13:16.772032 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e128526a-8489-47e1-a711-4977f2b0d46d-marketplace-trusted-ca\") pod \"e128526a-8489-47e1-a711-4977f2b0d46d\" (UID: \"e128526a-8489-47e1-a711-4977f2b0d46d\") " Nov 21 14:13:16 crc kubenswrapper[4897]: I1121 14:13:16.772105 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8dde9f26-f67d-49d2-98ac-7483796f1212-catalog-content\") pod \"8dde9f26-f67d-49d2-98ac-7483796f1212\" (UID: \"8dde9f26-f67d-49d2-98ac-7483796f1212\") " Nov 21 14:13:16 crc kubenswrapper[4897]: I1121 14:13:16.772161 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2wp82\" (UniqueName: \"kubernetes.io/projected/e128526a-8489-47e1-a711-4977f2b0d46d-kube-api-access-2wp82\") pod \"e128526a-8489-47e1-a711-4977f2b0d46d\" (UID: \"e128526a-8489-47e1-a711-4977f2b0d46d\") " Nov 21 14:13:16 crc kubenswrapper[4897]: I1121 14:13:16.772184 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf875c4e-691b-4d77-ab68-5ac658b6f39b-utilities\") pod \"bf875c4e-691b-4d77-ab68-5ac658b6f39b\" (UID: \"bf875c4e-691b-4d77-ab68-5ac658b6f39b\") " Nov 21 14:13:16 crc kubenswrapper[4897]: I1121 14:13:16.772203 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/e128526a-8489-47e1-a711-4977f2b0d46d-marketplace-operator-metrics\") pod \"e128526a-8489-47e1-a711-4977f2b0d46d\" (UID: \"e128526a-8489-47e1-a711-4977f2b0d46d\") " Nov 21 14:13:16 crc kubenswrapper[4897]: I1121 14:13:16.772233 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8dde9f26-f67d-49d2-98ac-7483796f1212-utilities\") pod \"8dde9f26-f67d-49d2-98ac-7483796f1212\" (UID: \"8dde9f26-f67d-49d2-98ac-7483796f1212\") " Nov 21 14:13:16 crc kubenswrapper[4897]: I1121 14:13:16.772260 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f94h4\" (UniqueName: \"kubernetes.io/projected/bf875c4e-691b-4d77-ab68-5ac658b6f39b-kube-api-access-f94h4\") pod \"bf875c4e-691b-4d77-ab68-5ac658b6f39b\" (UID: \"bf875c4e-691b-4d77-ab68-5ac658b6f39b\") " Nov 21 14:13:16 crc kubenswrapper[4897]: I1121 14:13:16.772291 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4x5qq\" (UniqueName: \"kubernetes.io/projected/8dde9f26-f67d-49d2-98ac-7483796f1212-kube-api-access-4x5qq\") pod \"8dde9f26-f67d-49d2-98ac-7483796f1212\" (UID: \"8dde9f26-f67d-49d2-98ac-7483796f1212\") " Nov 21 14:13:16 crc kubenswrapper[4897]: I1121 14:13:16.772326 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf875c4e-691b-4d77-ab68-5ac658b6f39b-catalog-content\") pod \"bf875c4e-691b-4d77-ab68-5ac658b6f39b\" (UID: \"bf875c4e-691b-4d77-ab68-5ac658b6f39b\") " Nov 21 14:13:16 crc kubenswrapper[4897]: I1121 14:13:16.772632 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e128526a-8489-47e1-a711-4977f2b0d46d-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "e128526a-8489-47e1-a711-4977f2b0d46d" (UID: "e128526a-8489-47e1-a711-4977f2b0d46d"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:13:16 crc kubenswrapper[4897]: I1121 14:13:16.773430 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bf875c4e-691b-4d77-ab68-5ac658b6f39b-utilities" (OuterVolumeSpecName: "utilities") pod "bf875c4e-691b-4d77-ab68-5ac658b6f39b" (UID: "bf875c4e-691b-4d77-ab68-5ac658b6f39b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:13:16 crc kubenswrapper[4897]: I1121 14:13:16.773499 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8dde9f26-f67d-49d2-98ac-7483796f1212-utilities" (OuterVolumeSpecName: "utilities") pod "8dde9f26-f67d-49d2-98ac-7483796f1212" (UID: "8dde9f26-f67d-49d2-98ac-7483796f1212"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:13:16 crc kubenswrapper[4897]: I1121 14:13:16.775454 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8dde9f26-f67d-49d2-98ac-7483796f1212-kube-api-access-4x5qq" (OuterVolumeSpecName: "kube-api-access-4x5qq") pod "8dde9f26-f67d-49d2-98ac-7483796f1212" (UID: "8dde9f26-f67d-49d2-98ac-7483796f1212"). InnerVolumeSpecName "kube-api-access-4x5qq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:13:16 crc kubenswrapper[4897]: I1121 14:13:16.775902 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e128526a-8489-47e1-a711-4977f2b0d46d-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "e128526a-8489-47e1-a711-4977f2b0d46d" (UID: "e128526a-8489-47e1-a711-4977f2b0d46d"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:13:16 crc kubenswrapper[4897]: I1121 14:13:16.776131 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e128526a-8489-47e1-a711-4977f2b0d46d-kube-api-access-2wp82" (OuterVolumeSpecName: "kube-api-access-2wp82") pod "e128526a-8489-47e1-a711-4977f2b0d46d" (UID: "e128526a-8489-47e1-a711-4977f2b0d46d"). InnerVolumeSpecName "kube-api-access-2wp82". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:13:16 crc kubenswrapper[4897]: I1121 14:13:16.777874 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf875c4e-691b-4d77-ab68-5ac658b6f39b-kube-api-access-f94h4" (OuterVolumeSpecName: "kube-api-access-f94h4") pod "bf875c4e-691b-4d77-ab68-5ac658b6f39b" (UID: "bf875c4e-691b-4d77-ab68-5ac658b6f39b"). InnerVolumeSpecName "kube-api-access-f94h4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:13:16 crc kubenswrapper[4897]: I1121 14:13:16.800712 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8dde9f26-f67d-49d2-98ac-7483796f1212-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8dde9f26-f67d-49d2-98ac-7483796f1212" (UID: "8dde9f26-f67d-49d2-98ac-7483796f1212"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:13:16 crc kubenswrapper[4897]: I1121 14:13:16.869557 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bf875c4e-691b-4d77-ab68-5ac658b6f39b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bf875c4e-691b-4d77-ab68-5ac658b6f39b" (UID: "bf875c4e-691b-4d77-ab68-5ac658b6f39b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:13:16 crc kubenswrapper[4897]: I1121 14:13:16.873224 4897 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf875c4e-691b-4d77-ab68-5ac658b6f39b-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 21 14:13:16 crc kubenswrapper[4897]: I1121 14:13:16.873257 4897 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e128526a-8489-47e1-a711-4977f2b0d46d-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 21 14:13:16 crc kubenswrapper[4897]: I1121 14:13:16.873269 4897 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8dde9f26-f67d-49d2-98ac-7483796f1212-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 21 14:13:16 crc kubenswrapper[4897]: I1121 14:13:16.873278 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2wp82\" (UniqueName: \"kubernetes.io/projected/e128526a-8489-47e1-a711-4977f2b0d46d-kube-api-access-2wp82\") on node \"crc\" DevicePath \"\"" Nov 21 14:13:16 crc kubenswrapper[4897]: I1121 14:13:16.873289 4897 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/e128526a-8489-47e1-a711-4977f2b0d46d-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Nov 21 14:13:16 crc kubenswrapper[4897]: I1121 14:13:16.873309 4897 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf875c4e-691b-4d77-ab68-5ac658b6f39b-utilities\") on node \"crc\" DevicePath \"\"" Nov 21 14:13:16 crc kubenswrapper[4897]: I1121 14:13:16.873320 4897 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8dde9f26-f67d-49d2-98ac-7483796f1212-utilities\") on node \"crc\" DevicePath \"\"" Nov 21 14:13:16 crc kubenswrapper[4897]: I1121 14:13:16.873332 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f94h4\" (UniqueName: \"kubernetes.io/projected/bf875c4e-691b-4d77-ab68-5ac658b6f39b-kube-api-access-f94h4\") on node \"crc\" DevicePath \"\"" Nov 21 14:13:16 crc kubenswrapper[4897]: I1121 14:13:16.873344 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4x5qq\" (UniqueName: \"kubernetes.io/projected/8dde9f26-f67d-49d2-98ac-7483796f1212-kube-api-access-4x5qq\") on node \"crc\" DevicePath \"\"" Nov 21 14:13:16 crc kubenswrapper[4897]: I1121 14:13:16.945585 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-5jpnn"] Nov 21 14:13:16 crc kubenswrapper[4897]: W1121 14:13:16.952756 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc6d71827_f5bb_4d61_bb8f_25addd3be667.slice/crio-f7674eb01cc64c83e3d912e3a73c2638a814baddc17e563b20d5e34f8927af65 WatchSource:0}: Error finding container f7674eb01cc64c83e3d912e3a73c2638a814baddc17e563b20d5e34f8927af65: Status 404 returned error can't find the container with id f7674eb01cc64c83e3d912e3a73c2638a814baddc17e563b20d5e34f8927af65 Nov 21 14:13:17 crc kubenswrapper[4897]: I1121 14:13:17.456065 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tqhs9" event={"ID":"767a13d7-0e18-4496-b0bb-5a39292622e2","Type":"ContainerDied","Data":"0bc586219d861cea2930e3953fe8cc3621f9fca40536d64da3db17d429927282"} Nov 21 14:13:17 crc kubenswrapper[4897]: I1121 14:13:17.456114 4897 scope.go:117] "RemoveContainer" containerID="01d955c394e426a56d3b10050ddce048da4f9940dbdf4902063201002c6aad75" Nov 21 14:13:17 crc kubenswrapper[4897]: I1121 14:13:17.456207 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tqhs9" Nov 21 14:13:17 crc kubenswrapper[4897]: I1121 14:13:17.460006 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-z9hjh" event={"ID":"e128526a-8489-47e1-a711-4977f2b0d46d","Type":"ContainerDied","Data":"c045494f8df7b7987645b495a3806d7fe2c61470dfc6a9479ad0ae14c20f251d"} Nov 21 14:13:17 crc kubenswrapper[4897]: I1121 14:13:17.460053 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-z9hjh" Nov 21 14:13:17 crc kubenswrapper[4897]: I1121 14:13:17.461434 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-5jpnn" event={"ID":"c6d71827-f5bb-4d61-bb8f-25addd3be667","Type":"ContainerStarted","Data":"32ea39fbf351358b15a05480f0eb4c1333853cbac36156d9cda162c4b79ebefc"} Nov 21 14:13:17 crc kubenswrapper[4897]: I1121 14:13:17.461458 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-5jpnn" event={"ID":"c6d71827-f5bb-4d61-bb8f-25addd3be667","Type":"ContainerStarted","Data":"f7674eb01cc64c83e3d912e3a73c2638a814baddc17e563b20d5e34f8927af65"} Nov 21 14:13:17 crc kubenswrapper[4897]: I1121 14:13:17.461604 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-5jpnn" Nov 21 14:13:17 crc kubenswrapper[4897]: I1121 14:13:17.463779 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-5jpnn" Nov 21 14:13:17 crc kubenswrapper[4897]: I1121 14:13:17.465804 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sqdvs" event={"ID":"bf875c4e-691b-4d77-ab68-5ac658b6f39b","Type":"ContainerDied","Data":"fddb7ffc183f994cf60ee875192cf65a4000ea02b86a4615f1a69859ad8eef58"} Nov 21 14:13:17 crc kubenswrapper[4897]: I1121 14:13:17.465897 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sqdvs" Nov 21 14:13:17 crc kubenswrapper[4897]: I1121 14:13:17.469177 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6fqg7" event={"ID":"8dde9f26-f67d-49d2-98ac-7483796f1212","Type":"ContainerDied","Data":"6f0c3aae606f7cac3c0a134f53733b7f6c1d62feb2cbedc2915445c42a62df52"} Nov 21 14:13:17 crc kubenswrapper[4897]: I1121 14:13:17.469185 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-q6dd2" Nov 21 14:13:17 crc kubenswrapper[4897]: I1121 14:13:17.469185 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6fqg7" Nov 21 14:13:17 crc kubenswrapper[4897]: I1121 14:13:17.475756 4897 scope.go:117] "RemoveContainer" containerID="404afe9d100084f246838f84effde9c75118dddb3d6f048061e3292125b4ca03" Nov 21 14:13:17 crc kubenswrapper[4897]: I1121 14:13:17.493341 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-5jpnn" podStartSLOduration=1.493320739 podStartE2EDuration="1.493320739s" podCreationTimestamp="2025-11-21 14:13:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:13:17.492545306 +0000 UTC m=+274.777138781" watchObservedRunningTime="2025-11-21 14:13:17.493320739 +0000 UTC m=+274.777914214" Nov 21 14:13:17 crc kubenswrapper[4897]: I1121 14:13:17.507679 4897 scope.go:117] "RemoveContainer" containerID="6be88453f6e893cdb6bf58762103e7d29f4937b2476a9dde9549dd05fc04fad8" Nov 21 14:13:17 crc kubenswrapper[4897]: I1121 14:13:17.510715 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tqhs9"] Nov 21 14:13:17 crc kubenswrapper[4897]: I1121 14:13:17.514807 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-tqhs9"] Nov 21 14:13:17 crc kubenswrapper[4897]: I1121 14:13:17.537994 4897 scope.go:117] "RemoveContainer" containerID="fd8a88f7278af250e4cea7de45d4fc4a9196d7593856d54edf70fd466c2f0ea0" Nov 21 14:13:17 crc kubenswrapper[4897]: I1121 14:13:17.547364 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-z9hjh"] Nov 21 14:13:17 crc kubenswrapper[4897]: I1121 14:13:17.554314 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-z9hjh"] Nov 21 14:13:17 crc kubenswrapper[4897]: I1121 14:13:17.559306 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6fqg7"] Nov 21 14:13:17 crc kubenswrapper[4897]: I1121 14:13:17.569479 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-6fqg7"] Nov 21 14:13:17 crc kubenswrapper[4897]: I1121 14:13:17.576968 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-q6dd2"] Nov 21 14:13:17 crc kubenswrapper[4897]: I1121 14:13:17.579099 4897 scope.go:117] "RemoveContainer" containerID="6936da99ee2db8e16a1af98a86c4cc9c32ec07a2c11853dff43fb6c255021576" Nov 21 14:13:17 crc kubenswrapper[4897]: I1121 14:13:17.587227 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-q6dd2"] Nov 21 14:13:17 crc kubenswrapper[4897]: I1121 14:13:17.591084 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-sqdvs"] Nov 21 14:13:17 crc kubenswrapper[4897]: I1121 14:13:17.594482 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-sqdvs"] Nov 21 14:13:17 crc kubenswrapper[4897]: I1121 14:13:17.602987 4897 scope.go:117] "RemoveContainer" containerID="52dc206c72235c5655018ca60c89bc6e0406c05e6ccad554b5c6930b171b6d52" Nov 21 14:13:17 crc kubenswrapper[4897]: I1121 14:13:17.618847 4897 scope.go:117] "RemoveContainer" containerID="423decd312997271961e228a25768e0f740f3fb7a02fcf8559757f5655bbf1d1" Nov 21 14:13:17 crc kubenswrapper[4897]: I1121 14:13:17.633132 4897 scope.go:117] "RemoveContainer" containerID="2ca36527ca91ac00bf01ae7ada651924905110730aba5170b1d3d43a5aa97ab7" Nov 21 14:13:17 crc kubenswrapper[4897]: I1121 14:13:17.645559 4897 scope.go:117] "RemoveContainer" containerID="46112836ee5444839fd76ed891b605d0f9a89cc0b735c1e6aefcc0e23623a60d" Nov 21 14:13:17 crc kubenswrapper[4897]: I1121 14:13:17.658093 4897 scope.go:117] "RemoveContainer" containerID="0f2fb4771e7be9d2354d41b4fee2027b5fce078a7d8ebb7f5824f08c597bbf83" Nov 21 14:13:18 crc kubenswrapper[4897]: I1121 14:13:18.098435 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1dc1b877-a8de-45b0-bab4-4411ecced04d" path="/var/lib/kubelet/pods/1dc1b877-a8de-45b0-bab4-4411ecced04d/volumes" Nov 21 14:13:18 crc kubenswrapper[4897]: I1121 14:13:18.099469 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="767a13d7-0e18-4496-b0bb-5a39292622e2" path="/var/lib/kubelet/pods/767a13d7-0e18-4496-b0bb-5a39292622e2/volumes" Nov 21 14:13:18 crc kubenswrapper[4897]: I1121 14:13:18.100198 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8dde9f26-f67d-49d2-98ac-7483796f1212" path="/var/lib/kubelet/pods/8dde9f26-f67d-49d2-98ac-7483796f1212/volumes" Nov 21 14:13:18 crc kubenswrapper[4897]: I1121 14:13:18.101448 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf875c4e-691b-4d77-ab68-5ac658b6f39b" path="/var/lib/kubelet/pods/bf875c4e-691b-4d77-ab68-5ac658b6f39b/volumes" Nov 21 14:13:18 crc kubenswrapper[4897]: I1121 14:13:18.102114 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e128526a-8489-47e1-a711-4977f2b0d46d" path="/var/lib/kubelet/pods/e128526a-8489-47e1-a711-4977f2b0d46d/volumes" Nov 21 14:13:18 crc kubenswrapper[4897]: I1121 14:13:18.279222 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-jxqbm"] Nov 21 14:13:18 crc kubenswrapper[4897]: E1121 14:13:18.279422 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8dde9f26-f67d-49d2-98ac-7483796f1212" containerName="extract-content" Nov 21 14:13:18 crc kubenswrapper[4897]: I1121 14:13:18.279434 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="8dde9f26-f67d-49d2-98ac-7483796f1212" containerName="extract-content" Nov 21 14:13:18 crc kubenswrapper[4897]: E1121 14:13:18.279444 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="767a13d7-0e18-4496-b0bb-5a39292622e2" containerName="extract-content" Nov 21 14:13:18 crc kubenswrapper[4897]: I1121 14:13:18.279451 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="767a13d7-0e18-4496-b0bb-5a39292622e2" containerName="extract-content" Nov 21 14:13:18 crc kubenswrapper[4897]: E1121 14:13:18.279460 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1dc1b877-a8de-45b0-bab4-4411ecced04d" containerName="registry-server" Nov 21 14:13:18 crc kubenswrapper[4897]: I1121 14:13:18.279468 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="1dc1b877-a8de-45b0-bab4-4411ecced04d" containerName="registry-server" Nov 21 14:13:18 crc kubenswrapper[4897]: E1121 14:13:18.279478 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1dc1b877-a8de-45b0-bab4-4411ecced04d" containerName="extract-utilities" Nov 21 14:13:18 crc kubenswrapper[4897]: I1121 14:13:18.279483 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="1dc1b877-a8de-45b0-bab4-4411ecced04d" containerName="extract-utilities" Nov 21 14:13:18 crc kubenswrapper[4897]: E1121 14:13:18.279493 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1dc1b877-a8de-45b0-bab4-4411ecced04d" containerName="extract-content" Nov 21 14:13:18 crc kubenswrapper[4897]: I1121 14:13:18.279499 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="1dc1b877-a8de-45b0-bab4-4411ecced04d" containerName="extract-content" Nov 21 14:13:18 crc kubenswrapper[4897]: E1121 14:13:18.279524 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8dde9f26-f67d-49d2-98ac-7483796f1212" containerName="extract-utilities" Nov 21 14:13:18 crc kubenswrapper[4897]: I1121 14:13:18.279530 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="8dde9f26-f67d-49d2-98ac-7483796f1212" containerName="extract-utilities" Nov 21 14:13:18 crc kubenswrapper[4897]: E1121 14:13:18.279539 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e128526a-8489-47e1-a711-4977f2b0d46d" containerName="marketplace-operator" Nov 21 14:13:18 crc kubenswrapper[4897]: I1121 14:13:18.279544 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="e128526a-8489-47e1-a711-4977f2b0d46d" containerName="marketplace-operator" Nov 21 14:13:18 crc kubenswrapper[4897]: E1121 14:13:18.279553 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf875c4e-691b-4d77-ab68-5ac658b6f39b" containerName="extract-content" Nov 21 14:13:18 crc kubenswrapper[4897]: I1121 14:13:18.279559 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf875c4e-691b-4d77-ab68-5ac658b6f39b" containerName="extract-content" Nov 21 14:13:18 crc kubenswrapper[4897]: E1121 14:13:18.279567 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf875c4e-691b-4d77-ab68-5ac658b6f39b" containerName="registry-server" Nov 21 14:13:18 crc kubenswrapper[4897]: I1121 14:13:18.279572 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf875c4e-691b-4d77-ab68-5ac658b6f39b" containerName="registry-server" Nov 21 14:13:18 crc kubenswrapper[4897]: E1121 14:13:18.279580 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="767a13d7-0e18-4496-b0bb-5a39292622e2" containerName="registry-server" Nov 21 14:13:18 crc kubenswrapper[4897]: I1121 14:13:18.279585 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="767a13d7-0e18-4496-b0bb-5a39292622e2" containerName="registry-server" Nov 21 14:13:18 crc kubenswrapper[4897]: E1121 14:13:18.279594 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf875c4e-691b-4d77-ab68-5ac658b6f39b" containerName="extract-utilities" Nov 21 14:13:18 crc kubenswrapper[4897]: I1121 14:13:18.279600 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf875c4e-691b-4d77-ab68-5ac658b6f39b" containerName="extract-utilities" Nov 21 14:13:18 crc kubenswrapper[4897]: E1121 14:13:18.279615 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="767a13d7-0e18-4496-b0bb-5a39292622e2" containerName="extract-utilities" Nov 21 14:13:18 crc kubenswrapper[4897]: I1121 14:13:18.279621 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="767a13d7-0e18-4496-b0bb-5a39292622e2" containerName="extract-utilities" Nov 21 14:13:18 crc kubenswrapper[4897]: E1121 14:13:18.279630 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8dde9f26-f67d-49d2-98ac-7483796f1212" containerName="registry-server" Nov 21 14:13:18 crc kubenswrapper[4897]: I1121 14:13:18.279636 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="8dde9f26-f67d-49d2-98ac-7483796f1212" containerName="registry-server" Nov 21 14:13:18 crc kubenswrapper[4897]: I1121 14:13:18.279707 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="e128526a-8489-47e1-a711-4977f2b0d46d" containerName="marketplace-operator" Nov 21 14:13:18 crc kubenswrapper[4897]: I1121 14:13:18.279718 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="767a13d7-0e18-4496-b0bb-5a39292622e2" containerName="registry-server" Nov 21 14:13:18 crc kubenswrapper[4897]: I1121 14:13:18.279725 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf875c4e-691b-4d77-ab68-5ac658b6f39b" containerName="registry-server" Nov 21 14:13:18 crc kubenswrapper[4897]: I1121 14:13:18.279734 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="1dc1b877-a8de-45b0-bab4-4411ecced04d" containerName="registry-server" Nov 21 14:13:18 crc kubenswrapper[4897]: I1121 14:13:18.279743 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="8dde9f26-f67d-49d2-98ac-7483796f1212" containerName="registry-server" Nov 21 14:13:18 crc kubenswrapper[4897]: I1121 14:13:18.280421 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jxqbm" Nov 21 14:13:18 crc kubenswrapper[4897]: I1121 14:13:18.282124 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 21 14:13:18 crc kubenswrapper[4897]: I1121 14:13:18.292878 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jxqbm"] Nov 21 14:13:18 crc kubenswrapper[4897]: I1121 14:13:18.293323 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0fcfc1e5-8927-4b4b-bd19-234497f86f23-utilities\") pod \"redhat-marketplace-jxqbm\" (UID: \"0fcfc1e5-8927-4b4b-bd19-234497f86f23\") " pod="openshift-marketplace/redhat-marketplace-jxqbm" Nov 21 14:13:18 crc kubenswrapper[4897]: I1121 14:13:18.293468 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-brhkc\" (UniqueName: \"kubernetes.io/projected/0fcfc1e5-8927-4b4b-bd19-234497f86f23-kube-api-access-brhkc\") pod \"redhat-marketplace-jxqbm\" (UID: \"0fcfc1e5-8927-4b4b-bd19-234497f86f23\") " pod="openshift-marketplace/redhat-marketplace-jxqbm" Nov 21 14:13:18 crc kubenswrapper[4897]: I1121 14:13:18.293607 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0fcfc1e5-8927-4b4b-bd19-234497f86f23-catalog-content\") pod \"redhat-marketplace-jxqbm\" (UID: \"0fcfc1e5-8927-4b4b-bd19-234497f86f23\") " pod="openshift-marketplace/redhat-marketplace-jxqbm" Nov 21 14:13:18 crc kubenswrapper[4897]: I1121 14:13:18.394098 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0fcfc1e5-8927-4b4b-bd19-234497f86f23-utilities\") pod \"redhat-marketplace-jxqbm\" (UID: \"0fcfc1e5-8927-4b4b-bd19-234497f86f23\") " pod="openshift-marketplace/redhat-marketplace-jxqbm" Nov 21 14:13:18 crc kubenswrapper[4897]: I1121 14:13:18.394154 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-brhkc\" (UniqueName: \"kubernetes.io/projected/0fcfc1e5-8927-4b4b-bd19-234497f86f23-kube-api-access-brhkc\") pod \"redhat-marketplace-jxqbm\" (UID: \"0fcfc1e5-8927-4b4b-bd19-234497f86f23\") " pod="openshift-marketplace/redhat-marketplace-jxqbm" Nov 21 14:13:18 crc kubenswrapper[4897]: I1121 14:13:18.394183 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0fcfc1e5-8927-4b4b-bd19-234497f86f23-catalog-content\") pod \"redhat-marketplace-jxqbm\" (UID: \"0fcfc1e5-8927-4b4b-bd19-234497f86f23\") " pod="openshift-marketplace/redhat-marketplace-jxqbm" Nov 21 14:13:18 crc kubenswrapper[4897]: I1121 14:13:18.394692 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0fcfc1e5-8927-4b4b-bd19-234497f86f23-catalog-content\") pod \"redhat-marketplace-jxqbm\" (UID: \"0fcfc1e5-8927-4b4b-bd19-234497f86f23\") " pod="openshift-marketplace/redhat-marketplace-jxqbm" Nov 21 14:13:18 crc kubenswrapper[4897]: I1121 14:13:18.394802 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0fcfc1e5-8927-4b4b-bd19-234497f86f23-utilities\") pod \"redhat-marketplace-jxqbm\" (UID: \"0fcfc1e5-8927-4b4b-bd19-234497f86f23\") " pod="openshift-marketplace/redhat-marketplace-jxqbm" Nov 21 14:13:18 crc kubenswrapper[4897]: I1121 14:13:18.412351 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-brhkc\" (UniqueName: \"kubernetes.io/projected/0fcfc1e5-8927-4b4b-bd19-234497f86f23-kube-api-access-brhkc\") pod \"redhat-marketplace-jxqbm\" (UID: \"0fcfc1e5-8927-4b4b-bd19-234497f86f23\") " pod="openshift-marketplace/redhat-marketplace-jxqbm" Nov 21 14:13:18 crc kubenswrapper[4897]: I1121 14:13:18.487751 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-7b657"] Nov 21 14:13:18 crc kubenswrapper[4897]: I1121 14:13:18.489012 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7b657" Nov 21 14:13:18 crc kubenswrapper[4897]: I1121 14:13:18.492837 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7b657"] Nov 21 14:13:18 crc kubenswrapper[4897]: I1121 14:13:18.495983 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 21 14:13:18 crc kubenswrapper[4897]: I1121 14:13:18.595932 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x9gbx\" (UniqueName: \"kubernetes.io/projected/d3417790-7956-4631-8c0c-f7959ce0ba19-kube-api-access-x9gbx\") pod \"redhat-operators-7b657\" (UID: \"d3417790-7956-4631-8c0c-f7959ce0ba19\") " pod="openshift-marketplace/redhat-operators-7b657" Nov 21 14:13:18 crc kubenswrapper[4897]: I1121 14:13:18.595973 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3417790-7956-4631-8c0c-f7959ce0ba19-utilities\") pod \"redhat-operators-7b657\" (UID: \"d3417790-7956-4631-8c0c-f7959ce0ba19\") " pod="openshift-marketplace/redhat-operators-7b657" Nov 21 14:13:18 crc kubenswrapper[4897]: I1121 14:13:18.596014 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3417790-7956-4631-8c0c-f7959ce0ba19-catalog-content\") pod \"redhat-operators-7b657\" (UID: \"d3417790-7956-4631-8c0c-f7959ce0ba19\") " pod="openshift-marketplace/redhat-operators-7b657" Nov 21 14:13:18 crc kubenswrapper[4897]: I1121 14:13:18.604405 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jxqbm" Nov 21 14:13:18 crc kubenswrapper[4897]: I1121 14:13:18.697209 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x9gbx\" (UniqueName: \"kubernetes.io/projected/d3417790-7956-4631-8c0c-f7959ce0ba19-kube-api-access-x9gbx\") pod \"redhat-operators-7b657\" (UID: \"d3417790-7956-4631-8c0c-f7959ce0ba19\") " pod="openshift-marketplace/redhat-operators-7b657" Nov 21 14:13:18 crc kubenswrapper[4897]: I1121 14:13:18.697269 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3417790-7956-4631-8c0c-f7959ce0ba19-utilities\") pod \"redhat-operators-7b657\" (UID: \"d3417790-7956-4631-8c0c-f7959ce0ba19\") " pod="openshift-marketplace/redhat-operators-7b657" Nov 21 14:13:18 crc kubenswrapper[4897]: I1121 14:13:18.697312 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3417790-7956-4631-8c0c-f7959ce0ba19-catalog-content\") pod \"redhat-operators-7b657\" (UID: \"d3417790-7956-4631-8c0c-f7959ce0ba19\") " pod="openshift-marketplace/redhat-operators-7b657" Nov 21 14:13:18 crc kubenswrapper[4897]: I1121 14:13:18.697999 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3417790-7956-4631-8c0c-f7959ce0ba19-catalog-content\") pod \"redhat-operators-7b657\" (UID: \"d3417790-7956-4631-8c0c-f7959ce0ba19\") " pod="openshift-marketplace/redhat-operators-7b657" Nov 21 14:13:18 crc kubenswrapper[4897]: I1121 14:13:18.698076 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3417790-7956-4631-8c0c-f7959ce0ba19-utilities\") pod \"redhat-operators-7b657\" (UID: \"d3417790-7956-4631-8c0c-f7959ce0ba19\") " pod="openshift-marketplace/redhat-operators-7b657" Nov 21 14:13:18 crc kubenswrapper[4897]: I1121 14:13:18.713991 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x9gbx\" (UniqueName: \"kubernetes.io/projected/d3417790-7956-4631-8c0c-f7959ce0ba19-kube-api-access-x9gbx\") pod \"redhat-operators-7b657\" (UID: \"d3417790-7956-4631-8c0c-f7959ce0ba19\") " pod="openshift-marketplace/redhat-operators-7b657" Nov 21 14:13:18 crc kubenswrapper[4897]: I1121 14:13:18.804682 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7b657" Nov 21 14:13:18 crc kubenswrapper[4897]: I1121 14:13:18.995115 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jxqbm"] Nov 21 14:13:19 crc kubenswrapper[4897]: I1121 14:13:19.173169 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7b657"] Nov 21 14:13:19 crc kubenswrapper[4897]: W1121 14:13:19.186483 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd3417790_7956_4631_8c0c_f7959ce0ba19.slice/crio-d6af59ab7741c7d5e178acb46d1b21d57aa31fd8f79c5141cbfb6ac10e744023 WatchSource:0}: Error finding container d6af59ab7741c7d5e178acb46d1b21d57aa31fd8f79c5141cbfb6ac10e744023: Status 404 returned error can't find the container with id d6af59ab7741c7d5e178acb46d1b21d57aa31fd8f79c5141cbfb6ac10e744023 Nov 21 14:13:19 crc kubenswrapper[4897]: I1121 14:13:19.494721 4897 generic.go:334] "Generic (PLEG): container finished" podID="0fcfc1e5-8927-4b4b-bd19-234497f86f23" containerID="f1d181b5d1bb77757c55fc0b401ef9cf62779ad6b646c99af3914b2f9971f880" exitCode=0 Nov 21 14:13:19 crc kubenswrapper[4897]: I1121 14:13:19.494801 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jxqbm" event={"ID":"0fcfc1e5-8927-4b4b-bd19-234497f86f23","Type":"ContainerDied","Data":"f1d181b5d1bb77757c55fc0b401ef9cf62779ad6b646c99af3914b2f9971f880"} Nov 21 14:13:19 crc kubenswrapper[4897]: I1121 14:13:19.494831 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jxqbm" event={"ID":"0fcfc1e5-8927-4b4b-bd19-234497f86f23","Type":"ContainerStarted","Data":"c880a91a9938f1411e9dc72f37eb88c2196d9154117c30acee4918dd044c0c5f"} Nov 21 14:13:19 crc kubenswrapper[4897]: I1121 14:13:19.499646 4897 generic.go:334] "Generic (PLEG): container finished" podID="d3417790-7956-4631-8c0c-f7959ce0ba19" containerID="4baa0e8a36ccbc33fce8a90a6ad7d5694864623c385d504f4b6bc18a87b56afd" exitCode=0 Nov 21 14:13:19 crc kubenswrapper[4897]: I1121 14:13:19.499705 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7b657" event={"ID":"d3417790-7956-4631-8c0c-f7959ce0ba19","Type":"ContainerDied","Data":"4baa0e8a36ccbc33fce8a90a6ad7d5694864623c385d504f4b6bc18a87b56afd"} Nov 21 14:13:19 crc kubenswrapper[4897]: I1121 14:13:19.499735 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7b657" event={"ID":"d3417790-7956-4631-8c0c-f7959ce0ba19","Type":"ContainerStarted","Data":"d6af59ab7741c7d5e178acb46d1b21d57aa31fd8f79c5141cbfb6ac10e744023"} Nov 21 14:13:20 crc kubenswrapper[4897]: I1121 14:13:20.679273 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-gt47r"] Nov 21 14:13:20 crc kubenswrapper[4897]: I1121 14:13:20.680772 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gt47r" Nov 21 14:13:20 crc kubenswrapper[4897]: I1121 14:13:20.682540 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 21 14:13:20 crc kubenswrapper[4897]: I1121 14:13:20.702491 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gt47r"] Nov 21 14:13:20 crc kubenswrapper[4897]: I1121 14:13:20.722649 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c288735-0f99-463c-bdce-0dc7e3a1a903-catalog-content\") pod \"community-operators-gt47r\" (UID: \"3c288735-0f99-463c-bdce-0dc7e3a1a903\") " pod="openshift-marketplace/community-operators-gt47r" Nov 21 14:13:20 crc kubenswrapper[4897]: I1121 14:13:20.722709 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c288735-0f99-463c-bdce-0dc7e3a1a903-utilities\") pod \"community-operators-gt47r\" (UID: \"3c288735-0f99-463c-bdce-0dc7e3a1a903\") " pod="openshift-marketplace/community-operators-gt47r" Nov 21 14:13:20 crc kubenswrapper[4897]: I1121 14:13:20.722775 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8nfj2\" (UniqueName: \"kubernetes.io/projected/3c288735-0f99-463c-bdce-0dc7e3a1a903-kube-api-access-8nfj2\") pod \"community-operators-gt47r\" (UID: \"3c288735-0f99-463c-bdce-0dc7e3a1a903\") " pod="openshift-marketplace/community-operators-gt47r" Nov 21 14:13:20 crc kubenswrapper[4897]: I1121 14:13:20.823174 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c288735-0f99-463c-bdce-0dc7e3a1a903-utilities\") pod \"community-operators-gt47r\" (UID: \"3c288735-0f99-463c-bdce-0dc7e3a1a903\") " pod="openshift-marketplace/community-operators-gt47r" Nov 21 14:13:20 crc kubenswrapper[4897]: I1121 14:13:20.823241 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8nfj2\" (UniqueName: \"kubernetes.io/projected/3c288735-0f99-463c-bdce-0dc7e3a1a903-kube-api-access-8nfj2\") pod \"community-operators-gt47r\" (UID: \"3c288735-0f99-463c-bdce-0dc7e3a1a903\") " pod="openshift-marketplace/community-operators-gt47r" Nov 21 14:13:20 crc kubenswrapper[4897]: I1121 14:13:20.823280 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c288735-0f99-463c-bdce-0dc7e3a1a903-catalog-content\") pod \"community-operators-gt47r\" (UID: \"3c288735-0f99-463c-bdce-0dc7e3a1a903\") " pod="openshift-marketplace/community-operators-gt47r" Nov 21 14:13:20 crc kubenswrapper[4897]: I1121 14:13:20.824110 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c288735-0f99-463c-bdce-0dc7e3a1a903-utilities\") pod \"community-operators-gt47r\" (UID: \"3c288735-0f99-463c-bdce-0dc7e3a1a903\") " pod="openshift-marketplace/community-operators-gt47r" Nov 21 14:13:20 crc kubenswrapper[4897]: I1121 14:13:20.824311 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c288735-0f99-463c-bdce-0dc7e3a1a903-catalog-content\") pod \"community-operators-gt47r\" (UID: \"3c288735-0f99-463c-bdce-0dc7e3a1a903\") " pod="openshift-marketplace/community-operators-gt47r" Nov 21 14:13:20 crc kubenswrapper[4897]: I1121 14:13:20.841312 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8nfj2\" (UniqueName: \"kubernetes.io/projected/3c288735-0f99-463c-bdce-0dc7e3a1a903-kube-api-access-8nfj2\") pod \"community-operators-gt47r\" (UID: \"3c288735-0f99-463c-bdce-0dc7e3a1a903\") " pod="openshift-marketplace/community-operators-gt47r" Nov 21 14:13:20 crc kubenswrapper[4897]: I1121 14:13:20.879240 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-s8gkj"] Nov 21 14:13:20 crc kubenswrapper[4897]: I1121 14:13:20.880994 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-s8gkj" Nov 21 14:13:20 crc kubenswrapper[4897]: I1121 14:13:20.883367 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 21 14:13:20 crc kubenswrapper[4897]: I1121 14:13:20.900236 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-s8gkj"] Nov 21 14:13:20 crc kubenswrapper[4897]: I1121 14:13:20.924568 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ms7xg\" (UniqueName: \"kubernetes.io/projected/955d92b4-cea8-4fba-86d9-f50592c64e89-kube-api-access-ms7xg\") pod \"certified-operators-s8gkj\" (UID: \"955d92b4-cea8-4fba-86d9-f50592c64e89\") " pod="openshift-marketplace/certified-operators-s8gkj" Nov 21 14:13:20 crc kubenswrapper[4897]: I1121 14:13:20.924644 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/955d92b4-cea8-4fba-86d9-f50592c64e89-utilities\") pod \"certified-operators-s8gkj\" (UID: \"955d92b4-cea8-4fba-86d9-f50592c64e89\") " pod="openshift-marketplace/certified-operators-s8gkj" Nov 21 14:13:20 crc kubenswrapper[4897]: I1121 14:13:20.924910 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/955d92b4-cea8-4fba-86d9-f50592c64e89-catalog-content\") pod \"certified-operators-s8gkj\" (UID: \"955d92b4-cea8-4fba-86d9-f50592c64e89\") " pod="openshift-marketplace/certified-operators-s8gkj" Nov 21 14:13:21 crc kubenswrapper[4897]: I1121 14:13:21.025728 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ms7xg\" (UniqueName: \"kubernetes.io/projected/955d92b4-cea8-4fba-86d9-f50592c64e89-kube-api-access-ms7xg\") pod \"certified-operators-s8gkj\" (UID: \"955d92b4-cea8-4fba-86d9-f50592c64e89\") " pod="openshift-marketplace/certified-operators-s8gkj" Nov 21 14:13:21 crc kubenswrapper[4897]: I1121 14:13:21.025795 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/955d92b4-cea8-4fba-86d9-f50592c64e89-utilities\") pod \"certified-operators-s8gkj\" (UID: \"955d92b4-cea8-4fba-86d9-f50592c64e89\") " pod="openshift-marketplace/certified-operators-s8gkj" Nov 21 14:13:21 crc kubenswrapper[4897]: I1121 14:13:21.025851 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/955d92b4-cea8-4fba-86d9-f50592c64e89-catalog-content\") pod \"certified-operators-s8gkj\" (UID: \"955d92b4-cea8-4fba-86d9-f50592c64e89\") " pod="openshift-marketplace/certified-operators-s8gkj" Nov 21 14:13:21 crc kubenswrapper[4897]: I1121 14:13:21.026363 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/955d92b4-cea8-4fba-86d9-f50592c64e89-catalog-content\") pod \"certified-operators-s8gkj\" (UID: \"955d92b4-cea8-4fba-86d9-f50592c64e89\") " pod="openshift-marketplace/certified-operators-s8gkj" Nov 21 14:13:21 crc kubenswrapper[4897]: I1121 14:13:21.026622 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/955d92b4-cea8-4fba-86d9-f50592c64e89-utilities\") pod \"certified-operators-s8gkj\" (UID: \"955d92b4-cea8-4fba-86d9-f50592c64e89\") " pod="openshift-marketplace/certified-operators-s8gkj" Nov 21 14:13:21 crc kubenswrapper[4897]: I1121 14:13:21.043047 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ms7xg\" (UniqueName: \"kubernetes.io/projected/955d92b4-cea8-4fba-86d9-f50592c64e89-kube-api-access-ms7xg\") pod \"certified-operators-s8gkj\" (UID: \"955d92b4-cea8-4fba-86d9-f50592c64e89\") " pod="openshift-marketplace/certified-operators-s8gkj" Nov 21 14:13:21 crc kubenswrapper[4897]: I1121 14:13:21.105363 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gt47r" Nov 21 14:13:21 crc kubenswrapper[4897]: I1121 14:13:21.203487 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-s8gkj" Nov 21 14:13:21 crc kubenswrapper[4897]: I1121 14:13:21.311086 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gt47r"] Nov 21 14:13:21 crc kubenswrapper[4897]: W1121 14:13:21.322949 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3c288735_0f99_463c_bdce_0dc7e3a1a903.slice/crio-b16ec0342bc101207d63ea9e92e46b7fa347bcc16f1e9d9010d826a8af2e4edc WatchSource:0}: Error finding container b16ec0342bc101207d63ea9e92e46b7fa347bcc16f1e9d9010d826a8af2e4edc: Status 404 returned error can't find the container with id b16ec0342bc101207d63ea9e92e46b7fa347bcc16f1e9d9010d826a8af2e4edc Nov 21 14:13:21 crc kubenswrapper[4897]: I1121 14:13:21.514747 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7b657" event={"ID":"d3417790-7956-4631-8c0c-f7959ce0ba19","Type":"ContainerDied","Data":"2e5f575f860d2a1e1b28486675495d09e26d0276ec597092353aec43eba6bf09"} Nov 21 14:13:21 crc kubenswrapper[4897]: I1121 14:13:21.516496 4897 generic.go:334] "Generic (PLEG): container finished" podID="d3417790-7956-4631-8c0c-f7959ce0ba19" containerID="2e5f575f860d2a1e1b28486675495d09e26d0276ec597092353aec43eba6bf09" exitCode=0 Nov 21 14:13:21 crc kubenswrapper[4897]: I1121 14:13:21.518768 4897 generic.go:334] "Generic (PLEG): container finished" podID="0fcfc1e5-8927-4b4b-bd19-234497f86f23" containerID="84acc2443addebb5dfe61af1cc8b0733ccbae80bf6044fd9e7ee713280e83e21" exitCode=0 Nov 21 14:13:21 crc kubenswrapper[4897]: I1121 14:13:21.518824 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jxqbm" event={"ID":"0fcfc1e5-8927-4b4b-bd19-234497f86f23","Type":"ContainerDied","Data":"84acc2443addebb5dfe61af1cc8b0733ccbae80bf6044fd9e7ee713280e83e21"} Nov 21 14:13:21 crc kubenswrapper[4897]: I1121 14:13:21.524591 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gt47r" event={"ID":"3c288735-0f99-463c-bdce-0dc7e3a1a903","Type":"ContainerStarted","Data":"b16ec0342bc101207d63ea9e92e46b7fa347bcc16f1e9d9010d826a8af2e4edc"} Nov 21 14:13:21 crc kubenswrapper[4897]: I1121 14:13:21.615128 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-s8gkj"] Nov 21 14:13:21 crc kubenswrapper[4897]: W1121 14:13:21.629442 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod955d92b4_cea8_4fba_86d9_f50592c64e89.slice/crio-ed261d9b588edca1b0cbe5e4585adfa7a1858c0d689d7191661719eea5055738 WatchSource:0}: Error finding container ed261d9b588edca1b0cbe5e4585adfa7a1858c0d689d7191661719eea5055738: Status 404 returned error can't find the container with id ed261d9b588edca1b0cbe5e4585adfa7a1858c0d689d7191661719eea5055738 Nov 21 14:13:22 crc kubenswrapper[4897]: I1121 14:13:22.536582 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jxqbm" event={"ID":"0fcfc1e5-8927-4b4b-bd19-234497f86f23","Type":"ContainerStarted","Data":"13b339b43dab4f7265598d2226cd45144a0ee3283134514b44e3bd205fb9807d"} Nov 21 14:13:22 crc kubenswrapper[4897]: I1121 14:13:22.540403 4897 generic.go:334] "Generic (PLEG): container finished" podID="955d92b4-cea8-4fba-86d9-f50592c64e89" containerID="c5e4e8cdc6791bbb529c7951647ea43868ac85dde3649dd3d076f06c445303bd" exitCode=0 Nov 21 14:13:22 crc kubenswrapper[4897]: I1121 14:13:22.540490 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s8gkj" event={"ID":"955d92b4-cea8-4fba-86d9-f50592c64e89","Type":"ContainerDied","Data":"c5e4e8cdc6791bbb529c7951647ea43868ac85dde3649dd3d076f06c445303bd"} Nov 21 14:13:22 crc kubenswrapper[4897]: I1121 14:13:22.540553 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s8gkj" event={"ID":"955d92b4-cea8-4fba-86d9-f50592c64e89","Type":"ContainerStarted","Data":"ed261d9b588edca1b0cbe5e4585adfa7a1858c0d689d7191661719eea5055738"} Nov 21 14:13:22 crc kubenswrapper[4897]: I1121 14:13:22.544895 4897 generic.go:334] "Generic (PLEG): container finished" podID="3c288735-0f99-463c-bdce-0dc7e3a1a903" containerID="8ae46e33df286596c0c8be819d3f279cf8f968d6017c7370b20e04274d1c19ac" exitCode=0 Nov 21 14:13:22 crc kubenswrapper[4897]: I1121 14:13:22.544975 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gt47r" event={"ID":"3c288735-0f99-463c-bdce-0dc7e3a1a903","Type":"ContainerDied","Data":"8ae46e33df286596c0c8be819d3f279cf8f968d6017c7370b20e04274d1c19ac"} Nov 21 14:13:22 crc kubenswrapper[4897]: I1121 14:13:22.549826 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7b657" event={"ID":"d3417790-7956-4631-8c0c-f7959ce0ba19","Type":"ContainerStarted","Data":"01e7f9db9921b11bb7a3c7c47b7cff60a8f3e6d4bfb17e272076667cd190fbe6"} Nov 21 14:13:22 crc kubenswrapper[4897]: I1121 14:13:22.554666 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-jxqbm" podStartSLOduration=2.145443651 podStartE2EDuration="4.554652331s" podCreationTimestamp="2025-11-21 14:13:18 +0000 UTC" firstStartedPulling="2025-11-21 14:13:19.497187663 +0000 UTC m=+276.781781138" lastFinishedPulling="2025-11-21 14:13:21.906396343 +0000 UTC m=+279.190989818" observedRunningTime="2025-11-21 14:13:22.552337181 +0000 UTC m=+279.836930656" watchObservedRunningTime="2025-11-21 14:13:22.554652331 +0000 UTC m=+279.839245806" Nov 21 14:13:22 crc kubenswrapper[4897]: I1121 14:13:22.583117 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-7b657" podStartSLOduration=2.147336308 podStartE2EDuration="4.583096517s" podCreationTimestamp="2025-11-21 14:13:18 +0000 UTC" firstStartedPulling="2025-11-21 14:13:19.501037209 +0000 UTC m=+276.785630684" lastFinishedPulling="2025-11-21 14:13:21.936797418 +0000 UTC m=+279.221390893" observedRunningTime="2025-11-21 14:13:22.568058885 +0000 UTC m=+279.852652360" watchObservedRunningTime="2025-11-21 14:13:22.583096517 +0000 UTC m=+279.867689992" Nov 21 14:13:23 crc kubenswrapper[4897]: I1121 14:13:23.557476 4897 generic.go:334] "Generic (PLEG): container finished" podID="3c288735-0f99-463c-bdce-0dc7e3a1a903" containerID="6aeccfd3199065d071f87108e6b98547b05ec2d5b9a6b542bc3be7c877cbc8b4" exitCode=0 Nov 21 14:13:23 crc kubenswrapper[4897]: I1121 14:13:23.557553 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gt47r" event={"ID":"3c288735-0f99-463c-bdce-0dc7e3a1a903","Type":"ContainerDied","Data":"6aeccfd3199065d071f87108e6b98547b05ec2d5b9a6b542bc3be7c877cbc8b4"} Nov 21 14:13:25 crc kubenswrapper[4897]: I1121 14:13:25.568135 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gt47r" event={"ID":"3c288735-0f99-463c-bdce-0dc7e3a1a903","Type":"ContainerStarted","Data":"c0343a6b11e78cbe1a8c243b88fbe29cb39ff5f1e40907ae95e8cb4cb6c136a3"} Nov 21 14:13:25 crc kubenswrapper[4897]: I1121 14:13:25.570288 4897 generic.go:334] "Generic (PLEG): container finished" podID="955d92b4-cea8-4fba-86d9-f50592c64e89" containerID="925c1b71ddce1126ef80ba78bfbeb43985caf8f035dbf2f81852978757acba45" exitCode=0 Nov 21 14:13:25 crc kubenswrapper[4897]: I1121 14:13:25.570318 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s8gkj" event={"ID":"955d92b4-cea8-4fba-86d9-f50592c64e89","Type":"ContainerDied","Data":"925c1b71ddce1126ef80ba78bfbeb43985caf8f035dbf2f81852978757acba45"} Nov 21 14:13:25 crc kubenswrapper[4897]: I1121 14:13:25.587139 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-gt47r" podStartSLOduration=2.290705663 podStartE2EDuration="5.587124366s" podCreationTimestamp="2025-11-21 14:13:20 +0000 UTC" firstStartedPulling="2025-11-21 14:13:21.525906886 +0000 UTC m=+278.810500361" lastFinishedPulling="2025-11-21 14:13:24.822325589 +0000 UTC m=+282.106919064" observedRunningTime="2025-11-21 14:13:25.585617651 +0000 UTC m=+282.870211126" watchObservedRunningTime="2025-11-21 14:13:25.587124366 +0000 UTC m=+282.871717841" Nov 21 14:13:26 crc kubenswrapper[4897]: I1121 14:13:26.578294 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s8gkj" event={"ID":"955d92b4-cea8-4fba-86d9-f50592c64e89","Type":"ContainerStarted","Data":"738f92462444b894cabeca211d5d0f35e618855760fec57a2d2d1e0bbbc278bc"} Nov 21 14:13:28 crc kubenswrapper[4897]: I1121 14:13:28.604716 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-s8gkj" podStartSLOduration=5.104308289 podStartE2EDuration="8.604698173s" podCreationTimestamp="2025-11-21 14:13:20 +0000 UTC" firstStartedPulling="2025-11-21 14:13:22.541836005 +0000 UTC m=+279.826429480" lastFinishedPulling="2025-11-21 14:13:26.042225889 +0000 UTC m=+283.326819364" observedRunningTime="2025-11-21 14:13:28.604518687 +0000 UTC m=+285.889112182" watchObservedRunningTime="2025-11-21 14:13:28.604698173 +0000 UTC m=+285.889291648" Nov 21 14:13:28 crc kubenswrapper[4897]: I1121 14:13:28.605607 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-jxqbm" Nov 21 14:13:28 crc kubenswrapper[4897]: I1121 14:13:28.605667 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-jxqbm" Nov 21 14:13:28 crc kubenswrapper[4897]: I1121 14:13:28.651922 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-jxqbm" Nov 21 14:13:28 crc kubenswrapper[4897]: I1121 14:13:28.805499 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-7b657" Nov 21 14:13:28 crc kubenswrapper[4897]: I1121 14:13:28.805578 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-7b657" Nov 21 14:13:28 crc kubenswrapper[4897]: I1121 14:13:28.857953 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-7b657" Nov 21 14:13:29 crc kubenswrapper[4897]: I1121 14:13:29.638862 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-jxqbm" Nov 21 14:13:29 crc kubenswrapper[4897]: I1121 14:13:29.644212 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-7b657" Nov 21 14:13:31 crc kubenswrapper[4897]: I1121 14:13:31.105548 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-gt47r" Nov 21 14:13:31 crc kubenswrapper[4897]: I1121 14:13:31.105881 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-gt47r" Nov 21 14:13:31 crc kubenswrapper[4897]: I1121 14:13:31.143997 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-gt47r" Nov 21 14:13:31 crc kubenswrapper[4897]: I1121 14:13:31.204431 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-s8gkj" Nov 21 14:13:31 crc kubenswrapper[4897]: I1121 14:13:31.204582 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-s8gkj" Nov 21 14:13:31 crc kubenswrapper[4897]: I1121 14:13:31.239713 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-s8gkj" Nov 21 14:13:31 crc kubenswrapper[4897]: I1121 14:13:31.643498 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-gt47r" Nov 21 14:13:31 crc kubenswrapper[4897]: I1121 14:13:31.656917 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-s8gkj" Nov 21 14:13:46 crc kubenswrapper[4897]: I1121 14:13:46.990910 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/cluster-monitoring-operator-6d5b84845-c5jbm"] Nov 21 14:13:46 crc kubenswrapper[4897]: I1121 14:13:46.996846 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-c5jbm" Nov 21 14:13:46 crc kubenswrapper[4897]: I1121 14:13:46.999162 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"openshift-service-ca.crt" Nov 21 14:13:46 crc kubenswrapper[4897]: I1121 14:13:46.999294 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"kube-root-ca.crt" Nov 21 14:13:46 crc kubenswrapper[4897]: I1121 14:13:46.999380 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"cluster-monitoring-operator-tls" Nov 21 14:13:46 crc kubenswrapper[4897]: I1121 14:13:46.999621 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"cluster-monitoring-operator-dockercfg-wwt9l" Nov 21 14:13:46 crc kubenswrapper[4897]: I1121 14:13:46.999791 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"telemetry-config" Nov 21 14:13:47 crc kubenswrapper[4897]: I1121 14:13:47.004396 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/cluster-monitoring-operator-6d5b84845-c5jbm"] Nov 21 14:13:47 crc kubenswrapper[4897]: I1121 14:13:47.168201 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-config\" (UniqueName: \"kubernetes.io/configmap/2685b6da-fab9-4e9f-a3df-c1653b8edfc6-telemetry-config\") pod \"cluster-monitoring-operator-6d5b84845-c5jbm\" (UID: \"2685b6da-fab9-4e9f-a3df-c1653b8edfc6\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-c5jbm" Nov 21 14:13:47 crc kubenswrapper[4897]: I1121 14:13:47.168256 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vl8tk\" (UniqueName: \"kubernetes.io/projected/2685b6da-fab9-4e9f-a3df-c1653b8edfc6-kube-api-access-vl8tk\") pod \"cluster-monitoring-operator-6d5b84845-c5jbm\" (UID: \"2685b6da-fab9-4e9f-a3df-c1653b8edfc6\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-c5jbm" Nov 21 14:13:47 crc kubenswrapper[4897]: I1121 14:13:47.168284 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cluster-monitoring-operator-tls\" (UniqueName: \"kubernetes.io/secret/2685b6da-fab9-4e9f-a3df-c1653b8edfc6-cluster-monitoring-operator-tls\") pod \"cluster-monitoring-operator-6d5b84845-c5jbm\" (UID: \"2685b6da-fab9-4e9f-a3df-c1653b8edfc6\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-c5jbm" Nov 21 14:13:47 crc kubenswrapper[4897]: I1121 14:13:47.269587 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-config\" (UniqueName: \"kubernetes.io/configmap/2685b6da-fab9-4e9f-a3df-c1653b8edfc6-telemetry-config\") pod \"cluster-monitoring-operator-6d5b84845-c5jbm\" (UID: \"2685b6da-fab9-4e9f-a3df-c1653b8edfc6\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-c5jbm" Nov 21 14:13:47 crc kubenswrapper[4897]: I1121 14:13:47.269676 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vl8tk\" (UniqueName: \"kubernetes.io/projected/2685b6da-fab9-4e9f-a3df-c1653b8edfc6-kube-api-access-vl8tk\") pod \"cluster-monitoring-operator-6d5b84845-c5jbm\" (UID: \"2685b6da-fab9-4e9f-a3df-c1653b8edfc6\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-c5jbm" Nov 21 14:13:47 crc kubenswrapper[4897]: I1121 14:13:47.270824 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-config\" (UniqueName: \"kubernetes.io/configmap/2685b6da-fab9-4e9f-a3df-c1653b8edfc6-telemetry-config\") pod \"cluster-monitoring-operator-6d5b84845-c5jbm\" (UID: \"2685b6da-fab9-4e9f-a3df-c1653b8edfc6\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-c5jbm" Nov 21 14:13:47 crc kubenswrapper[4897]: I1121 14:13:47.269717 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cluster-monitoring-operator-tls\" (UniqueName: \"kubernetes.io/secret/2685b6da-fab9-4e9f-a3df-c1653b8edfc6-cluster-monitoring-operator-tls\") pod \"cluster-monitoring-operator-6d5b84845-c5jbm\" (UID: \"2685b6da-fab9-4e9f-a3df-c1653b8edfc6\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-c5jbm" Nov 21 14:13:47 crc kubenswrapper[4897]: I1121 14:13:47.278434 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cluster-monitoring-operator-tls\" (UniqueName: \"kubernetes.io/secret/2685b6da-fab9-4e9f-a3df-c1653b8edfc6-cluster-monitoring-operator-tls\") pod \"cluster-monitoring-operator-6d5b84845-c5jbm\" (UID: \"2685b6da-fab9-4e9f-a3df-c1653b8edfc6\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-c5jbm" Nov 21 14:13:47 crc kubenswrapper[4897]: I1121 14:13:47.284593 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vl8tk\" (UniqueName: \"kubernetes.io/projected/2685b6da-fab9-4e9f-a3df-c1653b8edfc6-kube-api-access-vl8tk\") pod \"cluster-monitoring-operator-6d5b84845-c5jbm\" (UID: \"2685b6da-fab9-4e9f-a3df-c1653b8edfc6\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-c5jbm" Nov 21 14:13:47 crc kubenswrapper[4897]: I1121 14:13:47.316258 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-c5jbm" Nov 21 14:13:47 crc kubenswrapper[4897]: I1121 14:13:47.703342 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/cluster-monitoring-operator-6d5b84845-c5jbm"] Nov 21 14:13:47 crc kubenswrapper[4897]: W1121 14:13:47.710234 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2685b6da_fab9_4e9f_a3df_c1653b8edfc6.slice/crio-86d990a6b57d488c687973b79594464eeb23af9857568d253080f60229efd46c WatchSource:0}: Error finding container 86d990a6b57d488c687973b79594464eeb23af9857568d253080f60229efd46c: Status 404 returned error can't find the container with id 86d990a6b57d488c687973b79594464eeb23af9857568d253080f60229efd46c Nov 21 14:13:48 crc kubenswrapper[4897]: I1121 14:13:48.692734 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-c5jbm" event={"ID":"2685b6da-fab9-4e9f-a3df-c1653b8edfc6","Type":"ContainerStarted","Data":"86d990a6b57d488c687973b79594464eeb23af9857568d253080f60229efd46c"} Nov 21 14:13:50 crc kubenswrapper[4897]: I1121 14:13:50.229178 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-lsctq"] Nov 21 14:13:50 crc kubenswrapper[4897]: I1121 14:13:50.230098 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-lsctq" Nov 21 14:13:50 crc kubenswrapper[4897]: I1121 14:13:50.244212 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-lsctq"] Nov 21 14:13:50 crc kubenswrapper[4897]: I1121 14:13:50.347554 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-2sh55"] Nov 21 14:13:50 crc kubenswrapper[4897]: I1121 14:13:50.348366 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-2sh55" Nov 21 14:13:50 crc kubenswrapper[4897]: I1121 14:13:50.349850 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-operator-admission-webhook-dockercfg-vb5vk" Nov 21 14:13:50 crc kubenswrapper[4897]: I1121 14:13:50.350258 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-operator-admission-webhook-tls" Nov 21 14:13:50 crc kubenswrapper[4897]: I1121 14:13:50.356876 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-2sh55"] Nov 21 14:13:50 crc kubenswrapper[4897]: I1121 14:13:50.407417 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-lsctq\" (UID: \"71595772-afd3-4afd-aa27-3a501f39a1dc\") " pod="openshift-image-registry/image-registry-66df7c8f76-lsctq" Nov 21 14:13:50 crc kubenswrapper[4897]: I1121 14:13:50.407481 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6r9jz\" (UniqueName: \"kubernetes.io/projected/71595772-afd3-4afd-aa27-3a501f39a1dc-kube-api-access-6r9jz\") pod \"image-registry-66df7c8f76-lsctq\" (UID: \"71595772-afd3-4afd-aa27-3a501f39a1dc\") " pod="openshift-image-registry/image-registry-66df7c8f76-lsctq" Nov 21 14:13:50 crc kubenswrapper[4897]: I1121 14:13:50.407526 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/71595772-afd3-4afd-aa27-3a501f39a1dc-bound-sa-token\") pod \"image-registry-66df7c8f76-lsctq\" (UID: \"71595772-afd3-4afd-aa27-3a501f39a1dc\") " pod="openshift-image-registry/image-registry-66df7c8f76-lsctq" Nov 21 14:13:50 crc kubenswrapper[4897]: I1121 14:13:50.407552 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/71595772-afd3-4afd-aa27-3a501f39a1dc-registry-certificates\") pod \"image-registry-66df7c8f76-lsctq\" (UID: \"71595772-afd3-4afd-aa27-3a501f39a1dc\") " pod="openshift-image-registry/image-registry-66df7c8f76-lsctq" Nov 21 14:13:50 crc kubenswrapper[4897]: I1121 14:13:50.407573 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/71595772-afd3-4afd-aa27-3a501f39a1dc-registry-tls\") pod \"image-registry-66df7c8f76-lsctq\" (UID: \"71595772-afd3-4afd-aa27-3a501f39a1dc\") " pod="openshift-image-registry/image-registry-66df7c8f76-lsctq" Nov 21 14:13:50 crc kubenswrapper[4897]: I1121 14:13:50.407718 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/71595772-afd3-4afd-aa27-3a501f39a1dc-trusted-ca\") pod \"image-registry-66df7c8f76-lsctq\" (UID: \"71595772-afd3-4afd-aa27-3a501f39a1dc\") " pod="openshift-image-registry/image-registry-66df7c8f76-lsctq" Nov 21 14:13:50 crc kubenswrapper[4897]: I1121 14:13:50.407748 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/71595772-afd3-4afd-aa27-3a501f39a1dc-installation-pull-secrets\") pod \"image-registry-66df7c8f76-lsctq\" (UID: \"71595772-afd3-4afd-aa27-3a501f39a1dc\") " pod="openshift-image-registry/image-registry-66df7c8f76-lsctq" Nov 21 14:13:50 crc kubenswrapper[4897]: I1121 14:13:50.407779 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/71595772-afd3-4afd-aa27-3a501f39a1dc-ca-trust-extracted\") pod \"image-registry-66df7c8f76-lsctq\" (UID: \"71595772-afd3-4afd-aa27-3a501f39a1dc\") " pod="openshift-image-registry/image-registry-66df7c8f76-lsctq" Nov 21 14:13:50 crc kubenswrapper[4897]: I1121 14:13:50.455243 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-lsctq\" (UID: \"71595772-afd3-4afd-aa27-3a501f39a1dc\") " pod="openshift-image-registry/image-registry-66df7c8f76-lsctq" Nov 21 14:13:50 crc kubenswrapper[4897]: I1121 14:13:50.509318 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/71595772-afd3-4afd-aa27-3a501f39a1dc-ca-trust-extracted\") pod \"image-registry-66df7c8f76-lsctq\" (UID: \"71595772-afd3-4afd-aa27-3a501f39a1dc\") " pod="openshift-image-registry/image-registry-66df7c8f76-lsctq" Nov 21 14:13:50 crc kubenswrapper[4897]: I1121 14:13:50.509393 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6r9jz\" (UniqueName: \"kubernetes.io/projected/71595772-afd3-4afd-aa27-3a501f39a1dc-kube-api-access-6r9jz\") pod \"image-registry-66df7c8f76-lsctq\" (UID: \"71595772-afd3-4afd-aa27-3a501f39a1dc\") " pod="openshift-image-registry/image-registry-66df7c8f76-lsctq" Nov 21 14:13:50 crc kubenswrapper[4897]: I1121 14:13:50.509422 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/71595772-afd3-4afd-aa27-3a501f39a1dc-bound-sa-token\") pod \"image-registry-66df7c8f76-lsctq\" (UID: \"71595772-afd3-4afd-aa27-3a501f39a1dc\") " pod="openshift-image-registry/image-registry-66df7c8f76-lsctq" Nov 21 14:13:50 crc kubenswrapper[4897]: I1121 14:13:50.509441 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/71595772-afd3-4afd-aa27-3a501f39a1dc-registry-certificates\") pod \"image-registry-66df7c8f76-lsctq\" (UID: \"71595772-afd3-4afd-aa27-3a501f39a1dc\") " pod="openshift-image-registry/image-registry-66df7c8f76-lsctq" Nov 21 14:13:50 crc kubenswrapper[4897]: I1121 14:13:50.509463 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/71595772-afd3-4afd-aa27-3a501f39a1dc-registry-tls\") pod \"image-registry-66df7c8f76-lsctq\" (UID: \"71595772-afd3-4afd-aa27-3a501f39a1dc\") " pod="openshift-image-registry/image-registry-66df7c8f76-lsctq" Nov 21 14:13:50 crc kubenswrapper[4897]: I1121 14:13:50.509483 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/71595772-afd3-4afd-aa27-3a501f39a1dc-trusted-ca\") pod \"image-registry-66df7c8f76-lsctq\" (UID: \"71595772-afd3-4afd-aa27-3a501f39a1dc\") " pod="openshift-image-registry/image-registry-66df7c8f76-lsctq" Nov 21 14:13:50 crc kubenswrapper[4897]: I1121 14:13:50.509530 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/71595772-afd3-4afd-aa27-3a501f39a1dc-installation-pull-secrets\") pod \"image-registry-66df7c8f76-lsctq\" (UID: \"71595772-afd3-4afd-aa27-3a501f39a1dc\") " pod="openshift-image-registry/image-registry-66df7c8f76-lsctq" Nov 21 14:13:50 crc kubenswrapper[4897]: I1121 14:13:50.509563 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-certificates\" (UniqueName: \"kubernetes.io/secret/1ebe1a38-aff2-4347-9ca6-ccbffda18ac5-tls-certificates\") pod \"prometheus-operator-admission-webhook-f54c54754-2sh55\" (UID: \"1ebe1a38-aff2-4347-9ca6-ccbffda18ac5\") " pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-2sh55" Nov 21 14:13:50 crc kubenswrapper[4897]: I1121 14:13:50.509803 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/71595772-afd3-4afd-aa27-3a501f39a1dc-ca-trust-extracted\") pod \"image-registry-66df7c8f76-lsctq\" (UID: \"71595772-afd3-4afd-aa27-3a501f39a1dc\") " pod="openshift-image-registry/image-registry-66df7c8f76-lsctq" Nov 21 14:13:50 crc kubenswrapper[4897]: I1121 14:13:50.510879 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/71595772-afd3-4afd-aa27-3a501f39a1dc-trusted-ca\") pod \"image-registry-66df7c8f76-lsctq\" (UID: \"71595772-afd3-4afd-aa27-3a501f39a1dc\") " pod="openshift-image-registry/image-registry-66df7c8f76-lsctq" Nov 21 14:13:50 crc kubenswrapper[4897]: I1121 14:13:50.514161 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/71595772-afd3-4afd-aa27-3a501f39a1dc-registry-certificates\") pod \"image-registry-66df7c8f76-lsctq\" (UID: \"71595772-afd3-4afd-aa27-3a501f39a1dc\") " pod="openshift-image-registry/image-registry-66df7c8f76-lsctq" Nov 21 14:13:50 crc kubenswrapper[4897]: I1121 14:13:50.515598 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/71595772-afd3-4afd-aa27-3a501f39a1dc-registry-tls\") pod \"image-registry-66df7c8f76-lsctq\" (UID: \"71595772-afd3-4afd-aa27-3a501f39a1dc\") " pod="openshift-image-registry/image-registry-66df7c8f76-lsctq" Nov 21 14:13:50 crc kubenswrapper[4897]: I1121 14:13:50.516135 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/71595772-afd3-4afd-aa27-3a501f39a1dc-installation-pull-secrets\") pod \"image-registry-66df7c8f76-lsctq\" (UID: \"71595772-afd3-4afd-aa27-3a501f39a1dc\") " pod="openshift-image-registry/image-registry-66df7c8f76-lsctq" Nov 21 14:13:50 crc kubenswrapper[4897]: I1121 14:13:50.527104 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/71595772-afd3-4afd-aa27-3a501f39a1dc-bound-sa-token\") pod \"image-registry-66df7c8f76-lsctq\" (UID: \"71595772-afd3-4afd-aa27-3a501f39a1dc\") " pod="openshift-image-registry/image-registry-66df7c8f76-lsctq" Nov 21 14:13:50 crc kubenswrapper[4897]: I1121 14:13:50.527220 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6r9jz\" (UniqueName: \"kubernetes.io/projected/71595772-afd3-4afd-aa27-3a501f39a1dc-kube-api-access-6r9jz\") pod \"image-registry-66df7c8f76-lsctq\" (UID: \"71595772-afd3-4afd-aa27-3a501f39a1dc\") " pod="openshift-image-registry/image-registry-66df7c8f76-lsctq" Nov 21 14:13:50 crc kubenswrapper[4897]: I1121 14:13:50.544111 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-lsctq" Nov 21 14:13:50 crc kubenswrapper[4897]: I1121 14:13:50.610431 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-certificates\" (UniqueName: \"kubernetes.io/secret/1ebe1a38-aff2-4347-9ca6-ccbffda18ac5-tls-certificates\") pod \"prometheus-operator-admission-webhook-f54c54754-2sh55\" (UID: \"1ebe1a38-aff2-4347-9ca6-ccbffda18ac5\") " pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-2sh55" Nov 21 14:13:50 crc kubenswrapper[4897]: I1121 14:13:50.615996 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-certificates\" (UniqueName: \"kubernetes.io/secret/1ebe1a38-aff2-4347-9ca6-ccbffda18ac5-tls-certificates\") pod \"prometheus-operator-admission-webhook-f54c54754-2sh55\" (UID: \"1ebe1a38-aff2-4347-9ca6-ccbffda18ac5\") " pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-2sh55" Nov 21 14:13:50 crc kubenswrapper[4897]: I1121 14:13:50.659588 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-2sh55" Nov 21 14:13:50 crc kubenswrapper[4897]: I1121 14:13:50.703397 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-c5jbm" event={"ID":"2685b6da-fab9-4e9f-a3df-c1653b8edfc6","Type":"ContainerStarted","Data":"6eb097c6de29a87d9a4efa4b7e660810067d26670db8e8c1c1f91ebaa95157f2"} Nov 21 14:13:50 crc kubenswrapper[4897]: I1121 14:13:50.717023 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-c5jbm" podStartSLOduration=2.725267714 podStartE2EDuration="4.716993133s" podCreationTimestamp="2025-11-21 14:13:46 +0000 UTC" firstStartedPulling="2025-11-21 14:13:47.712318135 +0000 UTC m=+304.996911610" lastFinishedPulling="2025-11-21 14:13:49.704043554 +0000 UTC m=+306.988637029" observedRunningTime="2025-11-21 14:13:50.715480228 +0000 UTC m=+308.000073713" watchObservedRunningTime="2025-11-21 14:13:50.716993133 +0000 UTC m=+308.001586608" Nov 21 14:13:50 crc kubenswrapper[4897]: I1121 14:13:50.749858 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-lsctq"] Nov 21 14:13:50 crc kubenswrapper[4897]: I1121 14:13:50.845751 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-2sh55"] Nov 21 14:13:50 crc kubenswrapper[4897]: W1121 14:13:50.865538 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1ebe1a38_aff2_4347_9ca6_ccbffda18ac5.slice/crio-3387324a5379cdd2cdda33114e610c0ffc5e819af5bb2a22eb948543b4431ec1 WatchSource:0}: Error finding container 3387324a5379cdd2cdda33114e610c0ffc5e819af5bb2a22eb948543b4431ec1: Status 404 returned error can't find the container with id 3387324a5379cdd2cdda33114e610c0ffc5e819af5bb2a22eb948543b4431ec1 Nov 21 14:13:51 crc kubenswrapper[4897]: I1121 14:13:51.708108 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-2sh55" event={"ID":"1ebe1a38-aff2-4347-9ca6-ccbffda18ac5","Type":"ContainerStarted","Data":"3387324a5379cdd2cdda33114e610c0ffc5e819af5bb2a22eb948543b4431ec1"} Nov 21 14:13:51 crc kubenswrapper[4897]: I1121 14:13:51.709552 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-lsctq" event={"ID":"71595772-afd3-4afd-aa27-3a501f39a1dc","Type":"ContainerStarted","Data":"deb0a3dbeca059ba610a374458894027e7a81e9e6233653400faea19626dc7ee"} Nov 21 14:13:51 crc kubenswrapper[4897]: I1121 14:13:51.709850 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-lsctq" Nov 21 14:13:51 crc kubenswrapper[4897]: I1121 14:13:51.709882 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-lsctq" event={"ID":"71595772-afd3-4afd-aa27-3a501f39a1dc","Type":"ContainerStarted","Data":"5d12d09b52812d2ebe9d2815b319039fd7402d9946051f0576eff26ac1356278"} Nov 21 14:13:52 crc kubenswrapper[4897]: I1121 14:13:52.716884 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-2sh55" event={"ID":"1ebe1a38-aff2-4347-9ca6-ccbffda18ac5","Type":"ContainerStarted","Data":"49e480269cff6b5fcce3580709272ca1b51f3e749f471e18e9c550e1baf6989b"} Nov 21 14:13:52 crc kubenswrapper[4897]: I1121 14:13:52.717898 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-2sh55" Nov 21 14:13:52 crc kubenswrapper[4897]: I1121 14:13:52.723251 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-2sh55" Nov 21 14:13:52 crc kubenswrapper[4897]: I1121 14:13:52.733672 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-2sh55" podStartSLOduration=1.436484535 podStartE2EDuration="2.733635401s" podCreationTimestamp="2025-11-21 14:13:50 +0000 UTC" firstStartedPulling="2025-11-21 14:13:50.867394831 +0000 UTC m=+308.151988306" lastFinishedPulling="2025-11-21 14:13:52.164545697 +0000 UTC m=+309.449139172" observedRunningTime="2025-11-21 14:13:52.733346223 +0000 UTC m=+310.017939738" watchObservedRunningTime="2025-11-21 14:13:52.733635401 +0000 UTC m=+310.018228876" Nov 21 14:13:52 crc kubenswrapper[4897]: I1121 14:13:52.736739 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-lsctq" podStartSLOduration=2.736730344 podStartE2EDuration="2.736730344s" podCreationTimestamp="2025-11-21 14:13:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:13:51.7269229 +0000 UTC m=+309.011516395" watchObservedRunningTime="2025-11-21 14:13:52.736730344 +0000 UTC m=+310.021323819" Nov 21 14:13:53 crc kubenswrapper[4897]: I1121 14:13:53.412864 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/prometheus-operator-db54df47d-cl6x5"] Nov 21 14:13:53 crc kubenswrapper[4897]: I1121 14:13:53.414451 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-operator-db54df47d-cl6x5" Nov 21 14:13:53 crc kubenswrapper[4897]: I1121 14:13:53.416674 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"metrics-client-ca" Nov 21 14:13:53 crc kubenswrapper[4897]: I1121 14:13:53.417326 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-operator-tls" Nov 21 14:13:53 crc kubenswrapper[4897]: I1121 14:13:53.417390 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-operator-kube-rbac-proxy-config" Nov 21 14:13:53 crc kubenswrapper[4897]: I1121 14:13:53.425981 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-operator-dockercfg-wssfq" Nov 21 14:13:53 crc kubenswrapper[4897]: I1121 14:13:53.466374 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-operator-db54df47d-cl6x5"] Nov 21 14:13:53 crc kubenswrapper[4897]: I1121 14:13:53.550988 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-operator-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/be3bb5e8-ee93-4bdf-8bbd-a25e817d4f2f-prometheus-operator-kube-rbac-proxy-config\") pod \"prometheus-operator-db54df47d-cl6x5\" (UID: \"be3bb5e8-ee93-4bdf-8bbd-a25e817d4f2f\") " pod="openshift-monitoring/prometheus-operator-db54df47d-cl6x5" Nov 21 14:13:53 crc kubenswrapper[4897]: I1121 14:13:53.551048 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j2bw7\" (UniqueName: \"kubernetes.io/projected/be3bb5e8-ee93-4bdf-8bbd-a25e817d4f2f-kube-api-access-j2bw7\") pod \"prometheus-operator-db54df47d-cl6x5\" (UID: \"be3bb5e8-ee93-4bdf-8bbd-a25e817d4f2f\") " pod="openshift-monitoring/prometheus-operator-db54df47d-cl6x5" Nov 21 14:13:53 crc kubenswrapper[4897]: I1121 14:13:53.551091 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-operator-tls\" (UniqueName: \"kubernetes.io/secret/be3bb5e8-ee93-4bdf-8bbd-a25e817d4f2f-prometheus-operator-tls\") pod \"prometheus-operator-db54df47d-cl6x5\" (UID: \"be3bb5e8-ee93-4bdf-8bbd-a25e817d4f2f\") " pod="openshift-monitoring/prometheus-operator-db54df47d-cl6x5" Nov 21 14:13:53 crc kubenswrapper[4897]: I1121 14:13:53.551282 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/be3bb5e8-ee93-4bdf-8bbd-a25e817d4f2f-metrics-client-ca\") pod \"prometheus-operator-db54df47d-cl6x5\" (UID: \"be3bb5e8-ee93-4bdf-8bbd-a25e817d4f2f\") " pod="openshift-monitoring/prometheus-operator-db54df47d-cl6x5" Nov 21 14:13:53 crc kubenswrapper[4897]: I1121 14:13:53.652751 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-operator-tls\" (UniqueName: \"kubernetes.io/secret/be3bb5e8-ee93-4bdf-8bbd-a25e817d4f2f-prometheus-operator-tls\") pod \"prometheus-operator-db54df47d-cl6x5\" (UID: \"be3bb5e8-ee93-4bdf-8bbd-a25e817d4f2f\") " pod="openshift-monitoring/prometheus-operator-db54df47d-cl6x5" Nov 21 14:13:53 crc kubenswrapper[4897]: I1121 14:13:53.652840 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/be3bb5e8-ee93-4bdf-8bbd-a25e817d4f2f-metrics-client-ca\") pod \"prometheus-operator-db54df47d-cl6x5\" (UID: \"be3bb5e8-ee93-4bdf-8bbd-a25e817d4f2f\") " pod="openshift-monitoring/prometheus-operator-db54df47d-cl6x5" Nov 21 14:13:53 crc kubenswrapper[4897]: I1121 14:13:53.652904 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-operator-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/be3bb5e8-ee93-4bdf-8bbd-a25e817d4f2f-prometheus-operator-kube-rbac-proxy-config\") pod \"prometheus-operator-db54df47d-cl6x5\" (UID: \"be3bb5e8-ee93-4bdf-8bbd-a25e817d4f2f\") " pod="openshift-monitoring/prometheus-operator-db54df47d-cl6x5" Nov 21 14:13:53 crc kubenswrapper[4897]: I1121 14:13:53.652954 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j2bw7\" (UniqueName: \"kubernetes.io/projected/be3bb5e8-ee93-4bdf-8bbd-a25e817d4f2f-kube-api-access-j2bw7\") pod \"prometheus-operator-db54df47d-cl6x5\" (UID: \"be3bb5e8-ee93-4bdf-8bbd-a25e817d4f2f\") " pod="openshift-monitoring/prometheus-operator-db54df47d-cl6x5" Nov 21 14:13:53 crc kubenswrapper[4897]: I1121 14:13:53.653961 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/be3bb5e8-ee93-4bdf-8bbd-a25e817d4f2f-metrics-client-ca\") pod \"prometheus-operator-db54df47d-cl6x5\" (UID: \"be3bb5e8-ee93-4bdf-8bbd-a25e817d4f2f\") " pod="openshift-monitoring/prometheus-operator-db54df47d-cl6x5" Nov 21 14:13:53 crc kubenswrapper[4897]: I1121 14:13:53.657745 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-operator-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/be3bb5e8-ee93-4bdf-8bbd-a25e817d4f2f-prometheus-operator-kube-rbac-proxy-config\") pod \"prometheus-operator-db54df47d-cl6x5\" (UID: \"be3bb5e8-ee93-4bdf-8bbd-a25e817d4f2f\") " pod="openshift-monitoring/prometheus-operator-db54df47d-cl6x5" Nov 21 14:13:53 crc kubenswrapper[4897]: I1121 14:13:53.657813 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-operator-tls\" (UniqueName: \"kubernetes.io/secret/be3bb5e8-ee93-4bdf-8bbd-a25e817d4f2f-prometheus-operator-tls\") pod \"prometheus-operator-db54df47d-cl6x5\" (UID: \"be3bb5e8-ee93-4bdf-8bbd-a25e817d4f2f\") " pod="openshift-monitoring/prometheus-operator-db54df47d-cl6x5" Nov 21 14:13:53 crc kubenswrapper[4897]: I1121 14:13:53.670037 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j2bw7\" (UniqueName: \"kubernetes.io/projected/be3bb5e8-ee93-4bdf-8bbd-a25e817d4f2f-kube-api-access-j2bw7\") pod \"prometheus-operator-db54df47d-cl6x5\" (UID: \"be3bb5e8-ee93-4bdf-8bbd-a25e817d4f2f\") " pod="openshift-monitoring/prometheus-operator-db54df47d-cl6x5" Nov 21 14:13:53 crc kubenswrapper[4897]: I1121 14:13:53.733552 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-operator-db54df47d-cl6x5" Nov 21 14:13:53 crc kubenswrapper[4897]: I1121 14:13:53.900175 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-operator-db54df47d-cl6x5"] Nov 21 14:13:54 crc kubenswrapper[4897]: I1121 14:13:54.727761 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-db54df47d-cl6x5" event={"ID":"be3bb5e8-ee93-4bdf-8bbd-a25e817d4f2f","Type":"ContainerStarted","Data":"e373f3c302b7f58a8d713bd5d28c67cd85cf6f634271d58ba812df6711a97493"} Nov 21 14:13:56 crc kubenswrapper[4897]: I1121 14:13:56.738006 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-db54df47d-cl6x5" event={"ID":"be3bb5e8-ee93-4bdf-8bbd-a25e817d4f2f","Type":"ContainerStarted","Data":"0bde250090e155ff7e27363f6ca74be1dc30590bb3520dc270c4e8bfd6f94bda"} Nov 21 14:13:57 crc kubenswrapper[4897]: I1121 14:13:57.747189 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-db54df47d-cl6x5" event={"ID":"be3bb5e8-ee93-4bdf-8bbd-a25e817d4f2f","Type":"ContainerStarted","Data":"17344c507af3d8d1989f15df6951373981fa23379e8c7c563e3bf7cd8493eb47"} Nov 21 14:13:57 crc kubenswrapper[4897]: I1121 14:13:57.766556 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/prometheus-operator-db54df47d-cl6x5" podStartSLOduration=2.603767129 podStartE2EDuration="4.766529937s" podCreationTimestamp="2025-11-21 14:13:53 +0000 UTC" firstStartedPulling="2025-11-21 14:13:53.910220557 +0000 UTC m=+311.194814022" lastFinishedPulling="2025-11-21 14:13:56.072983355 +0000 UTC m=+313.357576830" observedRunningTime="2025-11-21 14:13:57.761785443 +0000 UTC m=+315.046379018" watchObservedRunningTime="2025-11-21 14:13:57.766529937 +0000 UTC m=+315.051123452" Nov 21 14:13:59 crc kubenswrapper[4897]: I1121 14:13:59.747124 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/openshift-state-metrics-566fddb674-hw2hf"] Nov 21 14:13:59 crc kubenswrapper[4897]: I1121 14:13:59.748800 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/openshift-state-metrics-566fddb674-hw2hf" Nov 21 14:13:59 crc kubenswrapper[4897]: I1121 14:13:59.750676 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"openshift-state-metrics-kube-rbac-proxy-config" Nov 21 14:13:59 crc kubenswrapper[4897]: I1121 14:13:59.750907 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"openshift-state-metrics-tls" Nov 21 14:13:59 crc kubenswrapper[4897]: I1121 14:13:59.751301 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"openshift-state-metrics-dockercfg-4wr9s" Nov 21 14:13:59 crc kubenswrapper[4897]: I1121 14:13:59.792903 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/kube-state-metrics-777cb5bd5d-tm69l"] Nov 21 14:13:59 crc kubenswrapper[4897]: I1121 14:13:59.794105 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-tm69l" Nov 21 14:13:59 crc kubenswrapper[4897]: I1121 14:13:59.795861 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"kube-state-metrics-tls" Nov 21 14:13:59 crc kubenswrapper[4897]: I1121 14:13:59.796062 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"kube-state-metrics-kube-rbac-proxy-config" Nov 21 14:13:59 crc kubenswrapper[4897]: I1121 14:13:59.797125 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"kube-state-metrics-dockercfg-g7tt8" Nov 21 14:13:59 crc kubenswrapper[4897]: I1121 14:13:59.797723 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"kube-state-metrics-custom-resource-state-configmap" Nov 21 14:13:59 crc kubenswrapper[4897]: I1121 14:13:59.802295 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/openshift-state-metrics-566fddb674-hw2hf"] Nov 21 14:13:59 crc kubenswrapper[4897]: I1121 14:13:59.816900 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/kube-state-metrics-777cb5bd5d-tm69l"] Nov 21 14:13:59 crc kubenswrapper[4897]: I1121 14:13:59.828160 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/8d0ea2ea-c9a1-495a-a626-7a1a735308df-openshift-state-metrics-tls\") pod \"openshift-state-metrics-566fddb674-hw2hf\" (UID: \"8d0ea2ea-c9a1-495a-a626-7a1a735308df\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-hw2hf" Nov 21 14:13:59 crc kubenswrapper[4897]: I1121 14:13:59.828204 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/8d0ea2ea-c9a1-495a-a626-7a1a735308df-openshift-state-metrics-kube-rbac-proxy-config\") pod \"openshift-state-metrics-566fddb674-hw2hf\" (UID: \"8d0ea2ea-c9a1-495a-a626-7a1a735308df\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-hw2hf" Nov 21 14:13:59 crc kubenswrapper[4897]: I1121 14:13:59.828247 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/8d0ea2ea-c9a1-495a-a626-7a1a735308df-metrics-client-ca\") pod \"openshift-state-metrics-566fddb674-hw2hf\" (UID: \"8d0ea2ea-c9a1-495a-a626-7a1a735308df\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-hw2hf" Nov 21 14:13:59 crc kubenswrapper[4897]: I1121 14:13:59.828265 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8nr4k\" (UniqueName: \"kubernetes.io/projected/8d0ea2ea-c9a1-495a-a626-7a1a735308df-kube-api-access-8nr4k\") pod \"openshift-state-metrics-566fddb674-hw2hf\" (UID: \"8d0ea2ea-c9a1-495a-a626-7a1a735308df\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-hw2hf" Nov 21 14:13:59 crc kubenswrapper[4897]: I1121 14:13:59.907247 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/node-exporter-9gpbl"] Nov 21 14:13:59 crc kubenswrapper[4897]: I1121 14:13:59.908442 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/node-exporter-9gpbl" Nov 21 14:13:59 crc kubenswrapper[4897]: I1121 14:13:59.910329 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"node-exporter-dockercfg-nvd2m" Nov 21 14:13:59 crc kubenswrapper[4897]: I1121 14:13:59.910419 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"node-exporter-kube-rbac-proxy-config" Nov 21 14:13:59 crc kubenswrapper[4897]: I1121 14:13:59.910470 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"node-exporter-tls" Nov 21 14:13:59 crc kubenswrapper[4897]: I1121 14:13:59.930146 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/91c21c9f-aed2-41c3-b3fa-371f9903350a-kube-state-metrics-kube-rbac-proxy-config\") pod \"kube-state-metrics-777cb5bd5d-tm69l\" (UID: \"91c21c9f-aed2-41c3-b3fa-371f9903350a\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-tm69l" Nov 21 14:13:59 crc kubenswrapper[4897]: I1121 14:13:59.930208 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f76gb\" (UniqueName: \"kubernetes.io/projected/91c21c9f-aed2-41c3-b3fa-371f9903350a-kube-api-access-f76gb\") pod \"kube-state-metrics-777cb5bd5d-tm69l\" (UID: \"91c21c9f-aed2-41c3-b3fa-371f9903350a\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-tm69l" Nov 21 14:13:59 crc kubenswrapper[4897]: I1121 14:13:59.930252 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/8d0ea2ea-c9a1-495a-a626-7a1a735308df-metrics-client-ca\") pod \"openshift-state-metrics-566fddb674-hw2hf\" (UID: \"8d0ea2ea-c9a1-495a-a626-7a1a735308df\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-hw2hf" Nov 21 14:13:59 crc kubenswrapper[4897]: I1121 14:13:59.930280 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8nr4k\" (UniqueName: \"kubernetes.io/projected/8d0ea2ea-c9a1-495a-a626-7a1a735308df-kube-api-access-8nr4k\") pod \"openshift-state-metrics-566fddb674-hw2hf\" (UID: \"8d0ea2ea-c9a1-495a-a626-7a1a735308df\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-hw2hf" Nov 21 14:13:59 crc kubenswrapper[4897]: I1121 14:13:59.930380 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-custom-resource-state-configmap\" (UniqueName: \"kubernetes.io/configmap/91c21c9f-aed2-41c3-b3fa-371f9903350a-kube-state-metrics-custom-resource-state-configmap\") pod \"kube-state-metrics-777cb5bd5d-tm69l\" (UID: \"91c21c9f-aed2-41c3-b3fa-371f9903350a\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-tm69l" Nov 21 14:13:59 crc kubenswrapper[4897]: I1121 14:13:59.930418 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/91c21c9f-aed2-41c3-b3fa-371f9903350a-kube-state-metrics-tls\") pod \"kube-state-metrics-777cb5bd5d-tm69l\" (UID: \"91c21c9f-aed2-41c3-b3fa-371f9903350a\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-tm69l" Nov 21 14:13:59 crc kubenswrapper[4897]: I1121 14:13:59.930449 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/91c21c9f-aed2-41c3-b3fa-371f9903350a-metrics-client-ca\") pod \"kube-state-metrics-777cb5bd5d-tm69l\" (UID: \"91c21c9f-aed2-41c3-b3fa-371f9903350a\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-tm69l" Nov 21 14:13:59 crc kubenswrapper[4897]: I1121 14:13:59.930473 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/8d0ea2ea-c9a1-495a-a626-7a1a735308df-openshift-state-metrics-tls\") pod \"openshift-state-metrics-566fddb674-hw2hf\" (UID: \"8d0ea2ea-c9a1-495a-a626-7a1a735308df\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-hw2hf" Nov 21 14:13:59 crc kubenswrapper[4897]: I1121 14:13:59.930516 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/8d0ea2ea-c9a1-495a-a626-7a1a735308df-openshift-state-metrics-kube-rbac-proxy-config\") pod \"openshift-state-metrics-566fddb674-hw2hf\" (UID: \"8d0ea2ea-c9a1-495a-a626-7a1a735308df\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-hw2hf" Nov 21 14:13:59 crc kubenswrapper[4897]: I1121 14:13:59.930645 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"volume-directive-shadow\" (UniqueName: \"kubernetes.io/empty-dir/91c21c9f-aed2-41c3-b3fa-371f9903350a-volume-directive-shadow\") pod \"kube-state-metrics-777cb5bd5d-tm69l\" (UID: \"91c21c9f-aed2-41c3-b3fa-371f9903350a\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-tm69l" Nov 21 14:13:59 crc kubenswrapper[4897]: E1121 14:13:59.930666 4897 secret.go:188] Couldn't get secret openshift-monitoring/openshift-state-metrics-tls: secret "openshift-state-metrics-tls" not found Nov 21 14:13:59 crc kubenswrapper[4897]: E1121 14:13:59.931019 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8d0ea2ea-c9a1-495a-a626-7a1a735308df-openshift-state-metrics-tls podName:8d0ea2ea-c9a1-495a-a626-7a1a735308df nodeName:}" failed. No retries permitted until 2025-11-21 14:14:00.430990946 +0000 UTC m=+317.715584421 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "openshift-state-metrics-tls" (UniqueName: "kubernetes.io/secret/8d0ea2ea-c9a1-495a-a626-7a1a735308df-openshift-state-metrics-tls") pod "openshift-state-metrics-566fddb674-hw2hf" (UID: "8d0ea2ea-c9a1-495a-a626-7a1a735308df") : secret "openshift-state-metrics-tls" not found Nov 21 14:13:59 crc kubenswrapper[4897]: I1121 14:13:59.931412 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/8d0ea2ea-c9a1-495a-a626-7a1a735308df-metrics-client-ca\") pod \"openshift-state-metrics-566fddb674-hw2hf\" (UID: \"8d0ea2ea-c9a1-495a-a626-7a1a735308df\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-hw2hf" Nov 21 14:13:59 crc kubenswrapper[4897]: I1121 14:13:59.942097 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/8d0ea2ea-c9a1-495a-a626-7a1a735308df-openshift-state-metrics-kube-rbac-proxy-config\") pod \"openshift-state-metrics-566fddb674-hw2hf\" (UID: \"8d0ea2ea-c9a1-495a-a626-7a1a735308df\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-hw2hf" Nov 21 14:13:59 crc kubenswrapper[4897]: I1121 14:13:59.966481 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8nr4k\" (UniqueName: \"kubernetes.io/projected/8d0ea2ea-c9a1-495a-a626-7a1a735308df-kube-api-access-8nr4k\") pod \"openshift-state-metrics-566fddb674-hw2hf\" (UID: \"8d0ea2ea-c9a1-495a-a626-7a1a735308df\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-hw2hf" Nov 21 14:14:00 crc kubenswrapper[4897]: I1121 14:14:00.032253 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-wtmp\" (UniqueName: \"kubernetes.io/host-path/8caaec7a-8eab-4ffe-9e68-f5c763d044eb-node-exporter-wtmp\") pod \"node-exporter-9gpbl\" (UID: \"8caaec7a-8eab-4ffe-9e68-f5c763d044eb\") " pod="openshift-monitoring/node-exporter-9gpbl" Nov 21 14:14:00 crc kubenswrapper[4897]: I1121 14:14:00.032312 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"root\" (UniqueName: \"kubernetes.io/host-path/8caaec7a-8eab-4ffe-9e68-f5c763d044eb-root\") pod \"node-exporter-9gpbl\" (UID: \"8caaec7a-8eab-4ffe-9e68-f5c763d044eb\") " pod="openshift-monitoring/node-exporter-9gpbl" Nov 21 14:14:00 crc kubenswrapper[4897]: I1121 14:14:00.032386 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/8caaec7a-8eab-4ffe-9e68-f5c763d044eb-node-exporter-kube-rbac-proxy-config\") pod \"node-exporter-9gpbl\" (UID: \"8caaec7a-8eab-4ffe-9e68-f5c763d044eb\") " pod="openshift-monitoring/node-exporter-9gpbl" Nov 21 14:14:00 crc kubenswrapper[4897]: I1121 14:14:00.032454 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-textfile\" (UniqueName: \"kubernetes.io/empty-dir/8caaec7a-8eab-4ffe-9e68-f5c763d044eb-node-exporter-textfile\") pod \"node-exporter-9gpbl\" (UID: \"8caaec7a-8eab-4ffe-9e68-f5c763d044eb\") " pod="openshift-monitoring/node-exporter-9gpbl" Nov 21 14:14:00 crc kubenswrapper[4897]: I1121 14:14:00.032550 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-custom-resource-state-configmap\" (UniqueName: \"kubernetes.io/configmap/91c21c9f-aed2-41c3-b3fa-371f9903350a-kube-state-metrics-custom-resource-state-configmap\") pod \"kube-state-metrics-777cb5bd5d-tm69l\" (UID: \"91c21c9f-aed2-41c3-b3fa-371f9903350a\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-tm69l" Nov 21 14:14:00 crc kubenswrapper[4897]: I1121 14:14:00.032594 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/91c21c9f-aed2-41c3-b3fa-371f9903350a-kube-state-metrics-tls\") pod \"kube-state-metrics-777cb5bd5d-tm69l\" (UID: \"91c21c9f-aed2-41c3-b3fa-371f9903350a\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-tm69l" Nov 21 14:14:00 crc kubenswrapper[4897]: I1121 14:14:00.032630 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/91c21c9f-aed2-41c3-b3fa-371f9903350a-metrics-client-ca\") pod \"kube-state-metrics-777cb5bd5d-tm69l\" (UID: \"91c21c9f-aed2-41c3-b3fa-371f9903350a\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-tm69l" Nov 21 14:14:00 crc kubenswrapper[4897]: I1121 14:14:00.032677 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"volume-directive-shadow\" (UniqueName: \"kubernetes.io/empty-dir/91c21c9f-aed2-41c3-b3fa-371f9903350a-volume-directive-shadow\") pod \"kube-state-metrics-777cb5bd5d-tm69l\" (UID: \"91c21c9f-aed2-41c3-b3fa-371f9903350a\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-tm69l" Nov 21 14:14:00 crc kubenswrapper[4897]: I1121 14:14:00.032726 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/8caaec7a-8eab-4ffe-9e68-f5c763d044eb-metrics-client-ca\") pod \"node-exporter-9gpbl\" (UID: \"8caaec7a-8eab-4ffe-9e68-f5c763d044eb\") " pod="openshift-monitoring/node-exporter-9gpbl" Nov 21 14:14:00 crc kubenswrapper[4897]: I1121 14:14:00.032746 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-28fnm\" (UniqueName: \"kubernetes.io/projected/8caaec7a-8eab-4ffe-9e68-f5c763d044eb-kube-api-access-28fnm\") pod \"node-exporter-9gpbl\" (UID: \"8caaec7a-8eab-4ffe-9e68-f5c763d044eb\") " pod="openshift-monitoring/node-exporter-9gpbl" Nov 21 14:14:00 crc kubenswrapper[4897]: I1121 14:14:00.032772 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/91c21c9f-aed2-41c3-b3fa-371f9903350a-kube-state-metrics-kube-rbac-proxy-config\") pod \"kube-state-metrics-777cb5bd5d-tm69l\" (UID: \"91c21c9f-aed2-41c3-b3fa-371f9903350a\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-tm69l" Nov 21 14:14:00 crc kubenswrapper[4897]: E1121 14:14:00.032788 4897 secret.go:188] Couldn't get secret openshift-monitoring/kube-state-metrics-tls: secret "kube-state-metrics-tls" not found Nov 21 14:14:00 crc kubenswrapper[4897]: E1121 14:14:00.032841 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/91c21c9f-aed2-41c3-b3fa-371f9903350a-kube-state-metrics-tls podName:91c21c9f-aed2-41c3-b3fa-371f9903350a nodeName:}" failed. No retries permitted until 2025-11-21 14:14:00.532821452 +0000 UTC m=+317.817414927 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-state-metrics-tls" (UniqueName: "kubernetes.io/secret/91c21c9f-aed2-41c3-b3fa-371f9903350a-kube-state-metrics-tls") pod "kube-state-metrics-777cb5bd5d-tm69l" (UID: "91c21c9f-aed2-41c3-b3fa-371f9903350a") : secret "kube-state-metrics-tls" not found Nov 21 14:14:00 crc kubenswrapper[4897]: I1121 14:14:00.032792 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f76gb\" (UniqueName: \"kubernetes.io/projected/91c21c9f-aed2-41c3-b3fa-371f9903350a-kube-api-access-f76gb\") pod \"kube-state-metrics-777cb5bd5d-tm69l\" (UID: \"91c21c9f-aed2-41c3-b3fa-371f9903350a\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-tm69l" Nov 21 14:14:00 crc kubenswrapper[4897]: I1121 14:14:00.033111 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/8caaec7a-8eab-4ffe-9e68-f5c763d044eb-sys\") pod \"node-exporter-9gpbl\" (UID: \"8caaec7a-8eab-4ffe-9e68-f5c763d044eb\") " pod="openshift-monitoring/node-exporter-9gpbl" Nov 21 14:14:00 crc kubenswrapper[4897]: I1121 14:14:00.033131 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/8caaec7a-8eab-4ffe-9e68-f5c763d044eb-node-exporter-tls\") pod \"node-exporter-9gpbl\" (UID: \"8caaec7a-8eab-4ffe-9e68-f5c763d044eb\") " pod="openshift-monitoring/node-exporter-9gpbl" Nov 21 14:14:00 crc kubenswrapper[4897]: I1121 14:14:00.033319 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"volume-directive-shadow\" (UniqueName: \"kubernetes.io/empty-dir/91c21c9f-aed2-41c3-b3fa-371f9903350a-volume-directive-shadow\") pod \"kube-state-metrics-777cb5bd5d-tm69l\" (UID: \"91c21c9f-aed2-41c3-b3fa-371f9903350a\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-tm69l" Nov 21 14:14:00 crc kubenswrapper[4897]: I1121 14:14:00.033698 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/91c21c9f-aed2-41c3-b3fa-371f9903350a-metrics-client-ca\") pod \"kube-state-metrics-777cb5bd5d-tm69l\" (UID: \"91c21c9f-aed2-41c3-b3fa-371f9903350a\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-tm69l" Nov 21 14:14:00 crc kubenswrapper[4897]: I1121 14:14:00.033777 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-custom-resource-state-configmap\" (UniqueName: \"kubernetes.io/configmap/91c21c9f-aed2-41c3-b3fa-371f9903350a-kube-state-metrics-custom-resource-state-configmap\") pod \"kube-state-metrics-777cb5bd5d-tm69l\" (UID: \"91c21c9f-aed2-41c3-b3fa-371f9903350a\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-tm69l" Nov 21 14:14:00 crc kubenswrapper[4897]: I1121 14:14:00.036320 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/91c21c9f-aed2-41c3-b3fa-371f9903350a-kube-state-metrics-kube-rbac-proxy-config\") pod \"kube-state-metrics-777cb5bd5d-tm69l\" (UID: \"91c21c9f-aed2-41c3-b3fa-371f9903350a\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-tm69l" Nov 21 14:14:00 crc kubenswrapper[4897]: I1121 14:14:00.047458 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f76gb\" (UniqueName: \"kubernetes.io/projected/91c21c9f-aed2-41c3-b3fa-371f9903350a-kube-api-access-f76gb\") pod \"kube-state-metrics-777cb5bd5d-tm69l\" (UID: \"91c21c9f-aed2-41c3-b3fa-371f9903350a\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-tm69l" Nov 21 14:14:00 crc kubenswrapper[4897]: I1121 14:14:00.133882 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-exporter-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/8caaec7a-8eab-4ffe-9e68-f5c763d044eb-node-exporter-kube-rbac-proxy-config\") pod \"node-exporter-9gpbl\" (UID: \"8caaec7a-8eab-4ffe-9e68-f5c763d044eb\") " pod="openshift-monitoring/node-exporter-9gpbl" Nov 21 14:14:00 crc kubenswrapper[4897]: I1121 14:14:00.133929 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-exporter-textfile\" (UniqueName: \"kubernetes.io/empty-dir/8caaec7a-8eab-4ffe-9e68-f5c763d044eb-node-exporter-textfile\") pod \"node-exporter-9gpbl\" (UID: \"8caaec7a-8eab-4ffe-9e68-f5c763d044eb\") " pod="openshift-monitoring/node-exporter-9gpbl" Nov 21 14:14:00 crc kubenswrapper[4897]: I1121 14:14:00.134021 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/8caaec7a-8eab-4ffe-9e68-f5c763d044eb-metrics-client-ca\") pod \"node-exporter-9gpbl\" (UID: \"8caaec7a-8eab-4ffe-9e68-f5c763d044eb\") " pod="openshift-monitoring/node-exporter-9gpbl" Nov 21 14:14:00 crc kubenswrapper[4897]: I1121 14:14:00.134041 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-28fnm\" (UniqueName: \"kubernetes.io/projected/8caaec7a-8eab-4ffe-9e68-f5c763d044eb-kube-api-access-28fnm\") pod \"node-exporter-9gpbl\" (UID: \"8caaec7a-8eab-4ffe-9e68-f5c763d044eb\") " pod="openshift-monitoring/node-exporter-9gpbl" Nov 21 14:14:00 crc kubenswrapper[4897]: I1121 14:14:00.134063 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/8caaec7a-8eab-4ffe-9e68-f5c763d044eb-sys\") pod \"node-exporter-9gpbl\" (UID: \"8caaec7a-8eab-4ffe-9e68-f5c763d044eb\") " pod="openshift-monitoring/node-exporter-9gpbl" Nov 21 14:14:00 crc kubenswrapper[4897]: I1121 14:14:00.134076 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/8caaec7a-8eab-4ffe-9e68-f5c763d044eb-node-exporter-tls\") pod \"node-exporter-9gpbl\" (UID: \"8caaec7a-8eab-4ffe-9e68-f5c763d044eb\") " pod="openshift-monitoring/node-exporter-9gpbl" Nov 21 14:14:00 crc kubenswrapper[4897]: I1121 14:14:00.134093 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-exporter-wtmp\" (UniqueName: \"kubernetes.io/host-path/8caaec7a-8eab-4ffe-9e68-f5c763d044eb-node-exporter-wtmp\") pod \"node-exporter-9gpbl\" (UID: \"8caaec7a-8eab-4ffe-9e68-f5c763d044eb\") " pod="openshift-monitoring/node-exporter-9gpbl" Nov 21 14:14:00 crc kubenswrapper[4897]: I1121 14:14:00.134111 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"root\" (UniqueName: \"kubernetes.io/host-path/8caaec7a-8eab-4ffe-9e68-f5c763d044eb-root\") pod \"node-exporter-9gpbl\" (UID: \"8caaec7a-8eab-4ffe-9e68-f5c763d044eb\") " pod="openshift-monitoring/node-exporter-9gpbl" Nov 21 14:14:00 crc kubenswrapper[4897]: I1121 14:14:00.134185 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"root\" (UniqueName: \"kubernetes.io/host-path/8caaec7a-8eab-4ffe-9e68-f5c763d044eb-root\") pod \"node-exporter-9gpbl\" (UID: \"8caaec7a-8eab-4ffe-9e68-f5c763d044eb\") " pod="openshift-monitoring/node-exporter-9gpbl" Nov 21 14:14:00 crc kubenswrapper[4897]: I1121 14:14:00.134434 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/8caaec7a-8eab-4ffe-9e68-f5c763d044eb-sys\") pod \"node-exporter-9gpbl\" (UID: \"8caaec7a-8eab-4ffe-9e68-f5c763d044eb\") " pod="openshift-monitoring/node-exporter-9gpbl" Nov 21 14:14:00 crc kubenswrapper[4897]: I1121 14:14:00.134608 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-exporter-textfile\" (UniqueName: \"kubernetes.io/empty-dir/8caaec7a-8eab-4ffe-9e68-f5c763d044eb-node-exporter-textfile\") pod \"node-exporter-9gpbl\" (UID: \"8caaec7a-8eab-4ffe-9e68-f5c763d044eb\") " pod="openshift-monitoring/node-exporter-9gpbl" Nov 21 14:14:00 crc kubenswrapper[4897]: I1121 14:14:00.134825 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-exporter-wtmp\" (UniqueName: \"kubernetes.io/host-path/8caaec7a-8eab-4ffe-9e68-f5c763d044eb-node-exporter-wtmp\") pod \"node-exporter-9gpbl\" (UID: \"8caaec7a-8eab-4ffe-9e68-f5c763d044eb\") " pod="openshift-monitoring/node-exporter-9gpbl" Nov 21 14:14:00 crc kubenswrapper[4897]: E1121 14:14:00.134831 4897 secret.go:188] Couldn't get secret openshift-monitoring/node-exporter-tls: secret "node-exporter-tls" not found Nov 21 14:14:00 crc kubenswrapper[4897]: E1121 14:14:00.135072 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8caaec7a-8eab-4ffe-9e68-f5c763d044eb-node-exporter-tls podName:8caaec7a-8eab-4ffe-9e68-f5c763d044eb nodeName:}" failed. No retries permitted until 2025-11-21 14:14:00.63505062 +0000 UTC m=+317.919644175 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "node-exporter-tls" (UniqueName: "kubernetes.io/secret/8caaec7a-8eab-4ffe-9e68-f5c763d044eb-node-exporter-tls") pod "node-exporter-9gpbl" (UID: "8caaec7a-8eab-4ffe-9e68-f5c763d044eb") : secret "node-exporter-tls" not found Nov 21 14:14:00 crc kubenswrapper[4897]: I1121 14:14:00.135476 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/8caaec7a-8eab-4ffe-9e68-f5c763d044eb-metrics-client-ca\") pod \"node-exporter-9gpbl\" (UID: \"8caaec7a-8eab-4ffe-9e68-f5c763d044eb\") " pod="openshift-monitoring/node-exporter-9gpbl" Nov 21 14:14:00 crc kubenswrapper[4897]: I1121 14:14:00.136805 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-exporter-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/8caaec7a-8eab-4ffe-9e68-f5c763d044eb-node-exporter-kube-rbac-proxy-config\") pod \"node-exporter-9gpbl\" (UID: \"8caaec7a-8eab-4ffe-9e68-f5c763d044eb\") " pod="openshift-monitoring/node-exporter-9gpbl" Nov 21 14:14:00 crc kubenswrapper[4897]: I1121 14:14:00.151439 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-28fnm\" (UniqueName: \"kubernetes.io/projected/8caaec7a-8eab-4ffe-9e68-f5c763d044eb-kube-api-access-28fnm\") pod \"node-exporter-9gpbl\" (UID: \"8caaec7a-8eab-4ffe-9e68-f5c763d044eb\") " pod="openshift-monitoring/node-exporter-9gpbl" Nov 21 14:14:00 crc kubenswrapper[4897]: I1121 14:14:00.439450 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/8d0ea2ea-c9a1-495a-a626-7a1a735308df-openshift-state-metrics-tls\") pod \"openshift-state-metrics-566fddb674-hw2hf\" (UID: \"8d0ea2ea-c9a1-495a-a626-7a1a735308df\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-hw2hf" Nov 21 14:14:00 crc kubenswrapper[4897]: I1121 14:14:00.443080 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/8d0ea2ea-c9a1-495a-a626-7a1a735308df-openshift-state-metrics-tls\") pod \"openshift-state-metrics-566fddb674-hw2hf\" (UID: \"8d0ea2ea-c9a1-495a-a626-7a1a735308df\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-hw2hf" Nov 21 14:14:00 crc kubenswrapper[4897]: I1121 14:14:00.540813 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/91c21c9f-aed2-41c3-b3fa-371f9903350a-kube-state-metrics-tls\") pod \"kube-state-metrics-777cb5bd5d-tm69l\" (UID: \"91c21c9f-aed2-41c3-b3fa-371f9903350a\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-tm69l" Nov 21 14:14:00 crc kubenswrapper[4897]: I1121 14:14:00.544259 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/91c21c9f-aed2-41c3-b3fa-371f9903350a-kube-state-metrics-tls\") pod \"kube-state-metrics-777cb5bd5d-tm69l\" (UID: \"91c21c9f-aed2-41c3-b3fa-371f9903350a\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-tm69l" Nov 21 14:14:00 crc kubenswrapper[4897]: I1121 14:14:00.642103 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/8caaec7a-8eab-4ffe-9e68-f5c763d044eb-node-exporter-tls\") pod \"node-exporter-9gpbl\" (UID: \"8caaec7a-8eab-4ffe-9e68-f5c763d044eb\") " pod="openshift-monitoring/node-exporter-9gpbl" Nov 21 14:14:00 crc kubenswrapper[4897]: I1121 14:14:00.645754 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/8caaec7a-8eab-4ffe-9e68-f5c763d044eb-node-exporter-tls\") pod \"node-exporter-9gpbl\" (UID: \"8caaec7a-8eab-4ffe-9e68-f5c763d044eb\") " pod="openshift-monitoring/node-exporter-9gpbl" Nov 21 14:14:00 crc kubenswrapper[4897]: I1121 14:14:00.667485 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/openshift-state-metrics-566fddb674-hw2hf" Nov 21 14:14:00 crc kubenswrapper[4897]: I1121 14:14:00.708748 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-tm69l" Nov 21 14:14:00 crc kubenswrapper[4897]: I1121 14:14:00.821800 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/node-exporter-9gpbl" Nov 21 14:14:00 crc kubenswrapper[4897]: I1121 14:14:00.824017 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/alertmanager-main-0"] Nov 21 14:14:00 crc kubenswrapper[4897]: I1121 14:14:00.830355 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/alertmanager-main-0" Nov 21 14:14:00 crc kubenswrapper[4897]: I1121 14:14:00.833474 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-main-tls-assets-0" Nov 21 14:14:00 crc kubenswrapper[4897]: I1121 14:14:00.833629 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-kube-rbac-proxy" Nov 21 14:14:00 crc kubenswrapper[4897]: I1121 14:14:00.833813 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-kube-rbac-proxy-web" Nov 21 14:14:00 crc kubenswrapper[4897]: I1121 14:14:00.833908 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-main-tls" Nov 21 14:14:00 crc kubenswrapper[4897]: I1121 14:14:00.836421 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-main-web-config" Nov 21 14:14:00 crc kubenswrapper[4897]: I1121 14:14:00.836722 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-kube-rbac-proxy-metric" Nov 21 14:14:00 crc kubenswrapper[4897]: I1121 14:14:00.836786 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-main-dockercfg-z4x9b" Nov 21 14:14:00 crc kubenswrapper[4897]: I1121 14:14:00.836964 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-main-generated" Nov 21 14:14:00 crc kubenswrapper[4897]: I1121 14:14:00.840229 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/alertmanager-main-0"] Nov 21 14:14:00 crc kubenswrapper[4897]: I1121 14:14:00.857644 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"alertmanager-trusted-ca-bundle" Nov 21 14:14:00 crc kubenswrapper[4897]: I1121 14:14:00.953182 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/d0fb95e9-9ac7-46bf-996a-0cef021092fe-secret-alertmanager-kube-rbac-proxy-web\") pod \"alertmanager-main-0\" (UID: \"d0fb95e9-9ac7-46bf-996a-0cef021092fe\") " pod="openshift-monitoring/alertmanager-main-0" Nov 21 14:14:00 crc kubenswrapper[4897]: I1121 14:14:00.953265 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/d0fb95e9-9ac7-46bf-996a-0cef021092fe-config-out\") pod \"alertmanager-main-0\" (UID: \"d0fb95e9-9ac7-46bf-996a-0cef021092fe\") " pod="openshift-monitoring/alertmanager-main-0" Nov 21 14:14:00 crc kubenswrapper[4897]: I1121 14:14:00.953293 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/d0fb95e9-9ac7-46bf-996a-0cef021092fe-web-config\") pod \"alertmanager-main-0\" (UID: \"d0fb95e9-9ac7-46bf-996a-0cef021092fe\") " pod="openshift-monitoring/alertmanager-main-0" Nov 21 14:14:00 crc kubenswrapper[4897]: I1121 14:14:00.953313 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/d0fb95e9-9ac7-46bf-996a-0cef021092fe-metrics-client-ca\") pod \"alertmanager-main-0\" (UID: \"d0fb95e9-9ac7-46bf-996a-0cef021092fe\") " pod="openshift-monitoring/alertmanager-main-0" Nov 21 14:14:00 crc kubenswrapper[4897]: I1121 14:14:00.953331 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/d0fb95e9-9ac7-46bf-996a-0cef021092fe-secret-alertmanager-kube-rbac-proxy\") pod \"alertmanager-main-0\" (UID: \"d0fb95e9-9ac7-46bf-996a-0cef021092fe\") " pod="openshift-monitoring/alertmanager-main-0" Nov 21 14:14:00 crc kubenswrapper[4897]: I1121 14:14:00.953348 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/d0fb95e9-9ac7-46bf-996a-0cef021092fe-config-volume\") pod \"alertmanager-main-0\" (UID: \"d0fb95e9-9ac7-46bf-996a-0cef021092fe\") " pod="openshift-monitoring/alertmanager-main-0" Nov 21 14:14:00 crc kubenswrapper[4897]: I1121 14:14:00.953362 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/d0fb95e9-9ac7-46bf-996a-0cef021092fe-tls-assets\") pod \"alertmanager-main-0\" (UID: \"d0fb95e9-9ac7-46bf-996a-0cef021092fe\") " pod="openshift-monitoring/alertmanager-main-0" Nov 21 14:14:00 crc kubenswrapper[4897]: I1121 14:14:00.953393 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"alertmanager-main-db\" (UniqueName: \"kubernetes.io/empty-dir/d0fb95e9-9ac7-46bf-996a-0cef021092fe-alertmanager-main-db\") pod \"alertmanager-main-0\" (UID: \"d0fb95e9-9ac7-46bf-996a-0cef021092fe\") " pod="openshift-monitoring/alertmanager-main-0" Nov 21 14:14:00 crc kubenswrapper[4897]: I1121 14:14:00.953421 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sz6z9\" (UniqueName: \"kubernetes.io/projected/d0fb95e9-9ac7-46bf-996a-0cef021092fe-kube-api-access-sz6z9\") pod \"alertmanager-main-0\" (UID: \"d0fb95e9-9ac7-46bf-996a-0cef021092fe\") " pod="openshift-monitoring/alertmanager-main-0" Nov 21 14:14:00 crc kubenswrapper[4897]: I1121 14:14:00.953535 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"alertmanager-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d0fb95e9-9ac7-46bf-996a-0cef021092fe-alertmanager-trusted-ca-bundle\") pod \"alertmanager-main-0\" (UID: \"d0fb95e9-9ac7-46bf-996a-0cef021092fe\") " pod="openshift-monitoring/alertmanager-main-0" Nov 21 14:14:00 crc kubenswrapper[4897]: I1121 14:14:00.953553 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-main-tls\" (UniqueName: \"kubernetes.io/secret/d0fb95e9-9ac7-46bf-996a-0cef021092fe-secret-alertmanager-main-tls\") pod \"alertmanager-main-0\" (UID: \"d0fb95e9-9ac7-46bf-996a-0cef021092fe\") " pod="openshift-monitoring/alertmanager-main-0" Nov 21 14:14:00 crc kubenswrapper[4897]: I1121 14:14:00.953582 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-kube-rbac-proxy-metric\" (UniqueName: \"kubernetes.io/secret/d0fb95e9-9ac7-46bf-996a-0cef021092fe-secret-alertmanager-kube-rbac-proxy-metric\") pod \"alertmanager-main-0\" (UID: \"d0fb95e9-9ac7-46bf-996a-0cef021092fe\") " pod="openshift-monitoring/alertmanager-main-0" Nov 21 14:14:01 crc kubenswrapper[4897]: I1121 14:14:01.054994 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-alertmanager-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/d0fb95e9-9ac7-46bf-996a-0cef021092fe-secret-alertmanager-kube-rbac-proxy-web\") pod \"alertmanager-main-0\" (UID: \"d0fb95e9-9ac7-46bf-996a-0cef021092fe\") " pod="openshift-monitoring/alertmanager-main-0" Nov 21 14:14:01 crc kubenswrapper[4897]: I1121 14:14:01.055066 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/d0fb95e9-9ac7-46bf-996a-0cef021092fe-config-out\") pod \"alertmanager-main-0\" (UID: \"d0fb95e9-9ac7-46bf-996a-0cef021092fe\") " pod="openshift-monitoring/alertmanager-main-0" Nov 21 14:14:01 crc kubenswrapper[4897]: I1121 14:14:01.055094 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/d0fb95e9-9ac7-46bf-996a-0cef021092fe-web-config\") pod \"alertmanager-main-0\" (UID: \"d0fb95e9-9ac7-46bf-996a-0cef021092fe\") " pod="openshift-monitoring/alertmanager-main-0" Nov 21 14:14:01 crc kubenswrapper[4897]: I1121 14:14:01.055120 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/d0fb95e9-9ac7-46bf-996a-0cef021092fe-metrics-client-ca\") pod \"alertmanager-main-0\" (UID: \"d0fb95e9-9ac7-46bf-996a-0cef021092fe\") " pod="openshift-monitoring/alertmanager-main-0" Nov 21 14:14:01 crc kubenswrapper[4897]: I1121 14:14:01.055148 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-alertmanager-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/d0fb95e9-9ac7-46bf-996a-0cef021092fe-secret-alertmanager-kube-rbac-proxy\") pod \"alertmanager-main-0\" (UID: \"d0fb95e9-9ac7-46bf-996a-0cef021092fe\") " pod="openshift-monitoring/alertmanager-main-0" Nov 21 14:14:01 crc kubenswrapper[4897]: I1121 14:14:01.055166 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/d0fb95e9-9ac7-46bf-996a-0cef021092fe-config-volume\") pod \"alertmanager-main-0\" (UID: \"d0fb95e9-9ac7-46bf-996a-0cef021092fe\") " pod="openshift-monitoring/alertmanager-main-0" Nov 21 14:14:01 crc kubenswrapper[4897]: I1121 14:14:01.055187 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/d0fb95e9-9ac7-46bf-996a-0cef021092fe-tls-assets\") pod \"alertmanager-main-0\" (UID: \"d0fb95e9-9ac7-46bf-996a-0cef021092fe\") " pod="openshift-monitoring/alertmanager-main-0" Nov 21 14:14:01 crc kubenswrapper[4897]: I1121 14:14:01.055231 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"alertmanager-main-db\" (UniqueName: \"kubernetes.io/empty-dir/d0fb95e9-9ac7-46bf-996a-0cef021092fe-alertmanager-main-db\") pod \"alertmanager-main-0\" (UID: \"d0fb95e9-9ac7-46bf-996a-0cef021092fe\") " pod="openshift-monitoring/alertmanager-main-0" Nov 21 14:14:01 crc kubenswrapper[4897]: I1121 14:14:01.055259 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sz6z9\" (UniqueName: \"kubernetes.io/projected/d0fb95e9-9ac7-46bf-996a-0cef021092fe-kube-api-access-sz6z9\") pod \"alertmanager-main-0\" (UID: \"d0fb95e9-9ac7-46bf-996a-0cef021092fe\") " pod="openshift-monitoring/alertmanager-main-0" Nov 21 14:14:01 crc kubenswrapper[4897]: I1121 14:14:01.055282 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"alertmanager-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d0fb95e9-9ac7-46bf-996a-0cef021092fe-alertmanager-trusted-ca-bundle\") pod \"alertmanager-main-0\" (UID: \"d0fb95e9-9ac7-46bf-996a-0cef021092fe\") " pod="openshift-monitoring/alertmanager-main-0" Nov 21 14:14:01 crc kubenswrapper[4897]: I1121 14:14:01.055310 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-alertmanager-main-tls\" (UniqueName: \"kubernetes.io/secret/d0fb95e9-9ac7-46bf-996a-0cef021092fe-secret-alertmanager-main-tls\") pod \"alertmanager-main-0\" (UID: \"d0fb95e9-9ac7-46bf-996a-0cef021092fe\") " pod="openshift-monitoring/alertmanager-main-0" Nov 21 14:14:01 crc kubenswrapper[4897]: I1121 14:14:01.055345 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-alertmanager-kube-rbac-proxy-metric\" (UniqueName: \"kubernetes.io/secret/d0fb95e9-9ac7-46bf-996a-0cef021092fe-secret-alertmanager-kube-rbac-proxy-metric\") pod \"alertmanager-main-0\" (UID: \"d0fb95e9-9ac7-46bf-996a-0cef021092fe\") " pod="openshift-monitoring/alertmanager-main-0" Nov 21 14:14:01 crc kubenswrapper[4897]: I1121 14:14:01.057425 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/d0fb95e9-9ac7-46bf-996a-0cef021092fe-metrics-client-ca\") pod \"alertmanager-main-0\" (UID: \"d0fb95e9-9ac7-46bf-996a-0cef021092fe\") " pod="openshift-monitoring/alertmanager-main-0" Nov 21 14:14:01 crc kubenswrapper[4897]: I1121 14:14:01.057878 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"alertmanager-main-db\" (UniqueName: \"kubernetes.io/empty-dir/d0fb95e9-9ac7-46bf-996a-0cef021092fe-alertmanager-main-db\") pod \"alertmanager-main-0\" (UID: \"d0fb95e9-9ac7-46bf-996a-0cef021092fe\") " pod="openshift-monitoring/alertmanager-main-0" Nov 21 14:14:01 crc kubenswrapper[4897]: I1121 14:14:01.057946 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"alertmanager-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d0fb95e9-9ac7-46bf-996a-0cef021092fe-alertmanager-trusted-ca-bundle\") pod \"alertmanager-main-0\" (UID: \"d0fb95e9-9ac7-46bf-996a-0cef021092fe\") " pod="openshift-monitoring/alertmanager-main-0" Nov 21 14:14:01 crc kubenswrapper[4897]: I1121 14:14:01.060331 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/d0fb95e9-9ac7-46bf-996a-0cef021092fe-tls-assets\") pod \"alertmanager-main-0\" (UID: \"d0fb95e9-9ac7-46bf-996a-0cef021092fe\") " pod="openshift-monitoring/alertmanager-main-0" Nov 21 14:14:01 crc kubenswrapper[4897]: I1121 14:14:01.060343 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/d0fb95e9-9ac7-46bf-996a-0cef021092fe-secret-alertmanager-kube-rbac-proxy\") pod \"alertmanager-main-0\" (UID: \"d0fb95e9-9ac7-46bf-996a-0cef021092fe\") " pod="openshift-monitoring/alertmanager-main-0" Nov 21 14:14:01 crc kubenswrapper[4897]: I1121 14:14:01.060828 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-main-tls\" (UniqueName: \"kubernetes.io/secret/d0fb95e9-9ac7-46bf-996a-0cef021092fe-secret-alertmanager-main-tls\") pod \"alertmanager-main-0\" (UID: \"d0fb95e9-9ac7-46bf-996a-0cef021092fe\") " pod="openshift-monitoring/alertmanager-main-0" Nov 21 14:14:01 crc kubenswrapper[4897]: I1121 14:14:01.061233 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/d0fb95e9-9ac7-46bf-996a-0cef021092fe-web-config\") pod \"alertmanager-main-0\" (UID: \"d0fb95e9-9ac7-46bf-996a-0cef021092fe\") " pod="openshift-monitoring/alertmanager-main-0" Nov 21 14:14:01 crc kubenswrapper[4897]: I1121 14:14:01.061447 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-kube-rbac-proxy-metric\" (UniqueName: \"kubernetes.io/secret/d0fb95e9-9ac7-46bf-996a-0cef021092fe-secret-alertmanager-kube-rbac-proxy-metric\") pod \"alertmanager-main-0\" (UID: \"d0fb95e9-9ac7-46bf-996a-0cef021092fe\") " pod="openshift-monitoring/alertmanager-main-0" Nov 21 14:14:01 crc kubenswrapper[4897]: I1121 14:14:01.062250 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/d0fb95e9-9ac7-46bf-996a-0cef021092fe-config-volume\") pod \"alertmanager-main-0\" (UID: \"d0fb95e9-9ac7-46bf-996a-0cef021092fe\") " pod="openshift-monitoring/alertmanager-main-0" Nov 21 14:14:01 crc kubenswrapper[4897]: I1121 14:14:01.065087 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/d0fb95e9-9ac7-46bf-996a-0cef021092fe-secret-alertmanager-kube-rbac-proxy-web\") pod \"alertmanager-main-0\" (UID: \"d0fb95e9-9ac7-46bf-996a-0cef021092fe\") " pod="openshift-monitoring/alertmanager-main-0" Nov 21 14:14:01 crc kubenswrapper[4897]: I1121 14:14:01.073675 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/d0fb95e9-9ac7-46bf-996a-0cef021092fe-config-out\") pod \"alertmanager-main-0\" (UID: \"d0fb95e9-9ac7-46bf-996a-0cef021092fe\") " pod="openshift-monitoring/alertmanager-main-0" Nov 21 14:14:01 crc kubenswrapper[4897]: I1121 14:14:01.074435 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sz6z9\" (UniqueName: \"kubernetes.io/projected/d0fb95e9-9ac7-46bf-996a-0cef021092fe-kube-api-access-sz6z9\") pod \"alertmanager-main-0\" (UID: \"d0fb95e9-9ac7-46bf-996a-0cef021092fe\") " pod="openshift-monitoring/alertmanager-main-0" Nov 21 14:14:01 crc kubenswrapper[4897]: I1121 14:14:01.115293 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/openshift-state-metrics-566fddb674-hw2hf"] Nov 21 14:14:01 crc kubenswrapper[4897]: W1121 14:14:01.121714 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8d0ea2ea_c9a1_495a_a626_7a1a735308df.slice/crio-55c66449bd74bbc9be3704c9733a21ddfcd8bc952939588a130f188fd7e49355 WatchSource:0}: Error finding container 55c66449bd74bbc9be3704c9733a21ddfcd8bc952939588a130f188fd7e49355: Status 404 returned error can't find the container with id 55c66449bd74bbc9be3704c9733a21ddfcd8bc952939588a130f188fd7e49355 Nov 21 14:14:01 crc kubenswrapper[4897]: I1121 14:14:01.158409 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/alertmanager-main-0" Nov 21 14:14:01 crc kubenswrapper[4897]: I1121 14:14:01.255463 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/kube-state-metrics-777cb5bd5d-tm69l"] Nov 21 14:14:01 crc kubenswrapper[4897]: W1121 14:14:01.263061 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod91c21c9f_aed2_41c3_b3fa_371f9903350a.slice/crio-3dd4e51edae07dda161f65f6aa51e825d61310f35157542f9cfdb33bc7781819 WatchSource:0}: Error finding container 3dd4e51edae07dda161f65f6aa51e825d61310f35157542f9cfdb33bc7781819: Status 404 returned error can't find the container with id 3dd4e51edae07dda161f65f6aa51e825d61310f35157542f9cfdb33bc7781819 Nov 21 14:14:01 crc kubenswrapper[4897]: I1121 14:14:01.338740 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/alertmanager-main-0"] Nov 21 14:14:01 crc kubenswrapper[4897]: W1121 14:14:01.343434 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd0fb95e9_9ac7_46bf_996a_0cef021092fe.slice/crio-8c1aa6c346d83b7250c792f88a153603b5df5e4a0b10758e582e08886d6f9b2a WatchSource:0}: Error finding container 8c1aa6c346d83b7250c792f88a153603b5df5e4a0b10758e582e08886d6f9b2a: Status 404 returned error can't find the container with id 8c1aa6c346d83b7250c792f88a153603b5df5e4a0b10758e582e08886d6f9b2a Nov 21 14:14:01 crc kubenswrapper[4897]: I1121 14:14:01.779707 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-tm69l" event={"ID":"91c21c9f-aed2-41c3-b3fa-371f9903350a","Type":"ContainerStarted","Data":"3dd4e51edae07dda161f65f6aa51e825d61310f35157542f9cfdb33bc7781819"} Nov 21 14:14:01 crc kubenswrapper[4897]: I1121 14:14:01.780727 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/openshift-state-metrics-566fddb674-hw2hf" event={"ID":"8d0ea2ea-c9a1-495a-a626-7a1a735308df","Type":"ContainerStarted","Data":"55c66449bd74bbc9be3704c9733a21ddfcd8bc952939588a130f188fd7e49355"} Nov 21 14:14:01 crc kubenswrapper[4897]: I1121 14:14:01.781642 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"d0fb95e9-9ac7-46bf-996a-0cef021092fe","Type":"ContainerStarted","Data":"8c1aa6c346d83b7250c792f88a153603b5df5e4a0b10758e582e08886d6f9b2a"} Nov 21 14:14:01 crc kubenswrapper[4897]: I1121 14:14:01.782455 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-9gpbl" event={"ID":"8caaec7a-8eab-4ffe-9e68-f5c763d044eb","Type":"ContainerStarted","Data":"54c9a89f8adb9c18d5b143a48df4eb7d4fcbe9a22712e8a702409e61614406a8"} Nov 21 14:14:01 crc kubenswrapper[4897]: I1121 14:14:01.847395 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/thanos-querier-7c6484df6-g5h9n"] Nov 21 14:14:01 crc kubenswrapper[4897]: I1121 14:14:01.849544 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/thanos-querier-7c6484df6-g5h9n" Nov 21 14:14:01 crc kubenswrapper[4897]: I1121 14:14:01.850899 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"thanos-querier-kube-rbac-proxy-metrics" Nov 21 14:14:01 crc kubenswrapper[4897]: I1121 14:14:01.852017 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"thanos-querier-kube-rbac-proxy-web" Nov 21 14:14:01 crc kubenswrapper[4897]: I1121 14:14:01.852050 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"thanos-querier-kube-rbac-proxy" Nov 21 14:14:01 crc kubenswrapper[4897]: I1121 14:14:01.852171 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"thanos-querier-tls" Nov 21 14:14:01 crc kubenswrapper[4897]: I1121 14:14:01.854859 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"thanos-querier-kube-rbac-proxy-rules" Nov 21 14:14:01 crc kubenswrapper[4897]: I1121 14:14:01.855541 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"thanos-querier-grpc-tls-9l01duj3cpf64" Nov 21 14:14:01 crc kubenswrapper[4897]: I1121 14:14:01.855759 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"thanos-querier-dockercfg-z4f42" Nov 21 14:14:01 crc kubenswrapper[4897]: I1121 14:14:01.867608 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/thanos-querier-7c6484df6-g5h9n"] Nov 21 14:14:01 crc kubenswrapper[4897]: I1121 14:14:01.968394 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-tls\" (UniqueName: \"kubernetes.io/secret/7ab7655b-d56b-4ba0-982a-c4649864016e-secret-thanos-querier-tls\") pod \"thanos-querier-7c6484df6-g5h9n\" (UID: \"7ab7655b-d56b-4ba0-982a-c4649864016e\") " pod="openshift-monitoring/thanos-querier-7c6484df6-g5h9n" Nov 21 14:14:01 crc kubenswrapper[4897]: I1121 14:14:01.968453 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-metrics\" (UniqueName: \"kubernetes.io/secret/7ab7655b-d56b-4ba0-982a-c4649864016e-secret-thanos-querier-kube-rbac-proxy-metrics\") pod \"thanos-querier-7c6484df6-g5h9n\" (UID: \"7ab7655b-d56b-4ba0-982a-c4649864016e\") " pod="openshift-monitoring/thanos-querier-7c6484df6-g5h9n" Nov 21 14:14:01 crc kubenswrapper[4897]: I1121 14:14:01.968476 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/7ab7655b-d56b-4ba0-982a-c4649864016e-metrics-client-ca\") pod \"thanos-querier-7c6484df6-g5h9n\" (UID: \"7ab7655b-d56b-4ba0-982a-c4649864016e\") " pod="openshift-monitoring/thanos-querier-7c6484df6-g5h9n" Nov 21 14:14:01 crc kubenswrapper[4897]: I1121 14:14:01.968494 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9s956\" (UniqueName: \"kubernetes.io/projected/7ab7655b-d56b-4ba0-982a-c4649864016e-kube-api-access-9s956\") pod \"thanos-querier-7c6484df6-g5h9n\" (UID: \"7ab7655b-d56b-4ba0-982a-c4649864016e\") " pod="openshift-monitoring/thanos-querier-7c6484df6-g5h9n" Nov 21 14:14:01 crc kubenswrapper[4897]: I1121 14:14:01.968544 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-rules\" (UniqueName: \"kubernetes.io/secret/7ab7655b-d56b-4ba0-982a-c4649864016e-secret-thanos-querier-kube-rbac-proxy-rules\") pod \"thanos-querier-7c6484df6-g5h9n\" (UID: \"7ab7655b-d56b-4ba0-982a-c4649864016e\") " pod="openshift-monitoring/thanos-querier-7c6484df6-g5h9n" Nov 21 14:14:01 crc kubenswrapper[4897]: I1121 14:14:01.968576 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/7ab7655b-d56b-4ba0-982a-c4649864016e-secret-grpc-tls\") pod \"thanos-querier-7c6484df6-g5h9n\" (UID: \"7ab7655b-d56b-4ba0-982a-c4649864016e\") " pod="openshift-monitoring/thanos-querier-7c6484df6-g5h9n" Nov 21 14:14:01 crc kubenswrapper[4897]: I1121 14:14:01.968604 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/7ab7655b-d56b-4ba0-982a-c4649864016e-secret-thanos-querier-kube-rbac-proxy\") pod \"thanos-querier-7c6484df6-g5h9n\" (UID: \"7ab7655b-d56b-4ba0-982a-c4649864016e\") " pod="openshift-monitoring/thanos-querier-7c6484df6-g5h9n" Nov 21 14:14:01 crc kubenswrapper[4897]: I1121 14:14:01.968620 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/7ab7655b-d56b-4ba0-982a-c4649864016e-secret-thanos-querier-kube-rbac-proxy-web\") pod \"thanos-querier-7c6484df6-g5h9n\" (UID: \"7ab7655b-d56b-4ba0-982a-c4649864016e\") " pod="openshift-monitoring/thanos-querier-7c6484df6-g5h9n" Nov 21 14:14:02 crc kubenswrapper[4897]: I1121 14:14:02.070427 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-rules\" (UniqueName: \"kubernetes.io/secret/7ab7655b-d56b-4ba0-982a-c4649864016e-secret-thanos-querier-kube-rbac-proxy-rules\") pod \"thanos-querier-7c6484df6-g5h9n\" (UID: \"7ab7655b-d56b-4ba0-982a-c4649864016e\") " pod="openshift-monitoring/thanos-querier-7c6484df6-g5h9n" Nov 21 14:14:02 crc kubenswrapper[4897]: I1121 14:14:02.070593 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/7ab7655b-d56b-4ba0-982a-c4649864016e-secret-grpc-tls\") pod \"thanos-querier-7c6484df6-g5h9n\" (UID: \"7ab7655b-d56b-4ba0-982a-c4649864016e\") " pod="openshift-monitoring/thanos-querier-7c6484df6-g5h9n" Nov 21 14:14:02 crc kubenswrapper[4897]: I1121 14:14:02.070690 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/7ab7655b-d56b-4ba0-982a-c4649864016e-secret-thanos-querier-kube-rbac-proxy\") pod \"thanos-querier-7c6484df6-g5h9n\" (UID: \"7ab7655b-d56b-4ba0-982a-c4649864016e\") " pod="openshift-monitoring/thanos-querier-7c6484df6-g5h9n" Nov 21 14:14:02 crc kubenswrapper[4897]: I1121 14:14:02.070741 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/7ab7655b-d56b-4ba0-982a-c4649864016e-secret-thanos-querier-kube-rbac-proxy-web\") pod \"thanos-querier-7c6484df6-g5h9n\" (UID: \"7ab7655b-d56b-4ba0-982a-c4649864016e\") " pod="openshift-monitoring/thanos-querier-7c6484df6-g5h9n" Nov 21 14:14:02 crc kubenswrapper[4897]: I1121 14:14:02.070828 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-tls\" (UniqueName: \"kubernetes.io/secret/7ab7655b-d56b-4ba0-982a-c4649864016e-secret-thanos-querier-tls\") pod \"thanos-querier-7c6484df6-g5h9n\" (UID: \"7ab7655b-d56b-4ba0-982a-c4649864016e\") " pod="openshift-monitoring/thanos-querier-7c6484df6-g5h9n" Nov 21 14:14:02 crc kubenswrapper[4897]: I1121 14:14:02.070922 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-metrics\" (UniqueName: \"kubernetes.io/secret/7ab7655b-d56b-4ba0-982a-c4649864016e-secret-thanos-querier-kube-rbac-proxy-metrics\") pod \"thanos-querier-7c6484df6-g5h9n\" (UID: \"7ab7655b-d56b-4ba0-982a-c4649864016e\") " pod="openshift-monitoring/thanos-querier-7c6484df6-g5h9n" Nov 21 14:14:02 crc kubenswrapper[4897]: I1121 14:14:02.070979 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/7ab7655b-d56b-4ba0-982a-c4649864016e-metrics-client-ca\") pod \"thanos-querier-7c6484df6-g5h9n\" (UID: \"7ab7655b-d56b-4ba0-982a-c4649864016e\") " pod="openshift-monitoring/thanos-querier-7c6484df6-g5h9n" Nov 21 14:14:02 crc kubenswrapper[4897]: I1121 14:14:02.071027 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9s956\" (UniqueName: \"kubernetes.io/projected/7ab7655b-d56b-4ba0-982a-c4649864016e-kube-api-access-9s956\") pod \"thanos-querier-7c6484df6-g5h9n\" (UID: \"7ab7655b-d56b-4ba0-982a-c4649864016e\") " pod="openshift-monitoring/thanos-querier-7c6484df6-g5h9n" Nov 21 14:14:02 crc kubenswrapper[4897]: I1121 14:14:02.074565 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/7ab7655b-d56b-4ba0-982a-c4649864016e-metrics-client-ca\") pod \"thanos-querier-7c6484df6-g5h9n\" (UID: \"7ab7655b-d56b-4ba0-982a-c4649864016e\") " pod="openshift-monitoring/thanos-querier-7c6484df6-g5h9n" Nov 21 14:14:02 crc kubenswrapper[4897]: I1121 14:14:02.079925 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-kube-rbac-proxy-metrics\" (UniqueName: \"kubernetes.io/secret/7ab7655b-d56b-4ba0-982a-c4649864016e-secret-thanos-querier-kube-rbac-proxy-metrics\") pod \"thanos-querier-7c6484df6-g5h9n\" (UID: \"7ab7655b-d56b-4ba0-982a-c4649864016e\") " pod="openshift-monitoring/thanos-querier-7c6484df6-g5h9n" Nov 21 14:14:02 crc kubenswrapper[4897]: I1121 14:14:02.079934 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/7ab7655b-d56b-4ba0-982a-c4649864016e-secret-thanos-querier-kube-rbac-proxy-web\") pod \"thanos-querier-7c6484df6-g5h9n\" (UID: \"7ab7655b-d56b-4ba0-982a-c4649864016e\") " pod="openshift-monitoring/thanos-querier-7c6484df6-g5h9n" Nov 21 14:14:02 crc kubenswrapper[4897]: I1121 14:14:02.080648 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/7ab7655b-d56b-4ba0-982a-c4649864016e-secret-grpc-tls\") pod \"thanos-querier-7c6484df6-g5h9n\" (UID: \"7ab7655b-d56b-4ba0-982a-c4649864016e\") " pod="openshift-monitoring/thanos-querier-7c6484df6-g5h9n" Nov 21 14:14:02 crc kubenswrapper[4897]: I1121 14:14:02.080681 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-tls\" (UniqueName: \"kubernetes.io/secret/7ab7655b-d56b-4ba0-982a-c4649864016e-secret-thanos-querier-tls\") pod \"thanos-querier-7c6484df6-g5h9n\" (UID: \"7ab7655b-d56b-4ba0-982a-c4649864016e\") " pod="openshift-monitoring/thanos-querier-7c6484df6-g5h9n" Nov 21 14:14:02 crc kubenswrapper[4897]: I1121 14:14:02.080699 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/7ab7655b-d56b-4ba0-982a-c4649864016e-secret-thanos-querier-kube-rbac-proxy\") pod \"thanos-querier-7c6484df6-g5h9n\" (UID: \"7ab7655b-d56b-4ba0-982a-c4649864016e\") " pod="openshift-monitoring/thanos-querier-7c6484df6-g5h9n" Nov 21 14:14:02 crc kubenswrapper[4897]: I1121 14:14:02.080681 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-kube-rbac-proxy-rules\" (UniqueName: \"kubernetes.io/secret/7ab7655b-d56b-4ba0-982a-c4649864016e-secret-thanos-querier-kube-rbac-proxy-rules\") pod \"thanos-querier-7c6484df6-g5h9n\" (UID: \"7ab7655b-d56b-4ba0-982a-c4649864016e\") " pod="openshift-monitoring/thanos-querier-7c6484df6-g5h9n" Nov 21 14:14:02 crc kubenswrapper[4897]: I1121 14:14:02.090489 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9s956\" (UniqueName: \"kubernetes.io/projected/7ab7655b-d56b-4ba0-982a-c4649864016e-kube-api-access-9s956\") pod \"thanos-querier-7c6484df6-g5h9n\" (UID: \"7ab7655b-d56b-4ba0-982a-c4649864016e\") " pod="openshift-monitoring/thanos-querier-7c6484df6-g5h9n" Nov 21 14:14:02 crc kubenswrapper[4897]: I1121 14:14:02.169639 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/thanos-querier-7c6484df6-g5h9n" Nov 21 14:14:02 crc kubenswrapper[4897]: I1121 14:14:02.382255 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/thanos-querier-7c6484df6-g5h9n"] Nov 21 14:14:02 crc kubenswrapper[4897]: I1121 14:14:02.789818 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-7c6484df6-g5h9n" event={"ID":"7ab7655b-d56b-4ba0-982a-c4649864016e","Type":"ContainerStarted","Data":"6540b5dab969b28c57bb04aa7452b64de2912203eb830f1922d8247625e06b5d"} Nov 21 14:14:05 crc kubenswrapper[4897]: I1121 14:14:05.160535 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/metrics-server-586c8bb899-r6zbz"] Nov 21 14:14:05 crc kubenswrapper[4897]: I1121 14:14:05.162381 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/metrics-server-586c8bb899-r6zbz" Nov 21 14:14:05 crc kubenswrapper[4897]: I1121 14:14:05.165721 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"metrics-server-audit-profiles" Nov 21 14:14:05 crc kubenswrapper[4897]: I1121 14:14:05.166952 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"metrics-client-certs" Nov 21 14:14:05 crc kubenswrapper[4897]: I1121 14:14:05.167256 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"kubelet-serving-ca-bundle" Nov 21 14:14:05 crc kubenswrapper[4897]: I1121 14:14:05.167555 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"metrics-server-tls" Nov 21 14:14:05 crc kubenswrapper[4897]: I1121 14:14:05.169426 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"metrics-server-dockercfg-j6nwb" Nov 21 14:14:05 crc kubenswrapper[4897]: I1121 14:14:05.173140 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/metrics-server-586c8bb899-r6zbz"] Nov 21 14:14:05 crc kubenswrapper[4897]: I1121 14:14:05.177420 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"metrics-server-2sh6n4btmg6no" Nov 21 14:14:05 crc kubenswrapper[4897]: I1121 14:14:05.315398 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97e1c0fe-f3e3-4afe-819e-eb85875bee86-client-ca-bundle\") pod \"metrics-server-586c8bb899-r6zbz\" (UID: \"97e1c0fe-f3e3-4afe-819e-eb85875bee86\") " pod="openshift-monitoring/metrics-server-586c8bb899-r6zbz" Nov 21 14:14:05 crc kubenswrapper[4897]: I1121 14:14:05.315469 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-server-audit-profiles\" (UniqueName: \"kubernetes.io/configmap/97e1c0fe-f3e3-4afe-819e-eb85875bee86-metrics-server-audit-profiles\") pod \"metrics-server-586c8bb899-r6zbz\" (UID: \"97e1c0fe-f3e3-4afe-819e-eb85875bee86\") " pod="openshift-monitoring/metrics-server-586c8bb899-r6zbz" Nov 21 14:14:05 crc kubenswrapper[4897]: I1121 14:14:05.315541 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-log\" (UniqueName: \"kubernetes.io/empty-dir/97e1c0fe-f3e3-4afe-819e-eb85875bee86-audit-log\") pod \"metrics-server-586c8bb899-r6zbz\" (UID: \"97e1c0fe-f3e3-4afe-819e-eb85875bee86\") " pod="openshift-monitoring/metrics-server-586c8bb899-r6zbz" Nov 21 14:14:05 crc kubenswrapper[4897]: I1121 14:14:05.315653 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nz7p7\" (UniqueName: \"kubernetes.io/projected/97e1c0fe-f3e3-4afe-819e-eb85875bee86-kube-api-access-nz7p7\") pod \"metrics-server-586c8bb899-r6zbz\" (UID: \"97e1c0fe-f3e3-4afe-819e-eb85875bee86\") " pod="openshift-monitoring/metrics-server-586c8bb899-r6zbz" Nov 21 14:14:05 crc kubenswrapper[4897]: I1121 14:14:05.315754 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-metrics-server-tls\" (UniqueName: \"kubernetes.io/secret/97e1c0fe-f3e3-4afe-819e-eb85875bee86-secret-metrics-server-tls\") pod \"metrics-server-586c8bb899-r6zbz\" (UID: \"97e1c0fe-f3e3-4afe-819e-eb85875bee86\") " pod="openshift-monitoring/metrics-server-586c8bb899-r6zbz" Nov 21 14:14:05 crc kubenswrapper[4897]: I1121 14:14:05.315894 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/97e1c0fe-f3e3-4afe-819e-eb85875bee86-configmap-kubelet-serving-ca-bundle\") pod \"metrics-server-586c8bb899-r6zbz\" (UID: \"97e1c0fe-f3e3-4afe-819e-eb85875bee86\") " pod="openshift-monitoring/metrics-server-586c8bb899-r6zbz" Nov 21 14:14:05 crc kubenswrapper[4897]: I1121 14:14:05.316001 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/97e1c0fe-f3e3-4afe-819e-eb85875bee86-secret-metrics-client-certs\") pod \"metrics-server-586c8bb899-r6zbz\" (UID: \"97e1c0fe-f3e3-4afe-819e-eb85875bee86\") " pod="openshift-monitoring/metrics-server-586c8bb899-r6zbz" Nov 21 14:14:05 crc kubenswrapper[4897]: I1121 14:14:05.417391 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97e1c0fe-f3e3-4afe-819e-eb85875bee86-client-ca-bundle\") pod \"metrics-server-586c8bb899-r6zbz\" (UID: \"97e1c0fe-f3e3-4afe-819e-eb85875bee86\") " pod="openshift-monitoring/metrics-server-586c8bb899-r6zbz" Nov 21 14:14:05 crc kubenswrapper[4897]: I1121 14:14:05.417436 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-server-audit-profiles\" (UniqueName: \"kubernetes.io/configmap/97e1c0fe-f3e3-4afe-819e-eb85875bee86-metrics-server-audit-profiles\") pod \"metrics-server-586c8bb899-r6zbz\" (UID: \"97e1c0fe-f3e3-4afe-819e-eb85875bee86\") " pod="openshift-monitoring/metrics-server-586c8bb899-r6zbz" Nov 21 14:14:05 crc kubenswrapper[4897]: I1121 14:14:05.417478 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-log\" (UniqueName: \"kubernetes.io/empty-dir/97e1c0fe-f3e3-4afe-819e-eb85875bee86-audit-log\") pod \"metrics-server-586c8bb899-r6zbz\" (UID: \"97e1c0fe-f3e3-4afe-819e-eb85875bee86\") " pod="openshift-monitoring/metrics-server-586c8bb899-r6zbz" Nov 21 14:14:05 crc kubenswrapper[4897]: I1121 14:14:05.417525 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nz7p7\" (UniqueName: \"kubernetes.io/projected/97e1c0fe-f3e3-4afe-819e-eb85875bee86-kube-api-access-nz7p7\") pod \"metrics-server-586c8bb899-r6zbz\" (UID: \"97e1c0fe-f3e3-4afe-819e-eb85875bee86\") " pod="openshift-monitoring/metrics-server-586c8bb899-r6zbz" Nov 21 14:14:05 crc kubenswrapper[4897]: I1121 14:14:05.417558 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-metrics-server-tls\" (UniqueName: \"kubernetes.io/secret/97e1c0fe-f3e3-4afe-819e-eb85875bee86-secret-metrics-server-tls\") pod \"metrics-server-586c8bb899-r6zbz\" (UID: \"97e1c0fe-f3e3-4afe-819e-eb85875bee86\") " pod="openshift-monitoring/metrics-server-586c8bb899-r6zbz" Nov 21 14:14:05 crc kubenswrapper[4897]: I1121 14:14:05.417616 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/97e1c0fe-f3e3-4afe-819e-eb85875bee86-configmap-kubelet-serving-ca-bundle\") pod \"metrics-server-586c8bb899-r6zbz\" (UID: \"97e1c0fe-f3e3-4afe-819e-eb85875bee86\") " pod="openshift-monitoring/metrics-server-586c8bb899-r6zbz" Nov 21 14:14:05 crc kubenswrapper[4897]: I1121 14:14:05.417662 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/97e1c0fe-f3e3-4afe-819e-eb85875bee86-secret-metrics-client-certs\") pod \"metrics-server-586c8bb899-r6zbz\" (UID: \"97e1c0fe-f3e3-4afe-819e-eb85875bee86\") " pod="openshift-monitoring/metrics-server-586c8bb899-r6zbz" Nov 21 14:14:05 crc kubenswrapper[4897]: I1121 14:14:05.418617 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-log\" (UniqueName: \"kubernetes.io/empty-dir/97e1c0fe-f3e3-4afe-819e-eb85875bee86-audit-log\") pod \"metrics-server-586c8bb899-r6zbz\" (UID: \"97e1c0fe-f3e3-4afe-819e-eb85875bee86\") " pod="openshift-monitoring/metrics-server-586c8bb899-r6zbz" Nov 21 14:14:05 crc kubenswrapper[4897]: I1121 14:14:05.418712 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/97e1c0fe-f3e3-4afe-819e-eb85875bee86-configmap-kubelet-serving-ca-bundle\") pod \"metrics-server-586c8bb899-r6zbz\" (UID: \"97e1c0fe-f3e3-4afe-819e-eb85875bee86\") " pod="openshift-monitoring/metrics-server-586c8bb899-r6zbz" Nov 21 14:14:05 crc kubenswrapper[4897]: I1121 14:14:05.420430 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-server-audit-profiles\" (UniqueName: \"kubernetes.io/configmap/97e1c0fe-f3e3-4afe-819e-eb85875bee86-metrics-server-audit-profiles\") pod \"metrics-server-586c8bb899-r6zbz\" (UID: \"97e1c0fe-f3e3-4afe-819e-eb85875bee86\") " pod="openshift-monitoring/metrics-server-586c8bb899-r6zbz" Nov 21 14:14:05 crc kubenswrapper[4897]: I1121 14:14:05.423378 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/97e1c0fe-f3e3-4afe-819e-eb85875bee86-secret-metrics-client-certs\") pod \"metrics-server-586c8bb899-r6zbz\" (UID: \"97e1c0fe-f3e3-4afe-819e-eb85875bee86\") " pod="openshift-monitoring/metrics-server-586c8bb899-r6zbz" Nov 21 14:14:05 crc kubenswrapper[4897]: I1121 14:14:05.423822 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-metrics-server-tls\" (UniqueName: \"kubernetes.io/secret/97e1c0fe-f3e3-4afe-819e-eb85875bee86-secret-metrics-server-tls\") pod \"metrics-server-586c8bb899-r6zbz\" (UID: \"97e1c0fe-f3e3-4afe-819e-eb85875bee86\") " pod="openshift-monitoring/metrics-server-586c8bb899-r6zbz" Nov 21 14:14:05 crc kubenswrapper[4897]: I1121 14:14:05.425617 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97e1c0fe-f3e3-4afe-819e-eb85875bee86-client-ca-bundle\") pod \"metrics-server-586c8bb899-r6zbz\" (UID: \"97e1c0fe-f3e3-4afe-819e-eb85875bee86\") " pod="openshift-monitoring/metrics-server-586c8bb899-r6zbz" Nov 21 14:14:05 crc kubenswrapper[4897]: I1121 14:14:05.433779 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nz7p7\" (UniqueName: \"kubernetes.io/projected/97e1c0fe-f3e3-4afe-819e-eb85875bee86-kube-api-access-nz7p7\") pod \"metrics-server-586c8bb899-r6zbz\" (UID: \"97e1c0fe-f3e3-4afe-819e-eb85875bee86\") " pod="openshift-monitoring/metrics-server-586c8bb899-r6zbz" Nov 21 14:14:05 crc kubenswrapper[4897]: I1121 14:14:05.478283 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/metrics-server-586c8bb899-r6zbz" Nov 21 14:14:05 crc kubenswrapper[4897]: I1121 14:14:05.602842 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/monitoring-plugin-7869668964-7jkch"] Nov 21 14:14:05 crc kubenswrapper[4897]: I1121 14:14:05.603856 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/monitoring-plugin-7869668964-7jkch" Nov 21 14:14:05 crc kubenswrapper[4897]: I1121 14:14:05.609930 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"monitoring-plugin-cert" Nov 21 14:14:05 crc kubenswrapper[4897]: I1121 14:14:05.610406 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"default-dockercfg-6tstp" Nov 21 14:14:05 crc kubenswrapper[4897]: I1121 14:14:05.614310 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/monitoring-plugin-7869668964-7jkch"] Nov 21 14:14:05 crc kubenswrapper[4897]: I1121 14:14:05.725989 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"monitoring-plugin-cert\" (UniqueName: \"kubernetes.io/secret/ee8ae79c-5683-46c4-909d-d9338c398678-monitoring-plugin-cert\") pod \"monitoring-plugin-7869668964-7jkch\" (UID: \"ee8ae79c-5683-46c4-909d-d9338c398678\") " pod="openshift-monitoring/monitoring-plugin-7869668964-7jkch" Nov 21 14:14:05 crc kubenswrapper[4897]: I1121 14:14:05.827572 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"monitoring-plugin-cert\" (UniqueName: \"kubernetes.io/secret/ee8ae79c-5683-46c4-909d-d9338c398678-monitoring-plugin-cert\") pod \"monitoring-plugin-7869668964-7jkch\" (UID: \"ee8ae79c-5683-46c4-909d-d9338c398678\") " pod="openshift-monitoring/monitoring-plugin-7869668964-7jkch" Nov 21 14:14:05 crc kubenswrapper[4897]: I1121 14:14:05.832563 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"monitoring-plugin-cert\" (UniqueName: \"kubernetes.io/secret/ee8ae79c-5683-46c4-909d-d9338c398678-monitoring-plugin-cert\") pod \"monitoring-plugin-7869668964-7jkch\" (UID: \"ee8ae79c-5683-46c4-909d-d9338c398678\") " pod="openshift-monitoring/monitoring-plugin-7869668964-7jkch" Nov 21 14:14:05 crc kubenswrapper[4897]: I1121 14:14:05.878424 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/metrics-server-586c8bb899-r6zbz"] Nov 21 14:14:05 crc kubenswrapper[4897]: W1121 14:14:05.891269 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod97e1c0fe_f3e3_4afe_819e_eb85875bee86.slice/crio-597514f28d1bb8444704658baf374eb3f977e9c5ee23cddcbdcff3e4f51e512d WatchSource:0}: Error finding container 597514f28d1bb8444704658baf374eb3f977e9c5ee23cddcbdcff3e4f51e512d: Status 404 returned error can't find the container with id 597514f28d1bb8444704658baf374eb3f977e9c5ee23cddcbdcff3e4f51e512d Nov 21 14:14:05 crc kubenswrapper[4897]: I1121 14:14:05.927554 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/monitoring-plugin-7869668964-7jkch" Nov 21 14:14:06 crc kubenswrapper[4897]: I1121 14:14:06.079125 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/prometheus-k8s-0"] Nov 21 14:14:06 crc kubenswrapper[4897]: I1121 14:14:06.080958 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-k8s-0" Nov 21 14:14:06 crc kubenswrapper[4897]: I1121 14:14:06.089246 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-thanos-prometheus-http-client-file" Nov 21 14:14:06 crc kubenswrapper[4897]: I1121 14:14:06.089414 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"kube-rbac-proxy" Nov 21 14:14:06 crc kubenswrapper[4897]: I1121 14:14:06.089462 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-web-config" Nov 21 14:14:06 crc kubenswrapper[4897]: I1121 14:14:06.090429 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-kube-rbac-proxy-web" Nov 21 14:14:06 crc kubenswrapper[4897]: I1121 14:14:06.090561 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"serving-certs-ca-bundle" Nov 21 14:14:06 crc kubenswrapper[4897]: I1121 14:14:06.090644 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-grpc-tls-ced1cvbut8157" Nov 21 14:14:06 crc kubenswrapper[4897]: I1121 14:14:06.090729 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-tls-assets-0" Nov 21 14:14:06 crc kubenswrapper[4897]: I1121 14:14:06.090769 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-thanos-sidecar-tls" Nov 21 14:14:06 crc kubenswrapper[4897]: I1121 14:14:06.090811 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-tls" Nov 21 14:14:06 crc kubenswrapper[4897]: I1121 14:14:06.090726 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s" Nov 21 14:14:06 crc kubenswrapper[4897]: I1121 14:14:06.091734 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-dockercfg-xfxtt" Nov 21 14:14:06 crc kubenswrapper[4897]: I1121 14:14:06.093064 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"prometheus-k8s-rulefiles-0" Nov 21 14:14:06 crc kubenswrapper[4897]: I1121 14:14:06.101776 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"prometheus-trusted-ca-bundle" Nov 21 14:14:06 crc kubenswrapper[4897]: I1121 14:14:06.109096 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-k8s-0"] Nov 21 14:14:06 crc kubenswrapper[4897]: I1121 14:14:06.238728 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/2ac1afd5-988d-447d-b7f1-07476917ac23-secret-metrics-client-certs\") pod \"prometheus-k8s-0\" (UID: \"2ac1afd5-988d-447d-b7f1-07476917ac23\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 21 14:14:06 crc kubenswrapper[4897]: I1121 14:14:06.238776 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-k8s-db\" (UniqueName: \"kubernetes.io/empty-dir/2ac1afd5-988d-447d-b7f1-07476917ac23-prometheus-k8s-db\") pod \"prometheus-k8s-0\" (UID: \"2ac1afd5-988d-447d-b7f1-07476917ac23\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 21 14:14:06 crc kubenswrapper[4897]: I1121 14:14:06.238800 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tkgj5\" (UniqueName: \"kubernetes.io/projected/2ac1afd5-988d-447d-b7f1-07476917ac23-kube-api-access-tkgj5\") pod \"prometheus-k8s-0\" (UID: \"2ac1afd5-988d-447d-b7f1-07476917ac23\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 21 14:14:06 crc kubenswrapper[4897]: I1121 14:14:06.238867 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2ac1afd5-988d-447d-b7f1-07476917ac23-configmap-kubelet-serving-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"2ac1afd5-988d-447d-b7f1-07476917ac23\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 21 14:14:06 crc kubenswrapper[4897]: I1121 14:14:06.238887 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-k8s-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/2ac1afd5-988d-447d-b7f1-07476917ac23-prometheus-k8s-rulefiles-0\") pod \"prometheus-k8s-0\" (UID: \"2ac1afd5-988d-447d-b7f1-07476917ac23\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 21 14:14:06 crc kubenswrapper[4897]: I1121 14:14:06.238905 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2ac1afd5-988d-447d-b7f1-07476917ac23-prometheus-trusted-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"2ac1afd5-988d-447d-b7f1-07476917ac23\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 21 14:14:06 crc kubenswrapper[4897]: I1121 14:14:06.239055 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/2ac1afd5-988d-447d-b7f1-07476917ac23-configmap-metrics-client-ca\") pod \"prometheus-k8s-0\" (UID: \"2ac1afd5-988d-447d-b7f1-07476917ac23\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 21 14:14:06 crc kubenswrapper[4897]: I1121 14:14:06.239109 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/2ac1afd5-988d-447d-b7f1-07476917ac23-web-config\") pod \"prometheus-k8s-0\" (UID: \"2ac1afd5-988d-447d-b7f1-07476917ac23\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 21 14:14:06 crc kubenswrapper[4897]: I1121 14:14:06.239153 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-prometheus-k8s-thanos-sidecar-tls\" (UniqueName: \"kubernetes.io/secret/2ac1afd5-988d-447d-b7f1-07476917ac23-secret-prometheus-k8s-thanos-sidecar-tls\") pod \"prometheus-k8s-0\" (UID: \"2ac1afd5-988d-447d-b7f1-07476917ac23\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 21 14:14:06 crc kubenswrapper[4897]: I1121 14:14:06.239170 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/2ac1afd5-988d-447d-b7f1-07476917ac23-thanos-prometheus-http-client-file\") pod \"prometheus-k8s-0\" (UID: \"2ac1afd5-988d-447d-b7f1-07476917ac23\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 21 14:14:06 crc kubenswrapper[4897]: I1121 14:14:06.239190 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/2ac1afd5-988d-447d-b7f1-07476917ac23-tls-assets\") pod \"prometheus-k8s-0\" (UID: \"2ac1afd5-988d-447d-b7f1-07476917ac23\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 21 14:14:06 crc kubenswrapper[4897]: I1121 14:14:06.239229 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-prometheus-k8s-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/2ac1afd5-988d-447d-b7f1-07476917ac23-secret-prometheus-k8s-kube-rbac-proxy-web\") pod \"prometheus-k8s-0\" (UID: \"2ac1afd5-988d-447d-b7f1-07476917ac23\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 21 14:14:06 crc kubenswrapper[4897]: I1121 14:14:06.239285 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/2ac1afd5-988d-447d-b7f1-07476917ac23-config-out\") pod \"prometheus-k8s-0\" (UID: \"2ac1afd5-988d-447d-b7f1-07476917ac23\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 21 14:14:06 crc kubenswrapper[4897]: I1121 14:14:06.239301 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/2ac1afd5-988d-447d-b7f1-07476917ac23-secret-grpc-tls\") pod \"prometheus-k8s-0\" (UID: \"2ac1afd5-988d-447d-b7f1-07476917ac23\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 21 14:14:06 crc kubenswrapper[4897]: I1121 14:14:06.239322 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/2ac1afd5-988d-447d-b7f1-07476917ac23-secret-kube-rbac-proxy\") pod \"prometheus-k8s-0\" (UID: \"2ac1afd5-988d-447d-b7f1-07476917ac23\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 21 14:14:06 crc kubenswrapper[4897]: I1121 14:14:06.239400 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2ac1afd5-988d-447d-b7f1-07476917ac23-configmap-serving-certs-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"2ac1afd5-988d-447d-b7f1-07476917ac23\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 21 14:14:06 crc kubenswrapper[4897]: I1121 14:14:06.239435 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/2ac1afd5-988d-447d-b7f1-07476917ac23-config\") pod \"prometheus-k8s-0\" (UID: \"2ac1afd5-988d-447d-b7f1-07476917ac23\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 21 14:14:06 crc kubenswrapper[4897]: I1121 14:14:06.239461 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-prometheus-k8s-tls\" (UniqueName: \"kubernetes.io/secret/2ac1afd5-988d-447d-b7f1-07476917ac23-secret-prometheus-k8s-tls\") pod \"prometheus-k8s-0\" (UID: \"2ac1afd5-988d-447d-b7f1-07476917ac23\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 21 14:14:06 crc kubenswrapper[4897]: I1121 14:14:06.340412 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-prometheus-k8s-tls\" (UniqueName: \"kubernetes.io/secret/2ac1afd5-988d-447d-b7f1-07476917ac23-secret-prometheus-k8s-tls\") pod \"prometheus-k8s-0\" (UID: \"2ac1afd5-988d-447d-b7f1-07476917ac23\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 21 14:14:06 crc kubenswrapper[4897]: I1121 14:14:06.340464 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-k8s-db\" (UniqueName: \"kubernetes.io/empty-dir/2ac1afd5-988d-447d-b7f1-07476917ac23-prometheus-k8s-db\") pod \"prometheus-k8s-0\" (UID: \"2ac1afd5-988d-447d-b7f1-07476917ac23\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 21 14:14:06 crc kubenswrapper[4897]: I1121 14:14:06.340479 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/2ac1afd5-988d-447d-b7f1-07476917ac23-secret-metrics-client-certs\") pod \"prometheus-k8s-0\" (UID: \"2ac1afd5-988d-447d-b7f1-07476917ac23\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 21 14:14:06 crc kubenswrapper[4897]: I1121 14:14:06.340496 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tkgj5\" (UniqueName: \"kubernetes.io/projected/2ac1afd5-988d-447d-b7f1-07476917ac23-kube-api-access-tkgj5\") pod \"prometheus-k8s-0\" (UID: \"2ac1afd5-988d-447d-b7f1-07476917ac23\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 21 14:14:06 crc kubenswrapper[4897]: I1121 14:14:06.340536 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2ac1afd5-988d-447d-b7f1-07476917ac23-configmap-kubelet-serving-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"2ac1afd5-988d-447d-b7f1-07476917ac23\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 21 14:14:06 crc kubenswrapper[4897]: I1121 14:14:06.340555 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-k8s-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/2ac1afd5-988d-447d-b7f1-07476917ac23-prometheus-k8s-rulefiles-0\") pod \"prometheus-k8s-0\" (UID: \"2ac1afd5-988d-447d-b7f1-07476917ac23\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 21 14:14:06 crc kubenswrapper[4897]: I1121 14:14:06.340572 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2ac1afd5-988d-447d-b7f1-07476917ac23-prometheus-trusted-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"2ac1afd5-988d-447d-b7f1-07476917ac23\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 21 14:14:06 crc kubenswrapper[4897]: I1121 14:14:06.340601 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"configmap-metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/2ac1afd5-988d-447d-b7f1-07476917ac23-configmap-metrics-client-ca\") pod \"prometheus-k8s-0\" (UID: \"2ac1afd5-988d-447d-b7f1-07476917ac23\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 21 14:14:06 crc kubenswrapper[4897]: I1121 14:14:06.340621 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/2ac1afd5-988d-447d-b7f1-07476917ac23-web-config\") pod \"prometheus-k8s-0\" (UID: \"2ac1afd5-988d-447d-b7f1-07476917ac23\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 21 14:14:06 crc kubenswrapper[4897]: I1121 14:14:06.340641 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-prometheus-k8s-thanos-sidecar-tls\" (UniqueName: \"kubernetes.io/secret/2ac1afd5-988d-447d-b7f1-07476917ac23-secret-prometheus-k8s-thanos-sidecar-tls\") pod \"prometheus-k8s-0\" (UID: \"2ac1afd5-988d-447d-b7f1-07476917ac23\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 21 14:14:06 crc kubenswrapper[4897]: I1121 14:14:06.340656 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/2ac1afd5-988d-447d-b7f1-07476917ac23-thanos-prometheus-http-client-file\") pod \"prometheus-k8s-0\" (UID: \"2ac1afd5-988d-447d-b7f1-07476917ac23\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 21 14:14:06 crc kubenswrapper[4897]: I1121 14:14:06.340670 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/2ac1afd5-988d-447d-b7f1-07476917ac23-tls-assets\") pod \"prometheus-k8s-0\" (UID: \"2ac1afd5-988d-447d-b7f1-07476917ac23\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 21 14:14:06 crc kubenswrapper[4897]: I1121 14:14:06.340686 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-prometheus-k8s-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/2ac1afd5-988d-447d-b7f1-07476917ac23-secret-prometheus-k8s-kube-rbac-proxy-web\") pod \"prometheus-k8s-0\" (UID: \"2ac1afd5-988d-447d-b7f1-07476917ac23\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 21 14:14:06 crc kubenswrapper[4897]: I1121 14:14:06.340709 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/2ac1afd5-988d-447d-b7f1-07476917ac23-config-out\") pod \"prometheus-k8s-0\" (UID: \"2ac1afd5-988d-447d-b7f1-07476917ac23\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 21 14:14:06 crc kubenswrapper[4897]: I1121 14:14:06.340726 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/2ac1afd5-988d-447d-b7f1-07476917ac23-secret-grpc-tls\") pod \"prometheus-k8s-0\" (UID: \"2ac1afd5-988d-447d-b7f1-07476917ac23\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 21 14:14:06 crc kubenswrapper[4897]: I1121 14:14:06.340742 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/2ac1afd5-988d-447d-b7f1-07476917ac23-secret-kube-rbac-proxy\") pod \"prometheus-k8s-0\" (UID: \"2ac1afd5-988d-447d-b7f1-07476917ac23\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 21 14:14:06 crc kubenswrapper[4897]: I1121 14:14:06.340774 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2ac1afd5-988d-447d-b7f1-07476917ac23-configmap-serving-certs-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"2ac1afd5-988d-447d-b7f1-07476917ac23\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 21 14:14:06 crc kubenswrapper[4897]: I1121 14:14:06.340795 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/2ac1afd5-988d-447d-b7f1-07476917ac23-config\") pod \"prometheus-k8s-0\" (UID: \"2ac1afd5-988d-447d-b7f1-07476917ac23\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 21 14:14:06 crc kubenswrapper[4897]: I1121 14:14:06.343757 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/2ac1afd5-988d-447d-b7f1-07476917ac23-config\") pod \"prometheus-k8s-0\" (UID: \"2ac1afd5-988d-447d-b7f1-07476917ac23\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 21 14:14:06 crc kubenswrapper[4897]: I1121 14:14:06.343877 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/2ac1afd5-988d-447d-b7f1-07476917ac23-web-config\") pod \"prometheus-k8s-0\" (UID: \"2ac1afd5-988d-447d-b7f1-07476917ac23\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 21 14:14:06 crc kubenswrapper[4897]: I1121 14:14:06.344653 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-prometheus-k8s-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/2ac1afd5-988d-447d-b7f1-07476917ac23-secret-prometheus-k8s-kube-rbac-proxy-web\") pod \"prometheus-k8s-0\" (UID: \"2ac1afd5-988d-447d-b7f1-07476917ac23\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 21 14:14:06 crc kubenswrapper[4897]: I1121 14:14:06.345139 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-k8s-db\" (UniqueName: \"kubernetes.io/empty-dir/2ac1afd5-988d-447d-b7f1-07476917ac23-prometheus-k8s-db\") pod \"prometheus-k8s-0\" (UID: \"2ac1afd5-988d-447d-b7f1-07476917ac23\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 21 14:14:06 crc kubenswrapper[4897]: I1121 14:14:06.345225 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/2ac1afd5-988d-447d-b7f1-07476917ac23-tls-assets\") pod \"prometheus-k8s-0\" (UID: \"2ac1afd5-988d-447d-b7f1-07476917ac23\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 21 14:14:06 crc kubenswrapper[4897]: I1121 14:14:06.345740 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2ac1afd5-988d-447d-b7f1-07476917ac23-configmap-kubelet-serving-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"2ac1afd5-988d-447d-b7f1-07476917ac23\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 21 14:14:06 crc kubenswrapper[4897]: I1121 14:14:06.346132 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-prometheus-k8s-thanos-sidecar-tls\" (UniqueName: \"kubernetes.io/secret/2ac1afd5-988d-447d-b7f1-07476917ac23-secret-prometheus-k8s-thanos-sidecar-tls\") pod \"prometheus-k8s-0\" (UID: \"2ac1afd5-988d-447d-b7f1-07476917ac23\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 21 14:14:06 crc kubenswrapper[4897]: I1121 14:14:06.347652 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2ac1afd5-988d-447d-b7f1-07476917ac23-configmap-serving-certs-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"2ac1afd5-988d-447d-b7f1-07476917ac23\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 21 14:14:06 crc kubenswrapper[4897]: I1121 14:14:06.347725 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/2ac1afd5-988d-447d-b7f1-07476917ac23-secret-grpc-tls\") pod \"prometheus-k8s-0\" (UID: \"2ac1afd5-988d-447d-b7f1-07476917ac23\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 21 14:14:06 crc kubenswrapper[4897]: I1121 14:14:06.347932 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"configmap-metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/2ac1afd5-988d-447d-b7f1-07476917ac23-configmap-metrics-client-ca\") pod \"prometheus-k8s-0\" (UID: \"2ac1afd5-988d-447d-b7f1-07476917ac23\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 21 14:14:06 crc kubenswrapper[4897]: I1121 14:14:06.348467 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/2ac1afd5-988d-447d-b7f1-07476917ac23-thanos-prometheus-http-client-file\") pod \"prometheus-k8s-0\" (UID: \"2ac1afd5-988d-447d-b7f1-07476917ac23\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 21 14:14:06 crc kubenswrapper[4897]: I1121 14:14:06.349117 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/2ac1afd5-988d-447d-b7f1-07476917ac23-secret-kube-rbac-proxy\") pod \"prometheus-k8s-0\" (UID: \"2ac1afd5-988d-447d-b7f1-07476917ac23\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 21 14:14:06 crc kubenswrapper[4897]: I1121 14:14:06.349208 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2ac1afd5-988d-447d-b7f1-07476917ac23-prometheus-trusted-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"2ac1afd5-988d-447d-b7f1-07476917ac23\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 21 14:14:06 crc kubenswrapper[4897]: I1121 14:14:06.349231 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-prometheus-k8s-tls\" (UniqueName: \"kubernetes.io/secret/2ac1afd5-988d-447d-b7f1-07476917ac23-secret-prometheus-k8s-tls\") pod \"prometheus-k8s-0\" (UID: \"2ac1afd5-988d-447d-b7f1-07476917ac23\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 21 14:14:06 crc kubenswrapper[4897]: I1121 14:14:06.350515 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/2ac1afd5-988d-447d-b7f1-07476917ac23-config-out\") pod \"prometheus-k8s-0\" (UID: \"2ac1afd5-988d-447d-b7f1-07476917ac23\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 21 14:14:06 crc kubenswrapper[4897]: I1121 14:14:06.352231 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/2ac1afd5-988d-447d-b7f1-07476917ac23-secret-metrics-client-certs\") pod \"prometheus-k8s-0\" (UID: \"2ac1afd5-988d-447d-b7f1-07476917ac23\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 21 14:14:06 crc kubenswrapper[4897]: I1121 14:14:06.356596 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-k8s-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/2ac1afd5-988d-447d-b7f1-07476917ac23-prometheus-k8s-rulefiles-0\") pod \"prometheus-k8s-0\" (UID: \"2ac1afd5-988d-447d-b7f1-07476917ac23\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 21 14:14:06 crc kubenswrapper[4897]: I1121 14:14:06.364097 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tkgj5\" (UniqueName: \"kubernetes.io/projected/2ac1afd5-988d-447d-b7f1-07476917ac23-kube-api-access-tkgj5\") pod \"prometheus-k8s-0\" (UID: \"2ac1afd5-988d-447d-b7f1-07476917ac23\") " pod="openshift-monitoring/prometheus-k8s-0" Nov 21 14:14:06 crc kubenswrapper[4897]: I1121 14:14:06.372357 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/monitoring-plugin-7869668964-7jkch"] Nov 21 14:14:06 crc kubenswrapper[4897]: I1121 14:14:06.404814 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-k8s-0" Nov 21 14:14:06 crc kubenswrapper[4897]: I1121 14:14:06.810283 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/metrics-server-586c8bb899-r6zbz" event={"ID":"97e1c0fe-f3e3-4afe-819e-eb85875bee86","Type":"ContainerStarted","Data":"597514f28d1bb8444704658baf374eb3f977e9c5ee23cddcbdcff3e4f51e512d"} Nov 21 14:14:06 crc kubenswrapper[4897]: I1121 14:14:06.811444 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/monitoring-plugin-7869668964-7jkch" event={"ID":"ee8ae79c-5683-46c4-909d-d9338c398678","Type":"ContainerStarted","Data":"183ba86e3b22ab83d57914222841f17dc733f2dc588c8a9046812624adcc3faf"} Nov 21 14:14:06 crc kubenswrapper[4897]: I1121 14:14:06.838381 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-k8s-0"] Nov 21 14:14:06 crc kubenswrapper[4897]: W1121 14:14:06.848219 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2ac1afd5_988d_447d_b7f1_07476917ac23.slice/crio-e726c620011ed459c104de52f2b2846a70cb6d4a66f697f2bb5baf5c2d927918 WatchSource:0}: Error finding container e726c620011ed459c104de52f2b2846a70cb6d4a66f697f2bb5baf5c2d927918: Status 404 returned error can't find the container with id e726c620011ed459c104de52f2b2846a70cb6d4a66f697f2bb5baf5c2d927918 Nov 21 14:14:07 crc kubenswrapper[4897]: I1121 14:14:07.819700 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"2ac1afd5-988d-447d-b7f1-07476917ac23","Type":"ContainerStarted","Data":"e726c620011ed459c104de52f2b2846a70cb6d4a66f697f2bb5baf5c2d927918"} Nov 21 14:14:10 crc kubenswrapper[4897]: I1121 14:14:10.549982 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-lsctq" Nov 21 14:14:10 crc kubenswrapper[4897]: I1121 14:14:10.603390 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-5z687"] Nov 21 14:14:12 crc kubenswrapper[4897]: I1121 14:14:12.031758 4897 patch_prober.go:28] interesting pod/router-default-5444994796-2kgg2 container/router namespace/openshift-ingress: Readiness probe status=failure output="Get \"http://localhost:1936/healthz/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 21 14:14:12 crc kubenswrapper[4897]: I1121 14:14:12.033017 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-ingress/router-default-5444994796-2kgg2" podUID="e095a18e-3a0b-4418-95a8-a2cc3de5bd7b" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 21 14:14:17 crc kubenswrapper[4897]: I1121 14:14:17.584987 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-5f888bf76d-vtk2t"] Nov 21 14:14:17 crc kubenswrapper[4897]: I1121 14:14:17.586283 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5f888bf76d-vtk2t" Nov 21 14:14:17 crc kubenswrapper[4897]: I1121 14:14:17.598353 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-5f888bf76d-vtk2t"] Nov 21 14:14:17 crc kubenswrapper[4897]: I1121 14:14:17.727178 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/062ef123-84be-49c9-9bb2-8fc1a415f431-console-serving-cert\") pod \"console-5f888bf76d-vtk2t\" (UID: \"062ef123-84be-49c9-9bb2-8fc1a415f431\") " pod="openshift-console/console-5f888bf76d-vtk2t" Nov 21 14:14:17 crc kubenswrapper[4897]: I1121 14:14:17.727346 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/062ef123-84be-49c9-9bb2-8fc1a415f431-service-ca\") pod \"console-5f888bf76d-vtk2t\" (UID: \"062ef123-84be-49c9-9bb2-8fc1a415f431\") " pod="openshift-console/console-5f888bf76d-vtk2t" Nov 21 14:14:17 crc kubenswrapper[4897]: I1121 14:14:17.727467 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/062ef123-84be-49c9-9bb2-8fc1a415f431-oauth-serving-cert\") pod \"console-5f888bf76d-vtk2t\" (UID: \"062ef123-84be-49c9-9bb2-8fc1a415f431\") " pod="openshift-console/console-5f888bf76d-vtk2t" Nov 21 14:14:17 crc kubenswrapper[4897]: I1121 14:14:17.727565 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/062ef123-84be-49c9-9bb2-8fc1a415f431-console-oauth-config\") pod \"console-5f888bf76d-vtk2t\" (UID: \"062ef123-84be-49c9-9bb2-8fc1a415f431\") " pod="openshift-console/console-5f888bf76d-vtk2t" Nov 21 14:14:17 crc kubenswrapper[4897]: I1121 14:14:17.727609 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/062ef123-84be-49c9-9bb2-8fc1a415f431-trusted-ca-bundle\") pod \"console-5f888bf76d-vtk2t\" (UID: \"062ef123-84be-49c9-9bb2-8fc1a415f431\") " pod="openshift-console/console-5f888bf76d-vtk2t" Nov 21 14:14:17 crc kubenswrapper[4897]: I1121 14:14:17.727792 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/062ef123-84be-49c9-9bb2-8fc1a415f431-console-config\") pod \"console-5f888bf76d-vtk2t\" (UID: \"062ef123-84be-49c9-9bb2-8fc1a415f431\") " pod="openshift-console/console-5f888bf76d-vtk2t" Nov 21 14:14:17 crc kubenswrapper[4897]: I1121 14:14:17.727944 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dwqrt\" (UniqueName: \"kubernetes.io/projected/062ef123-84be-49c9-9bb2-8fc1a415f431-kube-api-access-dwqrt\") pod \"console-5f888bf76d-vtk2t\" (UID: \"062ef123-84be-49c9-9bb2-8fc1a415f431\") " pod="openshift-console/console-5f888bf76d-vtk2t" Nov 21 14:14:17 crc kubenswrapper[4897]: I1121 14:14:17.829754 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/062ef123-84be-49c9-9bb2-8fc1a415f431-console-serving-cert\") pod \"console-5f888bf76d-vtk2t\" (UID: \"062ef123-84be-49c9-9bb2-8fc1a415f431\") " pod="openshift-console/console-5f888bf76d-vtk2t" Nov 21 14:14:17 crc kubenswrapper[4897]: I1121 14:14:17.829834 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/062ef123-84be-49c9-9bb2-8fc1a415f431-service-ca\") pod \"console-5f888bf76d-vtk2t\" (UID: \"062ef123-84be-49c9-9bb2-8fc1a415f431\") " pod="openshift-console/console-5f888bf76d-vtk2t" Nov 21 14:14:17 crc kubenswrapper[4897]: I1121 14:14:17.829879 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/062ef123-84be-49c9-9bb2-8fc1a415f431-oauth-serving-cert\") pod \"console-5f888bf76d-vtk2t\" (UID: \"062ef123-84be-49c9-9bb2-8fc1a415f431\") " pod="openshift-console/console-5f888bf76d-vtk2t" Nov 21 14:14:17 crc kubenswrapper[4897]: I1121 14:14:17.829980 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/062ef123-84be-49c9-9bb2-8fc1a415f431-console-oauth-config\") pod \"console-5f888bf76d-vtk2t\" (UID: \"062ef123-84be-49c9-9bb2-8fc1a415f431\") " pod="openshift-console/console-5f888bf76d-vtk2t" Nov 21 14:14:17 crc kubenswrapper[4897]: I1121 14:14:17.830008 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/062ef123-84be-49c9-9bb2-8fc1a415f431-trusted-ca-bundle\") pod \"console-5f888bf76d-vtk2t\" (UID: \"062ef123-84be-49c9-9bb2-8fc1a415f431\") " pod="openshift-console/console-5f888bf76d-vtk2t" Nov 21 14:14:17 crc kubenswrapper[4897]: I1121 14:14:17.830048 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/062ef123-84be-49c9-9bb2-8fc1a415f431-console-config\") pod \"console-5f888bf76d-vtk2t\" (UID: \"062ef123-84be-49c9-9bb2-8fc1a415f431\") " pod="openshift-console/console-5f888bf76d-vtk2t" Nov 21 14:14:17 crc kubenswrapper[4897]: I1121 14:14:17.830942 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/062ef123-84be-49c9-9bb2-8fc1a415f431-service-ca\") pod \"console-5f888bf76d-vtk2t\" (UID: \"062ef123-84be-49c9-9bb2-8fc1a415f431\") " pod="openshift-console/console-5f888bf76d-vtk2t" Nov 21 14:14:17 crc kubenswrapper[4897]: I1121 14:14:17.830962 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/062ef123-84be-49c9-9bb2-8fc1a415f431-oauth-serving-cert\") pod \"console-5f888bf76d-vtk2t\" (UID: \"062ef123-84be-49c9-9bb2-8fc1a415f431\") " pod="openshift-console/console-5f888bf76d-vtk2t" Nov 21 14:14:17 crc kubenswrapper[4897]: I1121 14:14:17.831228 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/062ef123-84be-49c9-9bb2-8fc1a415f431-console-config\") pod \"console-5f888bf76d-vtk2t\" (UID: \"062ef123-84be-49c9-9bb2-8fc1a415f431\") " pod="openshift-console/console-5f888bf76d-vtk2t" Nov 21 14:14:17 crc kubenswrapper[4897]: I1121 14:14:17.831571 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/062ef123-84be-49c9-9bb2-8fc1a415f431-trusted-ca-bundle\") pod \"console-5f888bf76d-vtk2t\" (UID: \"062ef123-84be-49c9-9bb2-8fc1a415f431\") " pod="openshift-console/console-5f888bf76d-vtk2t" Nov 21 14:14:17 crc kubenswrapper[4897]: I1121 14:14:17.831655 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dwqrt\" (UniqueName: \"kubernetes.io/projected/062ef123-84be-49c9-9bb2-8fc1a415f431-kube-api-access-dwqrt\") pod \"console-5f888bf76d-vtk2t\" (UID: \"062ef123-84be-49c9-9bb2-8fc1a415f431\") " pod="openshift-console/console-5f888bf76d-vtk2t" Nov 21 14:14:17 crc kubenswrapper[4897]: I1121 14:14:17.838217 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/062ef123-84be-49c9-9bb2-8fc1a415f431-console-oauth-config\") pod \"console-5f888bf76d-vtk2t\" (UID: \"062ef123-84be-49c9-9bb2-8fc1a415f431\") " pod="openshift-console/console-5f888bf76d-vtk2t" Nov 21 14:14:17 crc kubenswrapper[4897]: I1121 14:14:17.841313 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/062ef123-84be-49c9-9bb2-8fc1a415f431-console-serving-cert\") pod \"console-5f888bf76d-vtk2t\" (UID: \"062ef123-84be-49c9-9bb2-8fc1a415f431\") " pod="openshift-console/console-5f888bf76d-vtk2t" Nov 21 14:14:17 crc kubenswrapper[4897]: I1121 14:14:17.851696 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dwqrt\" (UniqueName: \"kubernetes.io/projected/062ef123-84be-49c9-9bb2-8fc1a415f431-kube-api-access-dwqrt\") pod \"console-5f888bf76d-vtk2t\" (UID: \"062ef123-84be-49c9-9bb2-8fc1a415f431\") " pod="openshift-console/console-5f888bf76d-vtk2t" Nov 21 14:14:17 crc kubenswrapper[4897]: I1121 14:14:17.907718 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5f888bf76d-vtk2t" Nov 21 14:14:18 crc kubenswrapper[4897]: E1121 14:14:18.496461 4897 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = writing blob: storing blob to file \"/var/tmp/container_images_storage1824190278/1\": happened during read: context canceled" image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e1456825b53fc7e6ea6aa2003b3f3626ad7846802f9fd9dc69874e349b849ad" Nov 21 14:14:18 crc kubenswrapper[4897]: E1121 14:14:18.496990 4897 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init-textfile,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e1456825b53fc7e6ea6aa2003b3f3626ad7846802f9fd9dc69874e349b849ad,Command:[/bin/sh -c [[ ! -d /node_exporter/collectors/init ]] || find /node_exporter/collectors/init -perm /111 -type f -exec {} \\;],Args:[],WorkingDir:/var/node_exporter/textfile,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:TMPDIR,Value:/tmp,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{1 -3} {} 1m DecimalSI},memory: {{1048576 0} {} 1Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:node-exporter-textfile,ReadOnly:false,MountPath:/var/node_exporter/textfile,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:node-exporter-wtmp,ReadOnly:true,MountPath:/var/log/wtmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-28fnm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod node-exporter-9gpbl_openshift-monitoring(8caaec7a-8eab-4ffe-9e68-f5c763d044eb): ErrImagePull: rpc error: code = Canceled desc = writing blob: storing blob to file \"/var/tmp/container_images_storage1824190278/1\": happened during read: context canceled" logger="UnhandledError" Nov 21 14:14:18 crc kubenswrapper[4897]: E1121 14:14:18.498165 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init-textfile\" with ErrImagePull: \"rpc error: code = Canceled desc = writing blob: storing blob to file \\\"/var/tmp/container_images_storage1824190278/1\\\": happened during read: context canceled\"" pod="openshift-monitoring/node-exporter-9gpbl" podUID="8caaec7a-8eab-4ffe-9e68-f5c763d044eb" Nov 21 14:14:18 crc kubenswrapper[4897]: E1121 14:14:18.819530 4897 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47dcd507a8ad265c7ebd6b128bb9bdaeb7688b5731503817b94ae1d1badd6a77" Nov 21 14:14:18 crc kubenswrapper[4897]: E1121 14:14:18.819696 4897 kuberuntime_manager.go:1274] "Unhandled Error" err=< Nov 21 14:14:18 crc kubenswrapper[4897]: container &Container{Name:kube-state-metrics,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47dcd507a8ad265c7ebd6b128bb9bdaeb7688b5731503817b94ae1d1badd6a77,Command:[],Args:[--host=127.0.0.1 --port=8081 --telemetry-host=127.0.0.1 --telemetry-port=8082 --metric-denylist= Nov 21 14:14:18 crc kubenswrapper[4897]: ^kube_secret_labels$, Nov 21 14:14:18 crc kubenswrapper[4897]: ^kube_.+_annotations$ Nov 21 14:14:18 crc kubenswrapper[4897]: ^kube_customresource_.+_annotations_info$, Nov 21 14:14:18 crc kubenswrapper[4897]: ^kube_customresource_.+_labels_info$, Nov 21 14:14:18 crc kubenswrapper[4897]: --metric-labels-allowlist=pods=[*],nodes=[*],namespaces=[*],persistentvolumes=[*],persistentvolumeclaims=[*],poddisruptionbudgets=[*] --metric-denylist= Nov 21 14:14:18 crc kubenswrapper[4897]: ^kube_.+_created$, Nov 21 14:14:18 crc kubenswrapper[4897]: ^kube_.+_metadata_resource_version$, Nov 21 14:14:18 crc kubenswrapper[4897]: ^kube_replicaset_metadata_generation$, Nov 21 14:14:18 crc kubenswrapper[4897]: ^kube_replicaset_status_observed_generation$, Nov 21 14:14:18 crc kubenswrapper[4897]: ^kube_pod_restart_policy$, Nov 21 14:14:18 crc kubenswrapper[4897]: ^kube_pod_init_container_status_terminated$, Nov 21 14:14:18 crc kubenswrapper[4897]: ^kube_pod_init_container_status_running$, Nov 21 14:14:18 crc kubenswrapper[4897]: ^kube_pod_container_status_terminated$, Nov 21 14:14:18 crc kubenswrapper[4897]: ^kube_pod_container_status_running$, Nov 21 14:14:18 crc kubenswrapper[4897]: ^kube_pod_completion_time$, Nov 21 14:14:18 crc kubenswrapper[4897]: ^kube_pod_status_scheduled$ Nov 21 14:14:18 crc kubenswrapper[4897]: ],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{2 -3} {} 2m DecimalSI},memory: {{83886080 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:volume-directive-shadow,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-state-metrics-custom-resource-state-configmap,ReadOnly:true,MountPath:/etc/kube-state-metrics,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-f76gb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000390000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod kube-state-metrics-777cb5bd5d-tm69l_openshift-monitoring(91c21c9f-aed2-41c3-b3fa-371f9903350a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled Nov 21 14:14:18 crc kubenswrapper[4897]: > logger="UnhandledError" Nov 21 14:14:18 crc kubenswrapper[4897]: I1121 14:14:18.905892 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/openshift-state-metrics-566fddb674-hw2hf" event={"ID":"8d0ea2ea-c9a1-495a-a626-7a1a735308df","Type":"ContainerStarted","Data":"43829a2121c850ac731c91985f4569cec89360473a803a1d176366cf900fafab"} Nov 21 14:14:18 crc kubenswrapper[4897]: E1121 14:14:18.908299 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init-textfile\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e1456825b53fc7e6ea6aa2003b3f3626ad7846802f9fd9dc69874e349b849ad\\\"\"" pod="openshift-monitoring/node-exporter-9gpbl" podUID="8caaec7a-8eab-4ffe-9e68-f5c763d044eb" Nov 21 14:14:19 crc kubenswrapper[4897]: E1121 14:14:19.288773 4897 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b51fe337495d490eda5164247b15073cb83da1f7bc629071c18ca9168b7f0461" Nov 21 14:14:19 crc kubenswrapper[4897]: E1121 14:14:19.288992 4897 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init-config-reloader,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b51fe337495d490eda5164247b15073cb83da1f7bc629071c18ca9168b7f0461,Command:[/bin/prometheus-config-reloader],Args:[--watch-interval=0 --listen-address=:8081 --config-file=/etc/alertmanager/config/alertmanager.yaml.gz --config-envsubst-file=/etc/alertmanager/config_out/alertmanager.env.yaml --watched-dir=/etc/alertmanager/config --watched-dir=/etc/alertmanager/secrets/alertmanager-main-tls --watched-dir=/etc/alertmanager/secrets/alertmanager-kube-rbac-proxy --watched-dir=/etc/alertmanager/secrets/alertmanager-kube-rbac-proxy-metric --watched-dir=/etc/alertmanager/secrets/alertmanager-kube-rbac-proxy-web],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:reloader-web,HostPort:0,ContainerPort:8081,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:SHARD,Value:-1,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{1 -3} {} 1m DecimalSI},memory: {{10485760 0} {} 10Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-volume,ReadOnly:true,MountPath:/etc/alertmanager/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-out,ReadOnly:false,MountPath:/etc/alertmanager/config_out,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:secret-alertmanager-main-tls,ReadOnly:true,MountPath:/etc/alertmanager/secrets/alertmanager-main-tls,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:secret-alertmanager-kube-rbac-proxy,ReadOnly:true,MountPath:/etc/alertmanager/secrets/alertmanager-kube-rbac-proxy,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:secret-alertmanager-kube-rbac-proxy-metric,ReadOnly:true,MountPath:/etc/alertmanager/secrets/alertmanager-kube-rbac-proxy-metric,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:secret-alertmanager-kube-rbac-proxy-web,ReadOnly:true,MountPath:/etc/alertmanager/secrets/alertmanager-kube-rbac-proxy-web,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:web-config,ReadOnly:true,MountPath:/etc/alertmanager/web_config/web-config.yaml,SubPath:web-config.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-sz6z9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL KILL MKNOD SETGID SETUID],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:*true,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod alertmanager-main-0_openshift-monitoring(d0fb95e9-9ac7-46bf-996a-0cef021092fe): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 21 14:14:19 crc kubenswrapper[4897]: E1121 14:14:19.290192 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init-config-reloader\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openshift-monitoring/alertmanager-main-0" podUID="d0fb95e9-9ac7-46bf-996a-0cef021092fe" Nov 21 14:14:19 crc kubenswrapper[4897]: I1121 14:14:19.686939 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-5f888bf76d-vtk2t"] Nov 21 14:14:19 crc kubenswrapper[4897]: I1121 14:14:19.920248 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5f888bf76d-vtk2t" event={"ID":"062ef123-84be-49c9-9bb2-8fc1a415f431","Type":"ContainerStarted","Data":"4b99cdf2d4a2a8555e1bf29923a5f96420c74764aff16d0be7205a25ad8329eb"} Nov 21 14:14:19 crc kubenswrapper[4897]: I1121 14:14:19.920615 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5f888bf76d-vtk2t" event={"ID":"062ef123-84be-49c9-9bb2-8fc1a415f431","Type":"ContainerStarted","Data":"17fda8582808ef239a7a28a430598982a5e5ace3de7b3fdd4b56a309d603211f"} Nov 21 14:14:19 crc kubenswrapper[4897]: I1121 14:14:19.924611 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/openshift-state-metrics-566fddb674-hw2hf" event={"ID":"8d0ea2ea-c9a1-495a-a626-7a1a735308df","Type":"ContainerStarted","Data":"dbbdf576d3a61a73b2db845de22beb5a54b4c7b54f232c5d022a31842b22d52f"} Nov 21 14:14:19 crc kubenswrapper[4897]: I1121 14:14:19.931924 4897 generic.go:334] "Generic (PLEG): container finished" podID="2ac1afd5-988d-447d-b7f1-07476917ac23" containerID="406f2f6cc0fa29506d8caa30932b44f13bf7c9bc90d5508c07883419ac40ab38" exitCode=0 Nov 21 14:14:19 crc kubenswrapper[4897]: I1121 14:14:19.931986 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"2ac1afd5-988d-447d-b7f1-07476917ac23","Type":"ContainerDied","Data":"406f2f6cc0fa29506d8caa30932b44f13bf7c9bc90d5508c07883419ac40ab38"} Nov 21 14:14:19 crc kubenswrapper[4897]: I1121 14:14:19.933247 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-tm69l" event={"ID":"91c21c9f-aed2-41c3-b3fa-371f9903350a","Type":"ContainerStarted","Data":"8e804cbb46797e2645eb8d919c88b0145e96a20c0fcd563c13bfddd359af7589"} Nov 21 14:14:19 crc kubenswrapper[4897]: I1121 14:14:19.951495 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-5f888bf76d-vtk2t" podStartSLOduration=2.951475024 podStartE2EDuration="2.951475024s" podCreationTimestamp="2025-11-21 14:14:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:14:19.9438524 +0000 UTC m=+337.228445895" watchObservedRunningTime="2025-11-21 14:14:19.951475024 +0000 UTC m=+337.236068499" Nov 21 14:14:20 crc kubenswrapper[4897]: E1121 14:14:20.232755 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-state-metrics\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-tm69l" podUID="91c21c9f-aed2-41c3-b3fa-371f9903350a" Nov 21 14:14:20 crc kubenswrapper[4897]: I1121 14:14:20.952799 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-tm69l" event={"ID":"91c21c9f-aed2-41c3-b3fa-371f9903350a","Type":"ContainerStarted","Data":"d76e8b24987440cb9ff4be6e20018081f851cc2640023e73355cd0524a98448d"} Nov 21 14:14:20 crc kubenswrapper[4897]: E1121 14:14:20.955096 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-state-metrics\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47dcd507a8ad265c7ebd6b128bb9bdaeb7688b5731503817b94ae1d1badd6a77\\\"\"" pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-tm69l" podUID="91c21c9f-aed2-41c3-b3fa-371f9903350a" Nov 21 14:14:20 crc kubenswrapper[4897]: I1121 14:14:20.955399 4897 generic.go:334] "Generic (PLEG): container finished" podID="d0fb95e9-9ac7-46bf-996a-0cef021092fe" containerID="6432a2338bff951cfec27f41f5dce6cdee3dc2e4c42bd051e8fedc5c4c6c7d23" exitCode=0 Nov 21 14:14:20 crc kubenswrapper[4897]: I1121 14:14:20.955479 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"d0fb95e9-9ac7-46bf-996a-0cef021092fe","Type":"ContainerDied","Data":"6432a2338bff951cfec27f41f5dce6cdee3dc2e4c42bd051e8fedc5c4c6c7d23"} Nov 21 14:14:21 crc kubenswrapper[4897]: E1121 14:14:21.964643 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-state-metrics\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47dcd507a8ad265c7ebd6b128bb9bdaeb7688b5731503817b94ae1d1badd6a77\\\"\"" pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-tm69l" podUID="91c21c9f-aed2-41c3-b3fa-371f9903350a" Nov 21 14:14:27 crc kubenswrapper[4897]: I1121 14:14:27.908489 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-5f888bf76d-vtk2t" Nov 21 14:14:27 crc kubenswrapper[4897]: I1121 14:14:27.909056 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-5f888bf76d-vtk2t" Nov 21 14:14:27 crc kubenswrapper[4897]: I1121 14:14:27.914327 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-5f888bf76d-vtk2t" Nov 21 14:14:28 crc kubenswrapper[4897]: I1121 14:14:28.016237 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-5f888bf76d-vtk2t" Nov 21 14:14:28 crc kubenswrapper[4897]: I1121 14:14:28.061316 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-7fsr9"] Nov 21 14:14:34 crc kubenswrapper[4897]: I1121 14:14:34.055550 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/openshift-state-metrics-566fddb674-hw2hf" event={"ID":"8d0ea2ea-c9a1-495a-a626-7a1a735308df","Type":"ContainerStarted","Data":"ca5f3148f9652cd2c2b506cdf8c916e748dae62fc40a53d1becba3c69ef445d6"} Nov 21 14:14:34 crc kubenswrapper[4897]: I1121 14:14:34.058596 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"d0fb95e9-9ac7-46bf-996a-0cef021092fe","Type":"ContainerStarted","Data":"5f933eaa48bcdde9fd1cb3ca242847f7ba95b424b57ff73807fa92c3cbd513a6"} Nov 21 14:14:34 crc kubenswrapper[4897]: I1121 14:14:34.060550 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/monitoring-plugin-7869668964-7jkch" event={"ID":"ee8ae79c-5683-46c4-909d-d9338c398678","Type":"ContainerStarted","Data":"009c34382a75150a48f7c33c22f750f8d641fc9c4f099cf38c6cf640a0f69c55"} Nov 21 14:14:35 crc kubenswrapper[4897]: I1121 14:14:35.071378 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"d0fb95e9-9ac7-46bf-996a-0cef021092fe","Type":"ContainerStarted","Data":"18df048fe9f5018bc21522d9ddaf50417f54d5587c62943a186942aab18e187b"} Nov 21 14:14:35 crc kubenswrapper[4897]: I1121 14:14:35.073485 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"2ac1afd5-988d-447d-b7f1-07476917ac23","Type":"ContainerStarted","Data":"bb69627dce10f84525e2c17d9ea421e47ce7a1aaa208493a3d09f3641873dd29"} Nov 21 14:14:35 crc kubenswrapper[4897]: I1121 14:14:35.073532 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"2ac1afd5-988d-447d-b7f1-07476917ac23","Type":"ContainerStarted","Data":"cb4f2ee0b052198e00cf05102ae3d1a31f758505e89408c470ade6b0b1c3ffb7"} Nov 21 14:14:35 crc kubenswrapper[4897]: I1121 14:14:35.075249 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-7c6484df6-g5h9n" event={"ID":"7ab7655b-d56b-4ba0-982a-c4649864016e","Type":"ContainerStarted","Data":"248e0c9d1e8d9305075ee5f208ba1a0a369f118eaaaa15242a0d1e738fa6facd"} Nov 21 14:14:35 crc kubenswrapper[4897]: I1121 14:14:35.075281 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-7c6484df6-g5h9n" event={"ID":"7ab7655b-d56b-4ba0-982a-c4649864016e","Type":"ContainerStarted","Data":"b05a1edafcc78686d3f2bc3d24ff218e0ba38f182d46de7ce8ee56c31f999872"} Nov 21 14:14:35 crc kubenswrapper[4897]: I1121 14:14:35.075296 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-7c6484df6-g5h9n" event={"ID":"7ab7655b-d56b-4ba0-982a-c4649864016e","Type":"ContainerStarted","Data":"1ef797699dd93300ea7838b52055ee6d838e1e7ed8706ffcd18dd38b3f3b6315"} Nov 21 14:14:35 crc kubenswrapper[4897]: I1121 14:14:35.077563 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/metrics-server-586c8bb899-r6zbz" event={"ID":"97e1c0fe-f3e3-4afe-819e-eb85875bee86","Type":"ContainerStarted","Data":"d09bbab8a515bb5c16ff147583b3d045d2dcc8463626e5a153f8372f28b96007"} Nov 21 14:14:35 crc kubenswrapper[4897]: I1121 14:14:35.078074 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/monitoring-plugin-7869668964-7jkch" Nov 21 14:14:35 crc kubenswrapper[4897]: I1121 14:14:35.083185 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/monitoring-plugin-7869668964-7jkch" Nov 21 14:14:35 crc kubenswrapper[4897]: I1121 14:14:35.090693 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/monitoring-plugin-7869668964-7jkch" podStartSLOduration=2.971789197 podStartE2EDuration="30.09067558s" podCreationTimestamp="2025-11-21 14:14:05 +0000 UTC" firstStartedPulling="2025-11-21 14:14:06.375894695 +0000 UTC m=+323.660488170" lastFinishedPulling="2025-11-21 14:14:33.494781078 +0000 UTC m=+350.779374553" observedRunningTime="2025-11-21 14:14:35.088423296 +0000 UTC m=+352.373016771" watchObservedRunningTime="2025-11-21 14:14:35.09067558 +0000 UTC m=+352.375269055" Nov 21 14:14:35 crc kubenswrapper[4897]: I1121 14:14:35.105961 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/openshift-state-metrics-566fddb674-hw2hf" podStartSLOduration=22.386079629 podStartE2EDuration="36.105942969s" podCreationTimestamp="2025-11-21 14:13:59 +0000 UTC" firstStartedPulling="2025-11-21 14:14:19.840198645 +0000 UTC m=+337.124792110" lastFinishedPulling="2025-11-21 14:14:33.560061975 +0000 UTC m=+350.844655450" observedRunningTime="2025-11-21 14:14:35.101296357 +0000 UTC m=+352.385889832" watchObservedRunningTime="2025-11-21 14:14:35.105942969 +0000 UTC m=+352.390536454" Nov 21 14:14:35 crc kubenswrapper[4897]: I1121 14:14:35.116640 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/metrics-server-586c8bb899-r6zbz" podStartSLOduration=2.516493076 podStartE2EDuration="30.116620157s" podCreationTimestamp="2025-11-21 14:14:05 +0000 UTC" firstStartedPulling="2025-11-21 14:14:05.894652557 +0000 UTC m=+323.179246032" lastFinishedPulling="2025-11-21 14:14:33.494779638 +0000 UTC m=+350.779373113" observedRunningTime="2025-11-21 14:14:35.11635375 +0000 UTC m=+352.400947225" watchObservedRunningTime="2025-11-21 14:14:35.116620157 +0000 UTC m=+352.401213642" Nov 21 14:14:35 crc kubenswrapper[4897]: I1121 14:14:35.646199 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-5z687" podUID="9f63b8dd-ffb7-4963-8fcc-0c133a2c045b" containerName="registry" containerID="cri-o://79ade4cd758c4fb4a6d2ad5e6f0ed42b487b6acd04c14586c067c870f5026815" gracePeriod=30 Nov 21 14:14:36 crc kubenswrapper[4897]: I1121 14:14:36.107177 4897 generic.go:334] "Generic (PLEG): container finished" podID="9f63b8dd-ffb7-4963-8fcc-0c133a2c045b" containerID="79ade4cd758c4fb4a6d2ad5e6f0ed42b487b6acd04c14586c067c870f5026815" exitCode=0 Nov 21 14:14:36 crc kubenswrapper[4897]: I1121 14:14:36.112316 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"d0fb95e9-9ac7-46bf-996a-0cef021092fe","Type":"ContainerStarted","Data":"43374bf6b29e75844acabf12a51a565deb02410df16ff020e0b9aa3d95e13f61"} Nov 21 14:14:36 crc kubenswrapper[4897]: I1121 14:14:36.112381 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"d0fb95e9-9ac7-46bf-996a-0cef021092fe","Type":"ContainerStarted","Data":"ac290e6f51a1c6d48503a5ef77e9336bcb74ba8643902897505cb7929d7a2aed"} Nov 21 14:14:36 crc kubenswrapper[4897]: I1121 14:14:36.112407 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-5z687" event={"ID":"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b","Type":"ContainerDied","Data":"79ade4cd758c4fb4a6d2ad5e6f0ed42b487b6acd04c14586c067c870f5026815"} Nov 21 14:14:36 crc kubenswrapper[4897]: I1121 14:14:36.112437 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-5z687" event={"ID":"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b","Type":"ContainerDied","Data":"d3abd9e080efefb4cc6168f2aa1c73f860bf0a037e32d66c5ff618ed90540b37"} Nov 21 14:14:36 crc kubenswrapper[4897]: I1121 14:14:36.112455 4897 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d3abd9e080efefb4cc6168f2aa1c73f860bf0a037e32d66c5ff618ed90540b37" Nov 21 14:14:36 crc kubenswrapper[4897]: I1121 14:14:36.119396 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-tm69l" event={"ID":"91c21c9f-aed2-41c3-b3fa-371f9903350a","Type":"ContainerStarted","Data":"251616e03b5d16b97c34ff443b2a48ba428640a454c31ea220342bb75610d23c"} Nov 21 14:14:36 crc kubenswrapper[4897]: I1121 14:14:36.124625 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"2ac1afd5-988d-447d-b7f1-07476917ac23","Type":"ContainerStarted","Data":"c9107508aa85b359c2aa51fa4bad2937068e6f07ab13dd875a554252e816714b"} Nov 21 14:14:36 crc kubenswrapper[4897]: I1121 14:14:36.124660 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"2ac1afd5-988d-447d-b7f1-07476917ac23","Type":"ContainerStarted","Data":"f3f30ae0b9cd651bbb69b4af93981c35347dc456ddff10219c5d11004346fead"} Nov 21 14:14:36 crc kubenswrapper[4897]: I1121 14:14:36.129187 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:14:36 crc kubenswrapper[4897]: I1121 14:14:36.144426 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-tm69l" podStartSLOduration=3.505489745 podStartE2EDuration="37.144404172s" podCreationTimestamp="2025-11-21 14:13:59 +0000 UTC" firstStartedPulling="2025-11-21 14:14:01.264874328 +0000 UTC m=+318.549467803" lastFinishedPulling="2025-11-21 14:14:34.903788745 +0000 UTC m=+352.188382230" observedRunningTime="2025-11-21 14:14:36.142769243 +0000 UTC m=+353.427362708" watchObservedRunningTime="2025-11-21 14:14:36.144404172 +0000 UTC m=+353.428997647" Nov 21 14:14:36 crc kubenswrapper[4897]: I1121 14:14:36.203494 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-86gsj\" (UniqueName: \"kubernetes.io/projected/9f63b8dd-ffb7-4963-8fcc-0c133a2c045b-kube-api-access-86gsj\") pod \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " Nov 21 14:14:36 crc kubenswrapper[4897]: I1121 14:14:36.204235 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/9f63b8dd-ffb7-4963-8fcc-0c133a2c045b-installation-pull-secrets\") pod \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " Nov 21 14:14:36 crc kubenswrapper[4897]: I1121 14:14:36.204298 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9f63b8dd-ffb7-4963-8fcc-0c133a2c045b-trusted-ca\") pod \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " Nov 21 14:14:36 crc kubenswrapper[4897]: I1121 14:14:36.204337 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/9f63b8dd-ffb7-4963-8fcc-0c133a2c045b-registry-tls\") pod \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " Nov 21 14:14:36 crc kubenswrapper[4897]: I1121 14:14:36.204361 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9f63b8dd-ffb7-4963-8fcc-0c133a2c045b-bound-sa-token\") pod \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " Nov 21 14:14:36 crc kubenswrapper[4897]: I1121 14:14:36.204394 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/9f63b8dd-ffb7-4963-8fcc-0c133a2c045b-ca-trust-extracted\") pod \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " Nov 21 14:14:36 crc kubenswrapper[4897]: I1121 14:14:36.205820 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9f63b8dd-ffb7-4963-8fcc-0c133a2c045b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9f63b8dd-ffb7-4963-8fcc-0c133a2c045b" (UID: "9f63b8dd-ffb7-4963-8fcc-0c133a2c045b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:14:36 crc kubenswrapper[4897]: I1121 14:14:36.205901 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " Nov 21 14:14:36 crc kubenswrapper[4897]: I1121 14:14:36.205995 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/9f63b8dd-ffb7-4963-8fcc-0c133a2c045b-registry-certificates\") pod \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\" (UID: \"9f63b8dd-ffb7-4963-8fcc-0c133a2c045b\") " Nov 21 14:14:36 crc kubenswrapper[4897]: I1121 14:14:36.206760 4897 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9f63b8dd-ffb7-4963-8fcc-0c133a2c045b-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 21 14:14:36 crc kubenswrapper[4897]: I1121 14:14:36.210433 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9f63b8dd-ffb7-4963-8fcc-0c133a2c045b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "9f63b8dd-ffb7-4963-8fcc-0c133a2c045b" (UID: "9f63b8dd-ffb7-4963-8fcc-0c133a2c045b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:14:36 crc kubenswrapper[4897]: I1121 14:14:36.212746 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f63b8dd-ffb7-4963-8fcc-0c133a2c045b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "9f63b8dd-ffb7-4963-8fcc-0c133a2c045b" (UID: "9f63b8dd-ffb7-4963-8fcc-0c133a2c045b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:14:36 crc kubenswrapper[4897]: I1121 14:14:36.216015 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f63b8dd-ffb7-4963-8fcc-0c133a2c045b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "9f63b8dd-ffb7-4963-8fcc-0c133a2c045b" (UID: "9f63b8dd-ffb7-4963-8fcc-0c133a2c045b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:14:36 crc kubenswrapper[4897]: I1121 14:14:36.219766 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f63b8dd-ffb7-4963-8fcc-0c133a2c045b-kube-api-access-86gsj" (OuterVolumeSpecName: "kube-api-access-86gsj") pod "9f63b8dd-ffb7-4963-8fcc-0c133a2c045b" (UID: "9f63b8dd-ffb7-4963-8fcc-0c133a2c045b"). InnerVolumeSpecName "kube-api-access-86gsj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:14:36 crc kubenswrapper[4897]: I1121 14:14:36.221021 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f63b8dd-ffb7-4963-8fcc-0c133a2c045b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "9f63b8dd-ffb7-4963-8fcc-0c133a2c045b" (UID: "9f63b8dd-ffb7-4963-8fcc-0c133a2c045b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:14:36 crc kubenswrapper[4897]: I1121 14:14:36.224433 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "9f63b8dd-ffb7-4963-8fcc-0c133a2c045b" (UID: "9f63b8dd-ffb7-4963-8fcc-0c133a2c045b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 21 14:14:36 crc kubenswrapper[4897]: I1121 14:14:36.247591 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9f63b8dd-ffb7-4963-8fcc-0c133a2c045b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "9f63b8dd-ffb7-4963-8fcc-0c133a2c045b" (UID: "9f63b8dd-ffb7-4963-8fcc-0c133a2c045b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:14:36 crc kubenswrapper[4897]: I1121 14:14:36.308579 4897 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/9f63b8dd-ffb7-4963-8fcc-0c133a2c045b-registry-tls\") on node \"crc\" DevicePath \"\"" Nov 21 14:14:36 crc kubenswrapper[4897]: I1121 14:14:36.308619 4897 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9f63b8dd-ffb7-4963-8fcc-0c133a2c045b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 21 14:14:36 crc kubenswrapper[4897]: I1121 14:14:36.308632 4897 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/9f63b8dd-ffb7-4963-8fcc-0c133a2c045b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Nov 21 14:14:36 crc kubenswrapper[4897]: I1121 14:14:36.308644 4897 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/9f63b8dd-ffb7-4963-8fcc-0c133a2c045b-registry-certificates\") on node \"crc\" DevicePath \"\"" Nov 21 14:14:36 crc kubenswrapper[4897]: I1121 14:14:36.308657 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-86gsj\" (UniqueName: \"kubernetes.io/projected/9f63b8dd-ffb7-4963-8fcc-0c133a2c045b-kube-api-access-86gsj\") on node \"crc\" DevicePath \"\"" Nov 21 14:14:36 crc kubenswrapper[4897]: I1121 14:14:36.308671 4897 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/9f63b8dd-ffb7-4963-8fcc-0c133a2c045b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Nov 21 14:14:37 crc kubenswrapper[4897]: I1121 14:14:37.132974 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"d0fb95e9-9ac7-46bf-996a-0cef021092fe","Type":"ContainerStarted","Data":"764e6b6b91228580d616ebe1ae0ce46b356f28ce956035e5112896162e0f7cc0"} Nov 21 14:14:37 crc kubenswrapper[4897]: I1121 14:14:37.133316 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"d0fb95e9-9ac7-46bf-996a-0cef021092fe","Type":"ContainerStarted","Data":"54250327bdf981617777cbc80390d4cf53903a83544cd81d0d5c190bf9646b94"} Nov 21 14:14:37 crc kubenswrapper[4897]: I1121 14:14:37.135522 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-9gpbl" event={"ID":"8caaec7a-8eab-4ffe-9e68-f5c763d044eb","Type":"ContainerDied","Data":"586e2b5a071e77566e65b64e96168c28a8278450f9759299e4fd249a4b6fa229"} Nov 21 14:14:37 crc kubenswrapper[4897]: I1121 14:14:37.135488 4897 generic.go:334] "Generic (PLEG): container finished" podID="8caaec7a-8eab-4ffe-9e68-f5c763d044eb" containerID="586e2b5a071e77566e65b64e96168c28a8278450f9759299e4fd249a4b6fa229" exitCode=0 Nov 21 14:14:37 crc kubenswrapper[4897]: I1121 14:14:37.142009 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"2ac1afd5-988d-447d-b7f1-07476917ac23","Type":"ContainerStarted","Data":"09f41d292ee2bd205be3c8fb1ce9c7d605f777cdf853a2939e3dda985bbe3c5f"} Nov 21 14:14:37 crc kubenswrapper[4897]: I1121 14:14:37.142037 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"2ac1afd5-988d-447d-b7f1-07476917ac23","Type":"ContainerStarted","Data":"c90c853f5b1774716b964eb39efee1434fd0a3e4e0040b69f5471c6aa604fc7d"} Nov 21 14:14:37 crc kubenswrapper[4897]: I1121 14:14:37.145397 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-5z687" Nov 21 14:14:37 crc kubenswrapper[4897]: I1121 14:14:37.146317 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-7c6484df6-g5h9n" event={"ID":"7ab7655b-d56b-4ba0-982a-c4649864016e","Type":"ContainerStarted","Data":"114be0e03a2e300ac2897f9f8de89a3f6c5bb0bb1d6d5f0b94e2dd3ca4914297"} Nov 21 14:14:37 crc kubenswrapper[4897]: I1121 14:14:37.146348 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-7c6484df6-g5h9n" event={"ID":"7ab7655b-d56b-4ba0-982a-c4649864016e","Type":"ContainerStarted","Data":"9f7a2faa84777b4aa04b0496f29542d73d3ba4888c231d5ba3fcbb1db1f4ec96"} Nov 21 14:14:37 crc kubenswrapper[4897]: I1121 14:14:37.146357 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-7c6484df6-g5h9n" event={"ID":"7ab7655b-d56b-4ba0-982a-c4649864016e","Type":"ContainerStarted","Data":"9094c906b8dc3cfcdfd8c75ac0b25684aa7d5e54025aa27974453097082a4537"} Nov 21 14:14:37 crc kubenswrapper[4897]: I1121 14:14:37.173045 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/alertmanager-main-0" podStartSLOduration=1.855179321 podStartE2EDuration="37.173021126s" podCreationTimestamp="2025-11-21 14:14:00 +0000 UTC" firstStartedPulling="2025-11-21 14:14:01.345712632 +0000 UTC m=+318.630306107" lastFinishedPulling="2025-11-21 14:14:36.663554437 +0000 UTC m=+353.948147912" observedRunningTime="2025-11-21 14:14:37.159671084 +0000 UTC m=+354.444264569" watchObservedRunningTime="2025-11-21 14:14:37.173021126 +0000 UTC m=+354.457614601" Nov 21 14:14:37 crc kubenswrapper[4897]: I1121 14:14:37.173876 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/thanos-querier-7c6484df6-g5h9n" Nov 21 14:14:37 crc kubenswrapper[4897]: I1121 14:14:37.206579 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-5z687"] Nov 21 14:14:37 crc kubenswrapper[4897]: I1121 14:14:37.212335 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-5z687"] Nov 21 14:14:37 crc kubenswrapper[4897]: I1121 14:14:37.224608 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/thanos-querier-7c6484df6-g5h9n" podStartSLOduration=2.057759209 podStartE2EDuration="36.224589412s" podCreationTimestamp="2025-11-21 14:14:01 +0000 UTC" firstStartedPulling="2025-11-21 14:14:02.394919833 +0000 UTC m=+319.679513308" lastFinishedPulling="2025-11-21 14:14:36.561750036 +0000 UTC m=+353.846343511" observedRunningTime="2025-11-21 14:14:37.223915227 +0000 UTC m=+354.508508712" watchObservedRunningTime="2025-11-21 14:14:37.224589412 +0000 UTC m=+354.509182887" Nov 21 14:14:37 crc kubenswrapper[4897]: I1121 14:14:37.257161 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/prometheus-k8s-0" podStartSLOduration=4.516729841 podStartE2EDuration="31.257143569s" podCreationTimestamp="2025-11-21 14:14:06 +0000 UTC" firstStartedPulling="2025-11-21 14:14:06.850725768 +0000 UTC m=+324.135319243" lastFinishedPulling="2025-11-21 14:14:33.591139496 +0000 UTC m=+350.875732971" observedRunningTime="2025-11-21 14:14:37.252924557 +0000 UTC m=+354.537518032" watchObservedRunningTime="2025-11-21 14:14:37.257143569 +0000 UTC m=+354.541737044" Nov 21 14:14:38 crc kubenswrapper[4897]: I1121 14:14:38.096685 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9f63b8dd-ffb7-4963-8fcc-0c133a2c045b" path="/var/lib/kubelet/pods/9f63b8dd-ffb7-4963-8fcc-0c133a2c045b/volumes" Nov 21 14:14:38 crc kubenswrapper[4897]: I1121 14:14:38.152156 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-9gpbl" event={"ID":"8caaec7a-8eab-4ffe-9e68-f5c763d044eb","Type":"ContainerStarted","Data":"eedff14632a942cb0006eb956a07146a9db29d905af39ebf12bf08553c57e0db"} Nov 21 14:14:38 crc kubenswrapper[4897]: I1121 14:14:38.152215 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-9gpbl" event={"ID":"8caaec7a-8eab-4ffe-9e68-f5c763d044eb","Type":"ContainerStarted","Data":"1e4ce66c2893332c73d38014063b92917386a181c535a84fe52381c00ccc9a33"} Nov 21 14:14:39 crc kubenswrapper[4897]: I1121 14:14:39.167877 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/thanos-querier-7c6484df6-g5h9n" Nov 21 14:14:39 crc kubenswrapper[4897]: I1121 14:14:39.189440 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/node-exporter-9gpbl" podStartSLOduration=5.052658807 podStartE2EDuration="40.189420999s" podCreationTimestamp="2025-11-21 14:13:59 +0000 UTC" firstStartedPulling="2025-11-21 14:14:00.869129773 +0000 UTC m=+318.153723248" lastFinishedPulling="2025-11-21 14:14:36.005891945 +0000 UTC m=+353.290485440" observedRunningTime="2025-11-21 14:14:38.171288037 +0000 UTC m=+355.455881532" watchObservedRunningTime="2025-11-21 14:14:39.189420999 +0000 UTC m=+356.474014474" Nov 21 14:14:41 crc kubenswrapper[4897]: I1121 14:14:41.405970 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/prometheus-k8s-0" Nov 21 14:14:45 crc kubenswrapper[4897]: I1121 14:14:45.478736 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/metrics-server-586c8bb899-r6zbz" Nov 21 14:14:45 crc kubenswrapper[4897]: I1121 14:14:45.479871 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-monitoring/metrics-server-586c8bb899-r6zbz" Nov 21 14:14:53 crc kubenswrapper[4897]: I1121 14:14:53.103674 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-7fsr9" podUID="8342c8a2-87f4-4bf9-a2ca-073c02b277ad" containerName="console" containerID="cri-o://8a96362523e6d8f3d19ad31a419973897ae10647ff86cb917812b0afe685363b" gracePeriod=15 Nov 21 14:14:53 crc kubenswrapper[4897]: I1121 14:14:53.236129 4897 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-7fsr9_8342c8a2-87f4-4bf9-a2ca-073c02b277ad/console/0.log" Nov 21 14:14:53 crc kubenswrapper[4897]: I1121 14:14:53.236409 4897 generic.go:334] "Generic (PLEG): container finished" podID="8342c8a2-87f4-4bf9-a2ca-073c02b277ad" containerID="8a96362523e6d8f3d19ad31a419973897ae10647ff86cb917812b0afe685363b" exitCode=2 Nov 21 14:14:53 crc kubenswrapper[4897]: I1121 14:14:53.236440 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-7fsr9" event={"ID":"8342c8a2-87f4-4bf9-a2ca-073c02b277ad","Type":"ContainerDied","Data":"8a96362523e6d8f3d19ad31a419973897ae10647ff86cb917812b0afe685363b"} Nov 21 14:14:53 crc kubenswrapper[4897]: I1121 14:14:53.440392 4897 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-7fsr9_8342c8a2-87f4-4bf9-a2ca-073c02b277ad/console/0.log" Nov 21 14:14:53 crc kubenswrapper[4897]: I1121 14:14:53.440451 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-7fsr9" Nov 21 14:14:53 crc kubenswrapper[4897]: I1121 14:14:53.550648 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/8342c8a2-87f4-4bf9-a2ca-073c02b277ad-console-serving-cert\") pod \"8342c8a2-87f4-4bf9-a2ca-073c02b277ad\" (UID: \"8342c8a2-87f4-4bf9-a2ca-073c02b277ad\") " Nov 21 14:14:53 crc kubenswrapper[4897]: I1121 14:14:53.550682 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/8342c8a2-87f4-4bf9-a2ca-073c02b277ad-oauth-serving-cert\") pod \"8342c8a2-87f4-4bf9-a2ca-073c02b277ad\" (UID: \"8342c8a2-87f4-4bf9-a2ca-073c02b277ad\") " Nov 21 14:14:53 crc kubenswrapper[4897]: I1121 14:14:53.550758 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lsdt5\" (UniqueName: \"kubernetes.io/projected/8342c8a2-87f4-4bf9-a2ca-073c02b277ad-kube-api-access-lsdt5\") pod \"8342c8a2-87f4-4bf9-a2ca-073c02b277ad\" (UID: \"8342c8a2-87f4-4bf9-a2ca-073c02b277ad\") " Nov 21 14:14:53 crc kubenswrapper[4897]: I1121 14:14:53.550814 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/8342c8a2-87f4-4bf9-a2ca-073c02b277ad-console-oauth-config\") pod \"8342c8a2-87f4-4bf9-a2ca-073c02b277ad\" (UID: \"8342c8a2-87f4-4bf9-a2ca-073c02b277ad\") " Nov 21 14:14:53 crc kubenswrapper[4897]: I1121 14:14:53.551678 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/8342c8a2-87f4-4bf9-a2ca-073c02b277ad-console-config\") pod \"8342c8a2-87f4-4bf9-a2ca-073c02b277ad\" (UID: \"8342c8a2-87f4-4bf9-a2ca-073c02b277ad\") " Nov 21 14:14:53 crc kubenswrapper[4897]: I1121 14:14:53.551729 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8342c8a2-87f4-4bf9-a2ca-073c02b277ad-trusted-ca-bundle\") pod \"8342c8a2-87f4-4bf9-a2ca-073c02b277ad\" (UID: \"8342c8a2-87f4-4bf9-a2ca-073c02b277ad\") " Nov 21 14:14:53 crc kubenswrapper[4897]: I1121 14:14:53.551761 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8342c8a2-87f4-4bf9-a2ca-073c02b277ad-service-ca\") pod \"8342c8a2-87f4-4bf9-a2ca-073c02b277ad\" (UID: \"8342c8a2-87f4-4bf9-a2ca-073c02b277ad\") " Nov 21 14:14:53 crc kubenswrapper[4897]: I1121 14:14:53.552846 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8342c8a2-87f4-4bf9-a2ca-073c02b277ad-service-ca" (OuterVolumeSpecName: "service-ca") pod "8342c8a2-87f4-4bf9-a2ca-073c02b277ad" (UID: "8342c8a2-87f4-4bf9-a2ca-073c02b277ad"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:14:53 crc kubenswrapper[4897]: I1121 14:14:53.552883 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8342c8a2-87f4-4bf9-a2ca-073c02b277ad-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "8342c8a2-87f4-4bf9-a2ca-073c02b277ad" (UID: "8342c8a2-87f4-4bf9-a2ca-073c02b277ad"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:14:53 crc kubenswrapper[4897]: I1121 14:14:53.553220 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8342c8a2-87f4-4bf9-a2ca-073c02b277ad-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "8342c8a2-87f4-4bf9-a2ca-073c02b277ad" (UID: "8342c8a2-87f4-4bf9-a2ca-073c02b277ad"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:14:53 crc kubenswrapper[4897]: I1121 14:14:53.554242 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8342c8a2-87f4-4bf9-a2ca-073c02b277ad-console-config" (OuterVolumeSpecName: "console-config") pod "8342c8a2-87f4-4bf9-a2ca-073c02b277ad" (UID: "8342c8a2-87f4-4bf9-a2ca-073c02b277ad"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:14:53 crc kubenswrapper[4897]: I1121 14:14:53.555827 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8342c8a2-87f4-4bf9-a2ca-073c02b277ad-kube-api-access-lsdt5" (OuterVolumeSpecName: "kube-api-access-lsdt5") pod "8342c8a2-87f4-4bf9-a2ca-073c02b277ad" (UID: "8342c8a2-87f4-4bf9-a2ca-073c02b277ad"). InnerVolumeSpecName "kube-api-access-lsdt5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:14:53 crc kubenswrapper[4897]: I1121 14:14:53.556137 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8342c8a2-87f4-4bf9-a2ca-073c02b277ad-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "8342c8a2-87f4-4bf9-a2ca-073c02b277ad" (UID: "8342c8a2-87f4-4bf9-a2ca-073c02b277ad"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:14:53 crc kubenswrapper[4897]: I1121 14:14:53.556272 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8342c8a2-87f4-4bf9-a2ca-073c02b277ad-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "8342c8a2-87f4-4bf9-a2ca-073c02b277ad" (UID: "8342c8a2-87f4-4bf9-a2ca-073c02b277ad"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:14:53 crc kubenswrapper[4897]: I1121 14:14:53.653023 4897 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/8342c8a2-87f4-4bf9-a2ca-073c02b277ad-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 21 14:14:53 crc kubenswrapper[4897]: I1121 14:14:53.653062 4897 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/8342c8a2-87f4-4bf9-a2ca-073c02b277ad-console-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 21 14:14:53 crc kubenswrapper[4897]: I1121 14:14:53.653072 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lsdt5\" (UniqueName: \"kubernetes.io/projected/8342c8a2-87f4-4bf9-a2ca-073c02b277ad-kube-api-access-lsdt5\") on node \"crc\" DevicePath \"\"" Nov 21 14:14:53 crc kubenswrapper[4897]: I1121 14:14:53.653087 4897 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/8342c8a2-87f4-4bf9-a2ca-073c02b277ad-console-oauth-config\") on node \"crc\" DevicePath \"\"" Nov 21 14:14:53 crc kubenswrapper[4897]: I1121 14:14:53.653098 4897 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/8342c8a2-87f4-4bf9-a2ca-073c02b277ad-console-config\") on node \"crc\" DevicePath \"\"" Nov 21 14:14:53 crc kubenswrapper[4897]: I1121 14:14:53.653108 4897 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8342c8a2-87f4-4bf9-a2ca-073c02b277ad-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 21 14:14:53 crc kubenswrapper[4897]: I1121 14:14:53.653116 4897 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8342c8a2-87f4-4bf9-a2ca-073c02b277ad-service-ca\") on node \"crc\" DevicePath \"\"" Nov 21 14:14:54 crc kubenswrapper[4897]: I1121 14:14:54.242736 4897 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-7fsr9_8342c8a2-87f4-4bf9-a2ca-073c02b277ad/console/0.log" Nov 21 14:14:54 crc kubenswrapper[4897]: I1121 14:14:54.242805 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-7fsr9" event={"ID":"8342c8a2-87f4-4bf9-a2ca-073c02b277ad","Type":"ContainerDied","Data":"59c4e9d029bb293f0c9708784213d9a311f8bc3f091225f8b32b8f91cd416fb6"} Nov 21 14:14:54 crc kubenswrapper[4897]: I1121 14:14:54.242859 4897 scope.go:117] "RemoveContainer" containerID="8a96362523e6d8f3d19ad31a419973897ae10647ff86cb917812b0afe685363b" Nov 21 14:14:54 crc kubenswrapper[4897]: I1121 14:14:54.242886 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-7fsr9" Nov 21 14:14:54 crc kubenswrapper[4897]: I1121 14:14:54.265075 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-7fsr9"] Nov 21 14:14:54 crc kubenswrapper[4897]: I1121 14:14:54.269056 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-7fsr9"] Nov 21 14:14:56 crc kubenswrapper[4897]: I1121 14:14:56.107345 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8342c8a2-87f4-4bf9-a2ca-073c02b277ad" path="/var/lib/kubelet/pods/8342c8a2-87f4-4bf9-a2ca-073c02b277ad/volumes" Nov 21 14:15:00 crc kubenswrapper[4897]: I1121 14:15:00.125622 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29395575-tx6lf"] Nov 21 14:15:00 crc kubenswrapper[4897]: E1121 14:15:00.126129 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8342c8a2-87f4-4bf9-a2ca-073c02b277ad" containerName="console" Nov 21 14:15:00 crc kubenswrapper[4897]: I1121 14:15:00.126141 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="8342c8a2-87f4-4bf9-a2ca-073c02b277ad" containerName="console" Nov 21 14:15:00 crc kubenswrapper[4897]: E1121 14:15:00.126155 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f63b8dd-ffb7-4963-8fcc-0c133a2c045b" containerName="registry" Nov 21 14:15:00 crc kubenswrapper[4897]: I1121 14:15:00.126161 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f63b8dd-ffb7-4963-8fcc-0c133a2c045b" containerName="registry" Nov 21 14:15:00 crc kubenswrapper[4897]: I1121 14:15:00.126271 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f63b8dd-ffb7-4963-8fcc-0c133a2c045b" containerName="registry" Nov 21 14:15:00 crc kubenswrapper[4897]: I1121 14:15:00.126284 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="8342c8a2-87f4-4bf9-a2ca-073c02b277ad" containerName="console" Nov 21 14:15:00 crc kubenswrapper[4897]: I1121 14:15:00.126671 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29395575-tx6lf" Nov 21 14:15:00 crc kubenswrapper[4897]: I1121 14:15:00.128350 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 21 14:15:00 crc kubenswrapper[4897]: I1121 14:15:00.128359 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 21 14:15:00 crc kubenswrapper[4897]: I1121 14:15:00.141561 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29395575-tx6lf"] Nov 21 14:15:00 crc kubenswrapper[4897]: I1121 14:15:00.238089 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e7def41c-7fc4-44a5-8ae4-71ebd66ff255-secret-volume\") pod \"collect-profiles-29395575-tx6lf\" (UID: \"e7def41c-7fc4-44a5-8ae4-71ebd66ff255\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29395575-tx6lf" Nov 21 14:15:00 crc kubenswrapper[4897]: I1121 14:15:00.238159 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tlg5b\" (UniqueName: \"kubernetes.io/projected/e7def41c-7fc4-44a5-8ae4-71ebd66ff255-kube-api-access-tlg5b\") pod \"collect-profiles-29395575-tx6lf\" (UID: \"e7def41c-7fc4-44a5-8ae4-71ebd66ff255\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29395575-tx6lf" Nov 21 14:15:00 crc kubenswrapper[4897]: I1121 14:15:00.238246 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e7def41c-7fc4-44a5-8ae4-71ebd66ff255-config-volume\") pod \"collect-profiles-29395575-tx6lf\" (UID: \"e7def41c-7fc4-44a5-8ae4-71ebd66ff255\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29395575-tx6lf" Nov 21 14:15:00 crc kubenswrapper[4897]: I1121 14:15:00.340525 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e7def41c-7fc4-44a5-8ae4-71ebd66ff255-secret-volume\") pod \"collect-profiles-29395575-tx6lf\" (UID: \"e7def41c-7fc4-44a5-8ae4-71ebd66ff255\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29395575-tx6lf" Nov 21 14:15:00 crc kubenswrapper[4897]: I1121 14:15:00.340584 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tlg5b\" (UniqueName: \"kubernetes.io/projected/e7def41c-7fc4-44a5-8ae4-71ebd66ff255-kube-api-access-tlg5b\") pod \"collect-profiles-29395575-tx6lf\" (UID: \"e7def41c-7fc4-44a5-8ae4-71ebd66ff255\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29395575-tx6lf" Nov 21 14:15:00 crc kubenswrapper[4897]: I1121 14:15:00.340633 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e7def41c-7fc4-44a5-8ae4-71ebd66ff255-config-volume\") pod \"collect-profiles-29395575-tx6lf\" (UID: \"e7def41c-7fc4-44a5-8ae4-71ebd66ff255\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29395575-tx6lf" Nov 21 14:15:00 crc kubenswrapper[4897]: I1121 14:15:00.341842 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e7def41c-7fc4-44a5-8ae4-71ebd66ff255-config-volume\") pod \"collect-profiles-29395575-tx6lf\" (UID: \"e7def41c-7fc4-44a5-8ae4-71ebd66ff255\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29395575-tx6lf" Nov 21 14:15:00 crc kubenswrapper[4897]: I1121 14:15:00.350166 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e7def41c-7fc4-44a5-8ae4-71ebd66ff255-secret-volume\") pod \"collect-profiles-29395575-tx6lf\" (UID: \"e7def41c-7fc4-44a5-8ae4-71ebd66ff255\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29395575-tx6lf" Nov 21 14:15:00 crc kubenswrapper[4897]: I1121 14:15:00.355810 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tlg5b\" (UniqueName: \"kubernetes.io/projected/e7def41c-7fc4-44a5-8ae4-71ebd66ff255-kube-api-access-tlg5b\") pod \"collect-profiles-29395575-tx6lf\" (UID: \"e7def41c-7fc4-44a5-8ae4-71ebd66ff255\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29395575-tx6lf" Nov 21 14:15:00 crc kubenswrapper[4897]: I1121 14:15:00.502135 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29395575-tx6lf" Nov 21 14:15:00 crc kubenswrapper[4897]: I1121 14:15:00.688177 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29395575-tx6lf"] Nov 21 14:15:00 crc kubenswrapper[4897]: W1121 14:15:00.698808 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode7def41c_7fc4_44a5_8ae4_71ebd66ff255.slice/crio-2bbe0d003f369ab56a3efb774016adb03d581171823c5d1cad9ec2721d8d1e52 WatchSource:0}: Error finding container 2bbe0d003f369ab56a3efb774016adb03d581171823c5d1cad9ec2721d8d1e52: Status 404 returned error can't find the container with id 2bbe0d003f369ab56a3efb774016adb03d581171823c5d1cad9ec2721d8d1e52 Nov 21 14:15:01 crc kubenswrapper[4897]: I1121 14:15:01.283020 4897 generic.go:334] "Generic (PLEG): container finished" podID="e7def41c-7fc4-44a5-8ae4-71ebd66ff255" containerID="044b03bc8a8b588143aadbdbb3974f4bf577f413dc45ae7a95dd44cd65eb0b47" exitCode=0 Nov 21 14:15:01 crc kubenswrapper[4897]: I1121 14:15:01.283122 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29395575-tx6lf" event={"ID":"e7def41c-7fc4-44a5-8ae4-71ebd66ff255","Type":"ContainerDied","Data":"044b03bc8a8b588143aadbdbb3974f4bf577f413dc45ae7a95dd44cd65eb0b47"} Nov 21 14:15:01 crc kubenswrapper[4897]: I1121 14:15:01.283304 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29395575-tx6lf" event={"ID":"e7def41c-7fc4-44a5-8ae4-71ebd66ff255","Type":"ContainerStarted","Data":"2bbe0d003f369ab56a3efb774016adb03d581171823c5d1cad9ec2721d8d1e52"} Nov 21 14:15:02 crc kubenswrapper[4897]: I1121 14:15:02.518257 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29395575-tx6lf" Nov 21 14:15:02 crc kubenswrapper[4897]: I1121 14:15:02.569717 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e7def41c-7fc4-44a5-8ae4-71ebd66ff255-secret-volume\") pod \"e7def41c-7fc4-44a5-8ae4-71ebd66ff255\" (UID: \"e7def41c-7fc4-44a5-8ae4-71ebd66ff255\") " Nov 21 14:15:02 crc kubenswrapper[4897]: I1121 14:15:02.569882 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e7def41c-7fc4-44a5-8ae4-71ebd66ff255-config-volume\") pod \"e7def41c-7fc4-44a5-8ae4-71ebd66ff255\" (UID: \"e7def41c-7fc4-44a5-8ae4-71ebd66ff255\") " Nov 21 14:15:02 crc kubenswrapper[4897]: I1121 14:15:02.569932 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tlg5b\" (UniqueName: \"kubernetes.io/projected/e7def41c-7fc4-44a5-8ae4-71ebd66ff255-kube-api-access-tlg5b\") pod \"e7def41c-7fc4-44a5-8ae4-71ebd66ff255\" (UID: \"e7def41c-7fc4-44a5-8ae4-71ebd66ff255\") " Nov 21 14:15:02 crc kubenswrapper[4897]: I1121 14:15:02.570551 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7def41c-7fc4-44a5-8ae4-71ebd66ff255-config-volume" (OuterVolumeSpecName: "config-volume") pod "e7def41c-7fc4-44a5-8ae4-71ebd66ff255" (UID: "e7def41c-7fc4-44a5-8ae4-71ebd66ff255"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:15:02 crc kubenswrapper[4897]: I1121 14:15:02.574701 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7def41c-7fc4-44a5-8ae4-71ebd66ff255-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "e7def41c-7fc4-44a5-8ae4-71ebd66ff255" (UID: "e7def41c-7fc4-44a5-8ae4-71ebd66ff255"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:15:02 crc kubenswrapper[4897]: I1121 14:15:02.574832 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7def41c-7fc4-44a5-8ae4-71ebd66ff255-kube-api-access-tlg5b" (OuterVolumeSpecName: "kube-api-access-tlg5b") pod "e7def41c-7fc4-44a5-8ae4-71ebd66ff255" (UID: "e7def41c-7fc4-44a5-8ae4-71ebd66ff255"). InnerVolumeSpecName "kube-api-access-tlg5b". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:15:02 crc kubenswrapper[4897]: I1121 14:15:02.670837 4897 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e7def41c-7fc4-44a5-8ae4-71ebd66ff255-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 21 14:15:02 crc kubenswrapper[4897]: I1121 14:15:02.671258 4897 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e7def41c-7fc4-44a5-8ae4-71ebd66ff255-config-volume\") on node \"crc\" DevicePath \"\"" Nov 21 14:15:02 crc kubenswrapper[4897]: I1121 14:15:02.671295 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tlg5b\" (UniqueName: \"kubernetes.io/projected/e7def41c-7fc4-44a5-8ae4-71ebd66ff255-kube-api-access-tlg5b\") on node \"crc\" DevicePath \"\"" Nov 21 14:15:03 crc kubenswrapper[4897]: I1121 14:15:03.296226 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29395575-tx6lf" event={"ID":"e7def41c-7fc4-44a5-8ae4-71ebd66ff255","Type":"ContainerDied","Data":"2bbe0d003f369ab56a3efb774016adb03d581171823c5d1cad9ec2721d8d1e52"} Nov 21 14:15:03 crc kubenswrapper[4897]: I1121 14:15:03.296270 4897 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2bbe0d003f369ab56a3efb774016adb03d581171823c5d1cad9ec2721d8d1e52" Nov 21 14:15:03 crc kubenswrapper[4897]: I1121 14:15:03.296281 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29395575-tx6lf" Nov 21 14:15:04 crc kubenswrapper[4897]: I1121 14:15:04.371126 4897 patch_prober.go:28] interesting pod/machine-config-daemon-krv5b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 21 14:15:04 crc kubenswrapper[4897]: I1121 14:15:04.371216 4897 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 21 14:15:05 crc kubenswrapper[4897]: I1121 14:15:05.487434 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-monitoring/metrics-server-586c8bb899-r6zbz" Nov 21 14:15:05 crc kubenswrapper[4897]: I1121 14:15:05.491314 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/metrics-server-586c8bb899-r6zbz" Nov 21 14:15:06 crc kubenswrapper[4897]: I1121 14:15:06.405756 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-monitoring/prometheus-k8s-0" Nov 21 14:15:06 crc kubenswrapper[4897]: I1121 14:15:06.437216 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-monitoring/prometheus-k8s-0" Nov 21 14:15:07 crc kubenswrapper[4897]: I1121 14:15:07.340622 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/prometheus-k8s-0" Nov 21 14:15:34 crc kubenswrapper[4897]: I1121 14:15:34.371041 4897 patch_prober.go:28] interesting pod/machine-config-daemon-krv5b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 21 14:15:34 crc kubenswrapper[4897]: I1121 14:15:34.373314 4897 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 21 14:15:34 crc kubenswrapper[4897]: I1121 14:15:34.677465 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-95c6d8978-l44nd"] Nov 21 14:15:34 crc kubenswrapper[4897]: E1121 14:15:34.677992 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7def41c-7fc4-44a5-8ae4-71ebd66ff255" containerName="collect-profiles" Nov 21 14:15:34 crc kubenswrapper[4897]: I1121 14:15:34.678084 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7def41c-7fc4-44a5-8ae4-71ebd66ff255" containerName="collect-profiles" Nov 21 14:15:34 crc kubenswrapper[4897]: I1121 14:15:34.678308 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="e7def41c-7fc4-44a5-8ae4-71ebd66ff255" containerName="collect-profiles" Nov 21 14:15:34 crc kubenswrapper[4897]: I1121 14:15:34.678930 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-95c6d8978-l44nd" Nov 21 14:15:34 crc kubenswrapper[4897]: I1121 14:15:34.696418 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-95c6d8978-l44nd"] Nov 21 14:15:34 crc kubenswrapper[4897]: I1121 14:15:34.816181 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/bb7f0923-4b62-4114-8119-34c39e469323-console-config\") pod \"console-95c6d8978-l44nd\" (UID: \"bb7f0923-4b62-4114-8119-34c39e469323\") " pod="openshift-console/console-95c6d8978-l44nd" Nov 21 14:15:34 crc kubenswrapper[4897]: I1121 14:15:34.816226 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/bb7f0923-4b62-4114-8119-34c39e469323-console-oauth-config\") pod \"console-95c6d8978-l44nd\" (UID: \"bb7f0923-4b62-4114-8119-34c39e469323\") " pod="openshift-console/console-95c6d8978-l44nd" Nov 21 14:15:34 crc kubenswrapper[4897]: I1121 14:15:34.816262 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xktjm\" (UniqueName: \"kubernetes.io/projected/bb7f0923-4b62-4114-8119-34c39e469323-kube-api-access-xktjm\") pod \"console-95c6d8978-l44nd\" (UID: \"bb7f0923-4b62-4114-8119-34c39e469323\") " pod="openshift-console/console-95c6d8978-l44nd" Nov 21 14:15:34 crc kubenswrapper[4897]: I1121 14:15:34.816319 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/bb7f0923-4b62-4114-8119-34c39e469323-console-serving-cert\") pod \"console-95c6d8978-l44nd\" (UID: \"bb7f0923-4b62-4114-8119-34c39e469323\") " pod="openshift-console/console-95c6d8978-l44nd" Nov 21 14:15:34 crc kubenswrapper[4897]: I1121 14:15:34.816346 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bb7f0923-4b62-4114-8119-34c39e469323-trusted-ca-bundle\") pod \"console-95c6d8978-l44nd\" (UID: \"bb7f0923-4b62-4114-8119-34c39e469323\") " pod="openshift-console/console-95c6d8978-l44nd" Nov 21 14:15:34 crc kubenswrapper[4897]: I1121 14:15:34.816366 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/bb7f0923-4b62-4114-8119-34c39e469323-oauth-serving-cert\") pod \"console-95c6d8978-l44nd\" (UID: \"bb7f0923-4b62-4114-8119-34c39e469323\") " pod="openshift-console/console-95c6d8978-l44nd" Nov 21 14:15:34 crc kubenswrapper[4897]: I1121 14:15:34.816472 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/bb7f0923-4b62-4114-8119-34c39e469323-service-ca\") pod \"console-95c6d8978-l44nd\" (UID: \"bb7f0923-4b62-4114-8119-34c39e469323\") " pod="openshift-console/console-95c6d8978-l44nd" Nov 21 14:15:34 crc kubenswrapper[4897]: I1121 14:15:34.918016 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/bb7f0923-4b62-4114-8119-34c39e469323-service-ca\") pod \"console-95c6d8978-l44nd\" (UID: \"bb7f0923-4b62-4114-8119-34c39e469323\") " pod="openshift-console/console-95c6d8978-l44nd" Nov 21 14:15:34 crc kubenswrapper[4897]: I1121 14:15:34.918407 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/bb7f0923-4b62-4114-8119-34c39e469323-console-config\") pod \"console-95c6d8978-l44nd\" (UID: \"bb7f0923-4b62-4114-8119-34c39e469323\") " pod="openshift-console/console-95c6d8978-l44nd" Nov 21 14:15:34 crc kubenswrapper[4897]: I1121 14:15:34.918666 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/bb7f0923-4b62-4114-8119-34c39e469323-console-oauth-config\") pod \"console-95c6d8978-l44nd\" (UID: \"bb7f0923-4b62-4114-8119-34c39e469323\") " pod="openshift-console/console-95c6d8978-l44nd" Nov 21 14:15:34 crc kubenswrapper[4897]: I1121 14:15:34.918853 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xktjm\" (UniqueName: \"kubernetes.io/projected/bb7f0923-4b62-4114-8119-34c39e469323-kube-api-access-xktjm\") pod \"console-95c6d8978-l44nd\" (UID: \"bb7f0923-4b62-4114-8119-34c39e469323\") " pod="openshift-console/console-95c6d8978-l44nd" Nov 21 14:15:34 crc kubenswrapper[4897]: I1121 14:15:34.919057 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/bb7f0923-4b62-4114-8119-34c39e469323-console-serving-cert\") pod \"console-95c6d8978-l44nd\" (UID: \"bb7f0923-4b62-4114-8119-34c39e469323\") " pod="openshift-console/console-95c6d8978-l44nd" Nov 21 14:15:34 crc kubenswrapper[4897]: I1121 14:15:34.919217 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bb7f0923-4b62-4114-8119-34c39e469323-trusted-ca-bundle\") pod \"console-95c6d8978-l44nd\" (UID: \"bb7f0923-4b62-4114-8119-34c39e469323\") " pod="openshift-console/console-95c6d8978-l44nd" Nov 21 14:15:34 crc kubenswrapper[4897]: I1121 14:15:34.919361 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/bb7f0923-4b62-4114-8119-34c39e469323-oauth-serving-cert\") pod \"console-95c6d8978-l44nd\" (UID: \"bb7f0923-4b62-4114-8119-34c39e469323\") " pod="openshift-console/console-95c6d8978-l44nd" Nov 21 14:15:34 crc kubenswrapper[4897]: I1121 14:15:34.919074 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/bb7f0923-4b62-4114-8119-34c39e469323-service-ca\") pod \"console-95c6d8978-l44nd\" (UID: \"bb7f0923-4b62-4114-8119-34c39e469323\") " pod="openshift-console/console-95c6d8978-l44nd" Nov 21 14:15:34 crc kubenswrapper[4897]: I1121 14:15:34.920471 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/bb7f0923-4b62-4114-8119-34c39e469323-console-config\") pod \"console-95c6d8978-l44nd\" (UID: \"bb7f0923-4b62-4114-8119-34c39e469323\") " pod="openshift-console/console-95c6d8978-l44nd" Nov 21 14:15:34 crc kubenswrapper[4897]: I1121 14:15:34.920917 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bb7f0923-4b62-4114-8119-34c39e469323-trusted-ca-bundle\") pod \"console-95c6d8978-l44nd\" (UID: \"bb7f0923-4b62-4114-8119-34c39e469323\") " pod="openshift-console/console-95c6d8978-l44nd" Nov 21 14:15:34 crc kubenswrapper[4897]: I1121 14:15:34.921402 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/bb7f0923-4b62-4114-8119-34c39e469323-oauth-serving-cert\") pod \"console-95c6d8978-l44nd\" (UID: \"bb7f0923-4b62-4114-8119-34c39e469323\") " pod="openshift-console/console-95c6d8978-l44nd" Nov 21 14:15:34 crc kubenswrapper[4897]: I1121 14:15:34.926396 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/bb7f0923-4b62-4114-8119-34c39e469323-console-oauth-config\") pod \"console-95c6d8978-l44nd\" (UID: \"bb7f0923-4b62-4114-8119-34c39e469323\") " pod="openshift-console/console-95c6d8978-l44nd" Nov 21 14:15:34 crc kubenswrapper[4897]: I1121 14:15:34.927726 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/bb7f0923-4b62-4114-8119-34c39e469323-console-serving-cert\") pod \"console-95c6d8978-l44nd\" (UID: \"bb7f0923-4b62-4114-8119-34c39e469323\") " pod="openshift-console/console-95c6d8978-l44nd" Nov 21 14:15:34 crc kubenswrapper[4897]: I1121 14:15:34.934212 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xktjm\" (UniqueName: \"kubernetes.io/projected/bb7f0923-4b62-4114-8119-34c39e469323-kube-api-access-xktjm\") pod \"console-95c6d8978-l44nd\" (UID: \"bb7f0923-4b62-4114-8119-34c39e469323\") " pod="openshift-console/console-95c6d8978-l44nd" Nov 21 14:15:34 crc kubenswrapper[4897]: I1121 14:15:34.997170 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-95c6d8978-l44nd" Nov 21 14:15:35 crc kubenswrapper[4897]: I1121 14:15:35.177718 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-95c6d8978-l44nd"] Nov 21 14:15:35 crc kubenswrapper[4897]: I1121 14:15:35.482998 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-95c6d8978-l44nd" event={"ID":"bb7f0923-4b62-4114-8119-34c39e469323","Type":"ContainerStarted","Data":"0f15d6842e5cb4a63dd02145ff446366682c58b091cbf480855eac870aec7321"} Nov 21 14:15:35 crc kubenswrapper[4897]: I1121 14:15:35.483044 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-95c6d8978-l44nd" event={"ID":"bb7f0923-4b62-4114-8119-34c39e469323","Type":"ContainerStarted","Data":"a4c7a342595f49d237da49e5fa18b725a5e1b6dcd7146c34cb6d10ed016b44d6"} Nov 21 14:15:35 crc kubenswrapper[4897]: I1121 14:15:35.502831 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-95c6d8978-l44nd" podStartSLOduration=1.502812217 podStartE2EDuration="1.502812217s" podCreationTimestamp="2025-11-21 14:15:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:15:35.502043757 +0000 UTC m=+412.786637242" watchObservedRunningTime="2025-11-21 14:15:35.502812217 +0000 UTC m=+412.787405692" Nov 21 14:15:44 crc kubenswrapper[4897]: I1121 14:15:44.998357 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-95c6d8978-l44nd" Nov 21 14:15:45 crc kubenswrapper[4897]: I1121 14:15:44.999062 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-95c6d8978-l44nd" Nov 21 14:15:45 crc kubenswrapper[4897]: I1121 14:15:45.006894 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-95c6d8978-l44nd" Nov 21 14:15:45 crc kubenswrapper[4897]: I1121 14:15:45.550579 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-95c6d8978-l44nd" Nov 21 14:15:45 crc kubenswrapper[4897]: I1121 14:15:45.650649 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-5f888bf76d-vtk2t"] Nov 21 14:16:04 crc kubenswrapper[4897]: I1121 14:16:04.370831 4897 patch_prober.go:28] interesting pod/machine-config-daemon-krv5b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 21 14:16:04 crc kubenswrapper[4897]: I1121 14:16:04.371944 4897 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 21 14:16:04 crc kubenswrapper[4897]: I1121 14:16:04.372266 4897 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" Nov 21 14:16:04 crc kubenswrapper[4897]: I1121 14:16:04.375775 4897 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d1709941d05f61c782c0e1fa4c44808bcba5b532f6ab33b1f154eb915fe3841b"} pod="openshift-machine-config-operator/machine-config-daemon-krv5b" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 21 14:16:04 crc kubenswrapper[4897]: I1121 14:16:04.376047 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" containerName="machine-config-daemon" containerID="cri-o://d1709941d05f61c782c0e1fa4c44808bcba5b532f6ab33b1f154eb915fe3841b" gracePeriod=600 Nov 21 14:16:04 crc kubenswrapper[4897]: I1121 14:16:04.682180 4897 generic.go:334] "Generic (PLEG): container finished" podID="e7670227-d280-4847-b882-754429f56b0f" containerID="d1709941d05f61c782c0e1fa4c44808bcba5b532f6ab33b1f154eb915fe3841b" exitCode=0 Nov 21 14:16:04 crc kubenswrapper[4897]: I1121 14:16:04.682249 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" event={"ID":"e7670227-d280-4847-b882-754429f56b0f","Type":"ContainerDied","Data":"d1709941d05f61c782c0e1fa4c44808bcba5b532f6ab33b1f154eb915fe3841b"} Nov 21 14:16:04 crc kubenswrapper[4897]: I1121 14:16:04.682584 4897 scope.go:117] "RemoveContainer" containerID="5854028821e6a360a259427b68e7bb4c2a7220e0c68f4cc91639b2ece0540a02" Nov 21 14:16:05 crc kubenswrapper[4897]: I1121 14:16:05.693061 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" event={"ID":"e7670227-d280-4847-b882-754429f56b0f","Type":"ContainerStarted","Data":"3ef2942b0389d06ca20fad79231df545f174b1eeae70986340588271efcd18e6"} Nov 21 14:16:10 crc kubenswrapper[4897]: I1121 14:16:10.692544 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-5f888bf76d-vtk2t" podUID="062ef123-84be-49c9-9bb2-8fc1a415f431" containerName="console" containerID="cri-o://4b99cdf2d4a2a8555e1bf29923a5f96420c74764aff16d0be7205a25ad8329eb" gracePeriod=15 Nov 21 14:16:11 crc kubenswrapper[4897]: I1121 14:16:11.088968 4897 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-5f888bf76d-vtk2t_062ef123-84be-49c9-9bb2-8fc1a415f431/console/0.log" Nov 21 14:16:11 crc kubenswrapper[4897]: I1121 14:16:11.089381 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5f888bf76d-vtk2t" Nov 21 14:16:11 crc kubenswrapper[4897]: I1121 14:16:11.213288 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dwqrt\" (UniqueName: \"kubernetes.io/projected/062ef123-84be-49c9-9bb2-8fc1a415f431-kube-api-access-dwqrt\") pod \"062ef123-84be-49c9-9bb2-8fc1a415f431\" (UID: \"062ef123-84be-49c9-9bb2-8fc1a415f431\") " Nov 21 14:16:11 crc kubenswrapper[4897]: I1121 14:16:11.213432 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/062ef123-84be-49c9-9bb2-8fc1a415f431-oauth-serving-cert\") pod \"062ef123-84be-49c9-9bb2-8fc1a415f431\" (UID: \"062ef123-84be-49c9-9bb2-8fc1a415f431\") " Nov 21 14:16:11 crc kubenswrapper[4897]: I1121 14:16:11.213545 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/062ef123-84be-49c9-9bb2-8fc1a415f431-trusted-ca-bundle\") pod \"062ef123-84be-49c9-9bb2-8fc1a415f431\" (UID: \"062ef123-84be-49c9-9bb2-8fc1a415f431\") " Nov 21 14:16:11 crc kubenswrapper[4897]: I1121 14:16:11.213641 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/062ef123-84be-49c9-9bb2-8fc1a415f431-service-ca\") pod \"062ef123-84be-49c9-9bb2-8fc1a415f431\" (UID: \"062ef123-84be-49c9-9bb2-8fc1a415f431\") " Nov 21 14:16:11 crc kubenswrapper[4897]: I1121 14:16:11.213689 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/062ef123-84be-49c9-9bb2-8fc1a415f431-console-oauth-config\") pod \"062ef123-84be-49c9-9bb2-8fc1a415f431\" (UID: \"062ef123-84be-49c9-9bb2-8fc1a415f431\") " Nov 21 14:16:11 crc kubenswrapper[4897]: I1121 14:16:11.213973 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/062ef123-84be-49c9-9bb2-8fc1a415f431-console-config\") pod \"062ef123-84be-49c9-9bb2-8fc1a415f431\" (UID: \"062ef123-84be-49c9-9bb2-8fc1a415f431\") " Nov 21 14:16:11 crc kubenswrapper[4897]: I1121 14:16:11.214142 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/062ef123-84be-49c9-9bb2-8fc1a415f431-console-serving-cert\") pod \"062ef123-84be-49c9-9bb2-8fc1a415f431\" (UID: \"062ef123-84be-49c9-9bb2-8fc1a415f431\") " Nov 21 14:16:11 crc kubenswrapper[4897]: I1121 14:16:11.214253 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/062ef123-84be-49c9-9bb2-8fc1a415f431-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "062ef123-84be-49c9-9bb2-8fc1a415f431" (UID: "062ef123-84be-49c9-9bb2-8fc1a415f431"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:16:11 crc kubenswrapper[4897]: I1121 14:16:11.214717 4897 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/062ef123-84be-49c9-9bb2-8fc1a415f431-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 21 14:16:11 crc kubenswrapper[4897]: I1121 14:16:11.214885 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/062ef123-84be-49c9-9bb2-8fc1a415f431-console-config" (OuterVolumeSpecName: "console-config") pod "062ef123-84be-49c9-9bb2-8fc1a415f431" (UID: "062ef123-84be-49c9-9bb2-8fc1a415f431"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:16:11 crc kubenswrapper[4897]: I1121 14:16:11.214981 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/062ef123-84be-49c9-9bb2-8fc1a415f431-service-ca" (OuterVolumeSpecName: "service-ca") pod "062ef123-84be-49c9-9bb2-8fc1a415f431" (UID: "062ef123-84be-49c9-9bb2-8fc1a415f431"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:16:11 crc kubenswrapper[4897]: I1121 14:16:11.215186 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/062ef123-84be-49c9-9bb2-8fc1a415f431-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "062ef123-84be-49c9-9bb2-8fc1a415f431" (UID: "062ef123-84be-49c9-9bb2-8fc1a415f431"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:16:11 crc kubenswrapper[4897]: I1121 14:16:11.222075 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/062ef123-84be-49c9-9bb2-8fc1a415f431-kube-api-access-dwqrt" (OuterVolumeSpecName: "kube-api-access-dwqrt") pod "062ef123-84be-49c9-9bb2-8fc1a415f431" (UID: "062ef123-84be-49c9-9bb2-8fc1a415f431"). InnerVolumeSpecName "kube-api-access-dwqrt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:16:11 crc kubenswrapper[4897]: I1121 14:16:11.222665 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/062ef123-84be-49c9-9bb2-8fc1a415f431-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "062ef123-84be-49c9-9bb2-8fc1a415f431" (UID: "062ef123-84be-49c9-9bb2-8fc1a415f431"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:16:11 crc kubenswrapper[4897]: I1121 14:16:11.222727 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/062ef123-84be-49c9-9bb2-8fc1a415f431-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "062ef123-84be-49c9-9bb2-8fc1a415f431" (UID: "062ef123-84be-49c9-9bb2-8fc1a415f431"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:16:11 crc kubenswrapper[4897]: I1121 14:16:11.316819 4897 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/062ef123-84be-49c9-9bb2-8fc1a415f431-service-ca\") on node \"crc\" DevicePath \"\"" Nov 21 14:16:11 crc kubenswrapper[4897]: I1121 14:16:11.316890 4897 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/062ef123-84be-49c9-9bb2-8fc1a415f431-console-oauth-config\") on node \"crc\" DevicePath \"\"" Nov 21 14:16:11 crc kubenswrapper[4897]: I1121 14:16:11.316922 4897 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/062ef123-84be-49c9-9bb2-8fc1a415f431-console-config\") on node \"crc\" DevicePath \"\"" Nov 21 14:16:11 crc kubenswrapper[4897]: I1121 14:16:11.316950 4897 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/062ef123-84be-49c9-9bb2-8fc1a415f431-console-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 21 14:16:11 crc kubenswrapper[4897]: I1121 14:16:11.316980 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dwqrt\" (UniqueName: \"kubernetes.io/projected/062ef123-84be-49c9-9bb2-8fc1a415f431-kube-api-access-dwqrt\") on node \"crc\" DevicePath \"\"" Nov 21 14:16:11 crc kubenswrapper[4897]: I1121 14:16:11.317007 4897 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/062ef123-84be-49c9-9bb2-8fc1a415f431-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 21 14:16:11 crc kubenswrapper[4897]: I1121 14:16:11.737176 4897 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-5f888bf76d-vtk2t_062ef123-84be-49c9-9bb2-8fc1a415f431/console/0.log" Nov 21 14:16:11 crc kubenswrapper[4897]: I1121 14:16:11.737226 4897 generic.go:334] "Generic (PLEG): container finished" podID="062ef123-84be-49c9-9bb2-8fc1a415f431" containerID="4b99cdf2d4a2a8555e1bf29923a5f96420c74764aff16d0be7205a25ad8329eb" exitCode=2 Nov 21 14:16:11 crc kubenswrapper[4897]: I1121 14:16:11.737255 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5f888bf76d-vtk2t" event={"ID":"062ef123-84be-49c9-9bb2-8fc1a415f431","Type":"ContainerDied","Data":"4b99cdf2d4a2a8555e1bf29923a5f96420c74764aff16d0be7205a25ad8329eb"} Nov 21 14:16:11 crc kubenswrapper[4897]: I1121 14:16:11.737281 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5f888bf76d-vtk2t" event={"ID":"062ef123-84be-49c9-9bb2-8fc1a415f431","Type":"ContainerDied","Data":"17fda8582808ef239a7a28a430598982a5e5ace3de7b3fdd4b56a309d603211f"} Nov 21 14:16:11 crc kubenswrapper[4897]: I1121 14:16:11.737298 4897 scope.go:117] "RemoveContainer" containerID="4b99cdf2d4a2a8555e1bf29923a5f96420c74764aff16d0be7205a25ad8329eb" Nov 21 14:16:11 crc kubenswrapper[4897]: I1121 14:16:11.737341 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5f888bf76d-vtk2t" Nov 21 14:16:11 crc kubenswrapper[4897]: I1121 14:16:11.756615 4897 scope.go:117] "RemoveContainer" containerID="4b99cdf2d4a2a8555e1bf29923a5f96420c74764aff16d0be7205a25ad8329eb" Nov 21 14:16:11 crc kubenswrapper[4897]: E1121 14:16:11.757051 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4b99cdf2d4a2a8555e1bf29923a5f96420c74764aff16d0be7205a25ad8329eb\": container with ID starting with 4b99cdf2d4a2a8555e1bf29923a5f96420c74764aff16d0be7205a25ad8329eb not found: ID does not exist" containerID="4b99cdf2d4a2a8555e1bf29923a5f96420c74764aff16d0be7205a25ad8329eb" Nov 21 14:16:11 crc kubenswrapper[4897]: I1121 14:16:11.757082 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4b99cdf2d4a2a8555e1bf29923a5f96420c74764aff16d0be7205a25ad8329eb"} err="failed to get container status \"4b99cdf2d4a2a8555e1bf29923a5f96420c74764aff16d0be7205a25ad8329eb\": rpc error: code = NotFound desc = could not find container \"4b99cdf2d4a2a8555e1bf29923a5f96420c74764aff16d0be7205a25ad8329eb\": container with ID starting with 4b99cdf2d4a2a8555e1bf29923a5f96420c74764aff16d0be7205a25ad8329eb not found: ID does not exist" Nov 21 14:16:11 crc kubenswrapper[4897]: I1121 14:16:11.779780 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-5f888bf76d-vtk2t"] Nov 21 14:16:11 crc kubenswrapper[4897]: I1121 14:16:11.786234 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-5f888bf76d-vtk2t"] Nov 21 14:16:12 crc kubenswrapper[4897]: I1121 14:16:12.096084 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="062ef123-84be-49c9-9bb2-8fc1a415f431" path="/var/lib/kubelet/pods/062ef123-84be-49c9-9bb2-8fc1a415f431/volumes" Nov 21 14:17:44 crc kubenswrapper[4897]: I1121 14:17:44.356768 4897 scope.go:117] "RemoveContainer" containerID="b9c27b68540fcadb0a431d5f21d557435c273b2b80b22ddfe779d421a63947ca" Nov 21 14:17:44 crc kubenswrapper[4897]: I1121 14:17:44.380386 4897 scope.go:117] "RemoveContainer" containerID="79ade4cd758c4fb4a6d2ad5e6f0ed42b487b6acd04c14586c067c870f5026815" Nov 21 14:18:04 crc kubenswrapper[4897]: I1121 14:18:04.370752 4897 patch_prober.go:28] interesting pod/machine-config-daemon-krv5b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 21 14:18:04 crc kubenswrapper[4897]: I1121 14:18:04.371372 4897 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 21 14:18:34 crc kubenswrapper[4897]: I1121 14:18:34.370866 4897 patch_prober.go:28] interesting pod/machine-config-daemon-krv5b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 21 14:18:34 crc kubenswrapper[4897]: I1121 14:18:34.371498 4897 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 21 14:18:44 crc kubenswrapper[4897]: I1121 14:18:44.446376 4897 scope.go:117] "RemoveContainer" containerID="3c4b45b28a007a59596ee574c1cced8c1fe7d9eb0b178c0ca7e9f28ee9f3b830" Nov 21 14:18:44 crc kubenswrapper[4897]: I1121 14:18:44.488843 4897 scope.go:117] "RemoveContainer" containerID="57daea7e609a9d7bca2ddc01887c447f5a0488528509239c6a88833905c3e474" Nov 21 14:19:04 crc kubenswrapper[4897]: I1121 14:19:04.370675 4897 patch_prober.go:28] interesting pod/machine-config-daemon-krv5b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 21 14:19:04 crc kubenswrapper[4897]: I1121 14:19:04.371438 4897 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 21 14:19:04 crc kubenswrapper[4897]: I1121 14:19:04.371900 4897 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" Nov 21 14:19:04 crc kubenswrapper[4897]: I1121 14:19:04.373651 4897 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3ef2942b0389d06ca20fad79231df545f174b1eeae70986340588271efcd18e6"} pod="openshift-machine-config-operator/machine-config-daemon-krv5b" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 21 14:19:04 crc kubenswrapper[4897]: I1121 14:19:04.374060 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" containerName="machine-config-daemon" containerID="cri-o://3ef2942b0389d06ca20fad79231df545f174b1eeae70986340588271efcd18e6" gracePeriod=600 Nov 21 14:19:04 crc kubenswrapper[4897]: I1121 14:19:04.935243 4897 generic.go:334] "Generic (PLEG): container finished" podID="e7670227-d280-4847-b882-754429f56b0f" containerID="3ef2942b0389d06ca20fad79231df545f174b1eeae70986340588271efcd18e6" exitCode=0 Nov 21 14:19:04 crc kubenswrapper[4897]: I1121 14:19:04.935372 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" event={"ID":"e7670227-d280-4847-b882-754429f56b0f","Type":"ContainerDied","Data":"3ef2942b0389d06ca20fad79231df545f174b1eeae70986340588271efcd18e6"} Nov 21 14:19:04 crc kubenswrapper[4897]: I1121 14:19:04.936003 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" event={"ID":"e7670227-d280-4847-b882-754429f56b0f","Type":"ContainerStarted","Data":"cad6f00fa42dcd6129303195cd6590d896370bbb75a53d50c3950cca8fb66efa"} Nov 21 14:19:04 crc kubenswrapper[4897]: I1121 14:19:04.936077 4897 scope.go:117] "RemoveContainer" containerID="d1709941d05f61c782c0e1fa4c44808bcba5b532f6ab33b1f154eb915fe3841b" Nov 21 14:19:20 crc kubenswrapper[4897]: I1121 14:19:20.820489 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zdcxt"] Nov 21 14:19:20 crc kubenswrapper[4897]: E1121 14:19:20.821338 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="062ef123-84be-49c9-9bb2-8fc1a415f431" containerName="console" Nov 21 14:19:20 crc kubenswrapper[4897]: I1121 14:19:20.821354 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="062ef123-84be-49c9-9bb2-8fc1a415f431" containerName="console" Nov 21 14:19:20 crc kubenswrapper[4897]: I1121 14:19:20.821497 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="062ef123-84be-49c9-9bb2-8fc1a415f431" containerName="console" Nov 21 14:19:20 crc kubenswrapper[4897]: I1121 14:19:20.822646 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zdcxt" Nov 21 14:19:20 crc kubenswrapper[4897]: I1121 14:19:20.825110 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 21 14:19:20 crc kubenswrapper[4897]: I1121 14:19:20.828901 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zdcxt"] Nov 21 14:19:20 crc kubenswrapper[4897]: I1121 14:19:20.955047 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ldx6x\" (UniqueName: \"kubernetes.io/projected/c792eeb7-8d0e-44fc-a600-fe5c6e4f997e-kube-api-access-ldx6x\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zdcxt\" (UID: \"c792eeb7-8d0e-44fc-a600-fe5c6e4f997e\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zdcxt" Nov 21 14:19:20 crc kubenswrapper[4897]: I1121 14:19:20.955109 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c792eeb7-8d0e-44fc-a600-fe5c6e4f997e-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zdcxt\" (UID: \"c792eeb7-8d0e-44fc-a600-fe5c6e4f997e\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zdcxt" Nov 21 14:19:20 crc kubenswrapper[4897]: I1121 14:19:20.955137 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c792eeb7-8d0e-44fc-a600-fe5c6e4f997e-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zdcxt\" (UID: \"c792eeb7-8d0e-44fc-a600-fe5c6e4f997e\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zdcxt" Nov 21 14:19:21 crc kubenswrapper[4897]: I1121 14:19:21.056005 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ldx6x\" (UniqueName: \"kubernetes.io/projected/c792eeb7-8d0e-44fc-a600-fe5c6e4f997e-kube-api-access-ldx6x\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zdcxt\" (UID: \"c792eeb7-8d0e-44fc-a600-fe5c6e4f997e\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zdcxt" Nov 21 14:19:21 crc kubenswrapper[4897]: I1121 14:19:21.056088 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c792eeb7-8d0e-44fc-a600-fe5c6e4f997e-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zdcxt\" (UID: \"c792eeb7-8d0e-44fc-a600-fe5c6e4f997e\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zdcxt" Nov 21 14:19:21 crc kubenswrapper[4897]: I1121 14:19:21.056125 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c792eeb7-8d0e-44fc-a600-fe5c6e4f997e-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zdcxt\" (UID: \"c792eeb7-8d0e-44fc-a600-fe5c6e4f997e\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zdcxt" Nov 21 14:19:21 crc kubenswrapper[4897]: I1121 14:19:21.056883 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c792eeb7-8d0e-44fc-a600-fe5c6e4f997e-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zdcxt\" (UID: \"c792eeb7-8d0e-44fc-a600-fe5c6e4f997e\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zdcxt" Nov 21 14:19:21 crc kubenswrapper[4897]: I1121 14:19:21.056915 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c792eeb7-8d0e-44fc-a600-fe5c6e4f997e-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zdcxt\" (UID: \"c792eeb7-8d0e-44fc-a600-fe5c6e4f997e\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zdcxt" Nov 21 14:19:21 crc kubenswrapper[4897]: I1121 14:19:21.083092 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ldx6x\" (UniqueName: \"kubernetes.io/projected/c792eeb7-8d0e-44fc-a600-fe5c6e4f997e-kube-api-access-ldx6x\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zdcxt\" (UID: \"c792eeb7-8d0e-44fc-a600-fe5c6e4f997e\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zdcxt" Nov 21 14:19:21 crc kubenswrapper[4897]: I1121 14:19:21.155561 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zdcxt" Nov 21 14:19:21 crc kubenswrapper[4897]: I1121 14:19:21.363269 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zdcxt"] Nov 21 14:19:21 crc kubenswrapper[4897]: W1121 14:19:21.371422 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc792eeb7_8d0e_44fc_a600_fe5c6e4f997e.slice/crio-d372497ff69e5418b0c5281551bf8d5e877a141cbd763cf786688805ee396a0b WatchSource:0}: Error finding container d372497ff69e5418b0c5281551bf8d5e877a141cbd763cf786688805ee396a0b: Status 404 returned error can't find the container with id d372497ff69e5418b0c5281551bf8d5e877a141cbd763cf786688805ee396a0b Nov 21 14:19:22 crc kubenswrapper[4897]: I1121 14:19:22.054478 4897 generic.go:334] "Generic (PLEG): container finished" podID="c792eeb7-8d0e-44fc-a600-fe5c6e4f997e" containerID="0a3a16b5ee1fc0ff531c6431c3d518cf8e70b3c771edd0416320e71a23c40ffc" exitCode=0 Nov 21 14:19:22 crc kubenswrapper[4897]: I1121 14:19:22.054704 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zdcxt" event={"ID":"c792eeb7-8d0e-44fc-a600-fe5c6e4f997e","Type":"ContainerDied","Data":"0a3a16b5ee1fc0ff531c6431c3d518cf8e70b3c771edd0416320e71a23c40ffc"} Nov 21 14:19:22 crc kubenswrapper[4897]: I1121 14:19:22.055096 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zdcxt" event={"ID":"c792eeb7-8d0e-44fc-a600-fe5c6e4f997e","Type":"ContainerStarted","Data":"d372497ff69e5418b0c5281551bf8d5e877a141cbd763cf786688805ee396a0b"} Nov 21 14:19:22 crc kubenswrapper[4897]: I1121 14:19:22.058009 4897 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 21 14:19:24 crc kubenswrapper[4897]: I1121 14:19:24.069155 4897 generic.go:334] "Generic (PLEG): container finished" podID="c792eeb7-8d0e-44fc-a600-fe5c6e4f997e" containerID="102d26391d47294f45693fdf4a3c87a685aa67adc325a6b2363ea6c4b8559901" exitCode=0 Nov 21 14:19:24 crc kubenswrapper[4897]: I1121 14:19:24.069204 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zdcxt" event={"ID":"c792eeb7-8d0e-44fc-a600-fe5c6e4f997e","Type":"ContainerDied","Data":"102d26391d47294f45693fdf4a3c87a685aa67adc325a6b2363ea6c4b8559901"} Nov 21 14:19:25 crc kubenswrapper[4897]: I1121 14:19:25.077973 4897 generic.go:334] "Generic (PLEG): container finished" podID="c792eeb7-8d0e-44fc-a600-fe5c6e4f997e" containerID="73935550fb211d56fd6d7a10455a93d1024f1a0c9777f7a3f1b655c8f0f1da9a" exitCode=0 Nov 21 14:19:25 crc kubenswrapper[4897]: I1121 14:19:25.078018 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zdcxt" event={"ID":"c792eeb7-8d0e-44fc-a600-fe5c6e4f997e","Type":"ContainerDied","Data":"73935550fb211d56fd6d7a10455a93d1024f1a0c9777f7a3f1b655c8f0f1da9a"} Nov 21 14:19:26 crc kubenswrapper[4897]: I1121 14:19:26.340439 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zdcxt" Nov 21 14:19:26 crc kubenswrapper[4897]: I1121 14:19:26.427768 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ldx6x\" (UniqueName: \"kubernetes.io/projected/c792eeb7-8d0e-44fc-a600-fe5c6e4f997e-kube-api-access-ldx6x\") pod \"c792eeb7-8d0e-44fc-a600-fe5c6e4f997e\" (UID: \"c792eeb7-8d0e-44fc-a600-fe5c6e4f997e\") " Nov 21 14:19:26 crc kubenswrapper[4897]: I1121 14:19:26.427835 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c792eeb7-8d0e-44fc-a600-fe5c6e4f997e-util\") pod \"c792eeb7-8d0e-44fc-a600-fe5c6e4f997e\" (UID: \"c792eeb7-8d0e-44fc-a600-fe5c6e4f997e\") " Nov 21 14:19:26 crc kubenswrapper[4897]: I1121 14:19:26.427949 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c792eeb7-8d0e-44fc-a600-fe5c6e4f997e-bundle\") pod \"c792eeb7-8d0e-44fc-a600-fe5c6e4f997e\" (UID: \"c792eeb7-8d0e-44fc-a600-fe5c6e4f997e\") " Nov 21 14:19:26 crc kubenswrapper[4897]: I1121 14:19:26.430627 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c792eeb7-8d0e-44fc-a600-fe5c6e4f997e-bundle" (OuterVolumeSpecName: "bundle") pod "c792eeb7-8d0e-44fc-a600-fe5c6e4f997e" (UID: "c792eeb7-8d0e-44fc-a600-fe5c6e4f997e"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:19:26 crc kubenswrapper[4897]: I1121 14:19:26.434801 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c792eeb7-8d0e-44fc-a600-fe5c6e4f997e-kube-api-access-ldx6x" (OuterVolumeSpecName: "kube-api-access-ldx6x") pod "c792eeb7-8d0e-44fc-a600-fe5c6e4f997e" (UID: "c792eeb7-8d0e-44fc-a600-fe5c6e4f997e"). InnerVolumeSpecName "kube-api-access-ldx6x". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:19:26 crc kubenswrapper[4897]: I1121 14:19:26.446947 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c792eeb7-8d0e-44fc-a600-fe5c6e4f997e-util" (OuterVolumeSpecName: "util") pod "c792eeb7-8d0e-44fc-a600-fe5c6e4f997e" (UID: "c792eeb7-8d0e-44fc-a600-fe5c6e4f997e"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:19:26 crc kubenswrapper[4897]: I1121 14:19:26.529307 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ldx6x\" (UniqueName: \"kubernetes.io/projected/c792eeb7-8d0e-44fc-a600-fe5c6e4f997e-kube-api-access-ldx6x\") on node \"crc\" DevicePath \"\"" Nov 21 14:19:26 crc kubenswrapper[4897]: I1121 14:19:26.529378 4897 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c792eeb7-8d0e-44fc-a600-fe5c6e4f997e-util\") on node \"crc\" DevicePath \"\"" Nov 21 14:19:26 crc kubenswrapper[4897]: I1121 14:19:26.529392 4897 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c792eeb7-8d0e-44fc-a600-fe5c6e4f997e-bundle\") on node \"crc\" DevicePath \"\"" Nov 21 14:19:27 crc kubenswrapper[4897]: I1121 14:19:27.092638 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zdcxt" event={"ID":"c792eeb7-8d0e-44fc-a600-fe5c6e4f997e","Type":"ContainerDied","Data":"d372497ff69e5418b0c5281551bf8d5e877a141cbd763cf786688805ee396a0b"} Nov 21 14:19:27 crc kubenswrapper[4897]: I1121 14:19:27.092683 4897 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d372497ff69e5418b0c5281551bf8d5e877a141cbd763cf786688805ee396a0b" Nov 21 14:19:27 crc kubenswrapper[4897]: I1121 14:19:27.092736 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zdcxt" Nov 21 14:19:32 crc kubenswrapper[4897]: I1121 14:19:32.285152 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-tpfb6"] Nov 21 14:19:32 crc kubenswrapper[4897]: I1121 14:19:32.285987 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" podUID="222f65f2-0b82-4760-bcf7-779244720f01" containerName="ovn-controller" containerID="cri-o://9de095f3c760b13d5197f624eac3806a1f4186577f2b95eb522532d453eb266e" gracePeriod=30 Nov 21 14:19:32 crc kubenswrapper[4897]: I1121 14:19:32.286090 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" podUID="222f65f2-0b82-4760-bcf7-779244720f01" containerName="ovn-acl-logging" containerID="cri-o://ec279f2f558a8ac653006321b18cbaf13185876fe303c5cc1c65988aef09edbf" gracePeriod=30 Nov 21 14:19:32 crc kubenswrapper[4897]: I1121 14:19:32.286129 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" podUID="222f65f2-0b82-4760-bcf7-779244720f01" containerName="northd" containerID="cri-o://cfee19f82e35869a641318a0c45bc7a2ae71d80039f5790e60e277be15c9d3ec" gracePeriod=30 Nov 21 14:19:32 crc kubenswrapper[4897]: I1121 14:19:32.286160 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" podUID="222f65f2-0b82-4760-bcf7-779244720f01" containerName="sbdb" containerID="cri-o://b486a86e0e97ab409e590f30a1bf5cb9f584a7c67f96ff1236e42ca4f8675b56" gracePeriod=30 Nov 21 14:19:32 crc kubenswrapper[4897]: I1121 14:19:32.286156 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" podUID="222f65f2-0b82-4760-bcf7-779244720f01" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://ccd51aa0faf3b1327f207b355cddf86a7353c0051f7e5ce0396f3c4bc67e3b16" gracePeriod=30 Nov 21 14:19:32 crc kubenswrapper[4897]: I1121 14:19:32.286076 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" podUID="222f65f2-0b82-4760-bcf7-779244720f01" containerName="nbdb" containerID="cri-o://8971c38daad45bda3d61545f52120dad40992dbf2e1d6e1e776500ed58d73f36" gracePeriod=30 Nov 21 14:19:32 crc kubenswrapper[4897]: I1121 14:19:32.286565 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" podUID="222f65f2-0b82-4760-bcf7-779244720f01" containerName="kube-rbac-proxy-node" containerID="cri-o://92f1d932a83d65ba6a9a906cc489ff51e76547d07d42656630c59421f833c4db" gracePeriod=30 Nov 21 14:19:32 crc kubenswrapper[4897]: I1121 14:19:32.361697 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" podUID="222f65f2-0b82-4760-bcf7-779244720f01" containerName="ovnkube-controller" containerID="cri-o://eaaa7662c416148afbfc746cc7b87fac845ce54740988de2c6d0af4b1a21b9e7" gracePeriod=30 Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.136035 4897 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-6ljgx_5f3b70b5-9294-4f69-9528-500d28f34c89/kube-multus/2.log" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.136773 4897 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-6ljgx_5f3b70b5-9294-4f69-9528-500d28f34c89/kube-multus/1.log" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.136819 4897 generic.go:334] "Generic (PLEG): container finished" podID="5f3b70b5-9294-4f69-9528-500d28f34c89" containerID="382da75081f6e2afec6e1492c40398e4b099071e26f7162a289311aedd0d7777" exitCode=2 Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.136881 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-6ljgx" event={"ID":"5f3b70b5-9294-4f69-9528-500d28f34c89","Type":"ContainerDied","Data":"382da75081f6e2afec6e1492c40398e4b099071e26f7162a289311aedd0d7777"} Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.136922 4897 scope.go:117] "RemoveContainer" containerID="bf92ab70c5a3196bf83f427ccf03acf188d1995766a9154bcd6267e3926c3ac8" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.137444 4897 scope.go:117] "RemoveContainer" containerID="382da75081f6e2afec6e1492c40398e4b099071e26f7162a289311aedd0d7777" Nov 21 14:19:33 crc kubenswrapper[4897]: E1121 14:19:33.137639 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-6ljgx_openshift-multus(5f3b70b5-9294-4f69-9528-500d28f34c89)\"" pod="openshift-multus/multus-6ljgx" podUID="5f3b70b5-9294-4f69-9528-500d28f34c89" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.139132 4897 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tpfb6_222f65f2-0b82-4760-bcf7-779244720f01/ovnkube-controller/3.log" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.141107 4897 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tpfb6_222f65f2-0b82-4760-bcf7-779244720f01/ovn-acl-logging/0.log" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.141642 4897 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tpfb6_222f65f2-0b82-4760-bcf7-779244720f01/ovn-controller/0.log" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.142005 4897 generic.go:334] "Generic (PLEG): container finished" podID="222f65f2-0b82-4760-bcf7-779244720f01" containerID="eaaa7662c416148afbfc746cc7b87fac845ce54740988de2c6d0af4b1a21b9e7" exitCode=0 Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.142029 4897 generic.go:334] "Generic (PLEG): container finished" podID="222f65f2-0b82-4760-bcf7-779244720f01" containerID="b486a86e0e97ab409e590f30a1bf5cb9f584a7c67f96ff1236e42ca4f8675b56" exitCode=0 Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.142037 4897 generic.go:334] "Generic (PLEG): container finished" podID="222f65f2-0b82-4760-bcf7-779244720f01" containerID="8971c38daad45bda3d61545f52120dad40992dbf2e1d6e1e776500ed58d73f36" exitCode=0 Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.142033 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" event={"ID":"222f65f2-0b82-4760-bcf7-779244720f01","Type":"ContainerDied","Data":"eaaa7662c416148afbfc746cc7b87fac845ce54740988de2c6d0af4b1a21b9e7"} Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.142067 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" event={"ID":"222f65f2-0b82-4760-bcf7-779244720f01","Type":"ContainerDied","Data":"b486a86e0e97ab409e590f30a1bf5cb9f584a7c67f96ff1236e42ca4f8675b56"} Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.142082 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" event={"ID":"222f65f2-0b82-4760-bcf7-779244720f01","Type":"ContainerDied","Data":"8971c38daad45bda3d61545f52120dad40992dbf2e1d6e1e776500ed58d73f36"} Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.142094 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" event={"ID":"222f65f2-0b82-4760-bcf7-779244720f01","Type":"ContainerDied","Data":"cfee19f82e35869a641318a0c45bc7a2ae71d80039f5790e60e277be15c9d3ec"} Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.142045 4897 generic.go:334] "Generic (PLEG): container finished" podID="222f65f2-0b82-4760-bcf7-779244720f01" containerID="cfee19f82e35869a641318a0c45bc7a2ae71d80039f5790e60e277be15c9d3ec" exitCode=0 Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.142117 4897 generic.go:334] "Generic (PLEG): container finished" podID="222f65f2-0b82-4760-bcf7-779244720f01" containerID="ec279f2f558a8ac653006321b18cbaf13185876fe303c5cc1c65988aef09edbf" exitCode=143 Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.142130 4897 generic.go:334] "Generic (PLEG): container finished" podID="222f65f2-0b82-4760-bcf7-779244720f01" containerID="9de095f3c760b13d5197f624eac3806a1f4186577f2b95eb522532d453eb266e" exitCode=143 Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.142145 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" event={"ID":"222f65f2-0b82-4760-bcf7-779244720f01","Type":"ContainerDied","Data":"ec279f2f558a8ac653006321b18cbaf13185876fe303c5cc1c65988aef09edbf"} Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.142157 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" event={"ID":"222f65f2-0b82-4760-bcf7-779244720f01","Type":"ContainerDied","Data":"9de095f3c760b13d5197f624eac3806a1f4186577f2b95eb522532d453eb266e"} Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.167827 4897 scope.go:117] "RemoveContainer" containerID="40b0bddce48b3626ad889ae2f43deb866acd521a8aab71253506364bfb47c723" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.561451 4897 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tpfb6_222f65f2-0b82-4760-bcf7-779244720f01/ovn-acl-logging/0.log" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.561970 4897 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tpfb6_222f65f2-0b82-4760-bcf7-779244720f01/ovn-controller/0.log" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.562368 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.625317 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/222f65f2-0b82-4760-bcf7-779244720f01-var-lib-openvswitch\") pod \"222f65f2-0b82-4760-bcf7-779244720f01\" (UID: \"222f65f2-0b82-4760-bcf7-779244720f01\") " Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.625357 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/222f65f2-0b82-4760-bcf7-779244720f01-host-run-netns\") pod \"222f65f2-0b82-4760-bcf7-779244720f01\" (UID: \"222f65f2-0b82-4760-bcf7-779244720f01\") " Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.625384 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/222f65f2-0b82-4760-bcf7-779244720f01-host-slash\") pod \"222f65f2-0b82-4760-bcf7-779244720f01\" (UID: \"222f65f2-0b82-4760-bcf7-779244720f01\") " Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.625402 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/222f65f2-0b82-4760-bcf7-779244720f01-host-cni-netd\") pod \"222f65f2-0b82-4760-bcf7-779244720f01\" (UID: \"222f65f2-0b82-4760-bcf7-779244720f01\") " Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.625445 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/222f65f2-0b82-4760-bcf7-779244720f01-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "222f65f2-0b82-4760-bcf7-779244720f01" (UID: "222f65f2-0b82-4760-bcf7-779244720f01"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.625467 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/222f65f2-0b82-4760-bcf7-779244720f01-host-slash" (OuterVolumeSpecName: "host-slash") pod "222f65f2-0b82-4760-bcf7-779244720f01" (UID: "222f65f2-0b82-4760-bcf7-779244720f01"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.625416 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/222f65f2-0b82-4760-bcf7-779244720f01-host-kubelet\") pod \"222f65f2-0b82-4760-bcf7-779244720f01\" (UID: \"222f65f2-0b82-4760-bcf7-779244720f01\") " Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.625532 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/222f65f2-0b82-4760-bcf7-779244720f01-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "222f65f2-0b82-4760-bcf7-779244720f01" (UID: "222f65f2-0b82-4760-bcf7-779244720f01"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.625537 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/222f65f2-0b82-4760-bcf7-779244720f01-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "222f65f2-0b82-4760-bcf7-779244720f01" (UID: "222f65f2-0b82-4760-bcf7-779244720f01"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.625590 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/222f65f2-0b82-4760-bcf7-779244720f01-node-log\") pod \"222f65f2-0b82-4760-bcf7-779244720f01\" (UID: \"222f65f2-0b82-4760-bcf7-779244720f01\") " Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.625618 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/222f65f2-0b82-4760-bcf7-779244720f01-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "222f65f2-0b82-4760-bcf7-779244720f01" (UID: "222f65f2-0b82-4760-bcf7-779244720f01"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.625627 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/222f65f2-0b82-4760-bcf7-779244720f01-ovnkube-config\") pod \"222f65f2-0b82-4760-bcf7-779244720f01\" (UID: \"222f65f2-0b82-4760-bcf7-779244720f01\") " Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.625649 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/222f65f2-0b82-4760-bcf7-779244720f01-ovn-node-metrics-cert\") pod \"222f65f2-0b82-4760-bcf7-779244720f01\" (UID: \"222f65f2-0b82-4760-bcf7-779244720f01\") " Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.625670 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/222f65f2-0b82-4760-bcf7-779244720f01-host-var-lib-cni-networks-ovn-kubernetes\") pod \"222f65f2-0b82-4760-bcf7-779244720f01\" (UID: \"222f65f2-0b82-4760-bcf7-779244720f01\") " Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.625687 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/222f65f2-0b82-4760-bcf7-779244720f01-etc-openvswitch\") pod \"222f65f2-0b82-4760-bcf7-779244720f01\" (UID: \"222f65f2-0b82-4760-bcf7-779244720f01\") " Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.625704 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/222f65f2-0b82-4760-bcf7-779244720f01-run-ovn\") pod \"222f65f2-0b82-4760-bcf7-779244720f01\" (UID: \"222f65f2-0b82-4760-bcf7-779244720f01\") " Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.625723 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/222f65f2-0b82-4760-bcf7-779244720f01-env-overrides\") pod \"222f65f2-0b82-4760-bcf7-779244720f01\" (UID: \"222f65f2-0b82-4760-bcf7-779244720f01\") " Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.625746 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/222f65f2-0b82-4760-bcf7-779244720f01-ovnkube-script-lib\") pod \"222f65f2-0b82-4760-bcf7-779244720f01\" (UID: \"222f65f2-0b82-4760-bcf7-779244720f01\") " Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.625780 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/222f65f2-0b82-4760-bcf7-779244720f01-run-systemd\") pod \"222f65f2-0b82-4760-bcf7-779244720f01\" (UID: \"222f65f2-0b82-4760-bcf7-779244720f01\") " Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.625802 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/222f65f2-0b82-4760-bcf7-779244720f01-log-socket\") pod \"222f65f2-0b82-4760-bcf7-779244720f01\" (UID: \"222f65f2-0b82-4760-bcf7-779244720f01\") " Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.625832 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/222f65f2-0b82-4760-bcf7-779244720f01-systemd-units\") pod \"222f65f2-0b82-4760-bcf7-779244720f01\" (UID: \"222f65f2-0b82-4760-bcf7-779244720f01\") " Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.625862 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/222f65f2-0b82-4760-bcf7-779244720f01-host-cni-bin\") pod \"222f65f2-0b82-4760-bcf7-779244720f01\" (UID: \"222f65f2-0b82-4760-bcf7-779244720f01\") " Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.625882 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mmvwn\" (UniqueName: \"kubernetes.io/projected/222f65f2-0b82-4760-bcf7-779244720f01-kube-api-access-mmvwn\") pod \"222f65f2-0b82-4760-bcf7-779244720f01\" (UID: \"222f65f2-0b82-4760-bcf7-779244720f01\") " Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.625903 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/222f65f2-0b82-4760-bcf7-779244720f01-host-run-ovn-kubernetes\") pod \"222f65f2-0b82-4760-bcf7-779244720f01\" (UID: \"222f65f2-0b82-4760-bcf7-779244720f01\") " Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.625929 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/222f65f2-0b82-4760-bcf7-779244720f01-run-openvswitch\") pod \"222f65f2-0b82-4760-bcf7-779244720f01\" (UID: \"222f65f2-0b82-4760-bcf7-779244720f01\") " Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.626006 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/222f65f2-0b82-4760-bcf7-779244720f01-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "222f65f2-0b82-4760-bcf7-779244720f01" (UID: "222f65f2-0b82-4760-bcf7-779244720f01"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.626141 4897 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/222f65f2-0b82-4760-bcf7-779244720f01-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.626155 4897 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/222f65f2-0b82-4760-bcf7-779244720f01-host-run-netns\") on node \"crc\" DevicePath \"\"" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.626165 4897 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/222f65f2-0b82-4760-bcf7-779244720f01-host-cni-netd\") on node \"crc\" DevicePath \"\"" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.626174 4897 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/222f65f2-0b82-4760-bcf7-779244720f01-host-kubelet\") on node \"crc\" DevicePath \"\"" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.626182 4897 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/222f65f2-0b82-4760-bcf7-779244720f01-host-slash\") on node \"crc\" DevicePath \"\"" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.626189 4897 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/222f65f2-0b82-4760-bcf7-779244720f01-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.626211 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/222f65f2-0b82-4760-bcf7-779244720f01-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "222f65f2-0b82-4760-bcf7-779244720f01" (UID: "222f65f2-0b82-4760-bcf7-779244720f01"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.626229 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/222f65f2-0b82-4760-bcf7-779244720f01-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "222f65f2-0b82-4760-bcf7-779244720f01" (UID: "222f65f2-0b82-4760-bcf7-779244720f01"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.625645 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/222f65f2-0b82-4760-bcf7-779244720f01-node-log" (OuterVolumeSpecName: "node-log") pod "222f65f2-0b82-4760-bcf7-779244720f01" (UID: "222f65f2-0b82-4760-bcf7-779244720f01"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.627007 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/222f65f2-0b82-4760-bcf7-779244720f01-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "222f65f2-0b82-4760-bcf7-779244720f01" (UID: "222f65f2-0b82-4760-bcf7-779244720f01"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.627021 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/222f65f2-0b82-4760-bcf7-779244720f01-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "222f65f2-0b82-4760-bcf7-779244720f01" (UID: "222f65f2-0b82-4760-bcf7-779244720f01"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.627022 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/222f65f2-0b82-4760-bcf7-779244720f01-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "222f65f2-0b82-4760-bcf7-779244720f01" (UID: "222f65f2-0b82-4760-bcf7-779244720f01"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.627059 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/222f65f2-0b82-4760-bcf7-779244720f01-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "222f65f2-0b82-4760-bcf7-779244720f01" (UID: "222f65f2-0b82-4760-bcf7-779244720f01"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.627045 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/222f65f2-0b82-4760-bcf7-779244720f01-log-socket" (OuterVolumeSpecName: "log-socket") pod "222f65f2-0b82-4760-bcf7-779244720f01" (UID: "222f65f2-0b82-4760-bcf7-779244720f01"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.627087 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/222f65f2-0b82-4760-bcf7-779244720f01-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "222f65f2-0b82-4760-bcf7-779244720f01" (UID: "222f65f2-0b82-4760-bcf7-779244720f01"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.627102 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/222f65f2-0b82-4760-bcf7-779244720f01-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "222f65f2-0b82-4760-bcf7-779244720f01" (UID: "222f65f2-0b82-4760-bcf7-779244720f01"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.627331 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/222f65f2-0b82-4760-bcf7-779244720f01-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "222f65f2-0b82-4760-bcf7-779244720f01" (UID: "222f65f2-0b82-4760-bcf7-779244720f01"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.645073 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/222f65f2-0b82-4760-bcf7-779244720f01-kube-api-access-mmvwn" (OuterVolumeSpecName: "kube-api-access-mmvwn") pod "222f65f2-0b82-4760-bcf7-779244720f01" (UID: "222f65f2-0b82-4760-bcf7-779244720f01"). InnerVolumeSpecName "kube-api-access-mmvwn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.645585 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/222f65f2-0b82-4760-bcf7-779244720f01-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "222f65f2-0b82-4760-bcf7-779244720f01" (UID: "222f65f2-0b82-4760-bcf7-779244720f01"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.646744 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/222f65f2-0b82-4760-bcf7-779244720f01-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "222f65f2-0b82-4760-bcf7-779244720f01" (UID: "222f65f2-0b82-4760-bcf7-779244720f01"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.672719 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-7d6sl"] Nov 21 14:19:33 crc kubenswrapper[4897]: E1121 14:19:33.672925 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="222f65f2-0b82-4760-bcf7-779244720f01" containerName="northd" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.672935 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="222f65f2-0b82-4760-bcf7-779244720f01" containerName="northd" Nov 21 14:19:33 crc kubenswrapper[4897]: E1121 14:19:33.672944 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="222f65f2-0b82-4760-bcf7-779244720f01" containerName="kube-rbac-proxy-ovn-metrics" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.672951 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="222f65f2-0b82-4760-bcf7-779244720f01" containerName="kube-rbac-proxy-ovn-metrics" Nov 21 14:19:33 crc kubenswrapper[4897]: E1121 14:19:33.672962 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="222f65f2-0b82-4760-bcf7-779244720f01" containerName="ovnkube-controller" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.672968 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="222f65f2-0b82-4760-bcf7-779244720f01" containerName="ovnkube-controller" Nov 21 14:19:33 crc kubenswrapper[4897]: E1121 14:19:33.672974 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c792eeb7-8d0e-44fc-a600-fe5c6e4f997e" containerName="util" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.672980 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="c792eeb7-8d0e-44fc-a600-fe5c6e4f997e" containerName="util" Nov 21 14:19:33 crc kubenswrapper[4897]: E1121 14:19:33.672990 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c792eeb7-8d0e-44fc-a600-fe5c6e4f997e" containerName="extract" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.672995 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="c792eeb7-8d0e-44fc-a600-fe5c6e4f997e" containerName="extract" Nov 21 14:19:33 crc kubenswrapper[4897]: E1121 14:19:33.673004 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="222f65f2-0b82-4760-bcf7-779244720f01" containerName="sbdb" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.673009 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="222f65f2-0b82-4760-bcf7-779244720f01" containerName="sbdb" Nov 21 14:19:33 crc kubenswrapper[4897]: E1121 14:19:33.673018 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="222f65f2-0b82-4760-bcf7-779244720f01" containerName="ovnkube-controller" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.673024 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="222f65f2-0b82-4760-bcf7-779244720f01" containerName="ovnkube-controller" Nov 21 14:19:33 crc kubenswrapper[4897]: E1121 14:19:33.673032 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="222f65f2-0b82-4760-bcf7-779244720f01" containerName="nbdb" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.673038 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="222f65f2-0b82-4760-bcf7-779244720f01" containerName="nbdb" Nov 21 14:19:33 crc kubenswrapper[4897]: E1121 14:19:33.673050 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="222f65f2-0b82-4760-bcf7-779244720f01" containerName="ovnkube-controller" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.673056 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="222f65f2-0b82-4760-bcf7-779244720f01" containerName="ovnkube-controller" Nov 21 14:19:33 crc kubenswrapper[4897]: E1121 14:19:33.673064 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="222f65f2-0b82-4760-bcf7-779244720f01" containerName="kube-rbac-proxy-node" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.673070 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="222f65f2-0b82-4760-bcf7-779244720f01" containerName="kube-rbac-proxy-node" Nov 21 14:19:33 crc kubenswrapper[4897]: E1121 14:19:33.673077 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="222f65f2-0b82-4760-bcf7-779244720f01" containerName="kubecfg-setup" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.673083 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="222f65f2-0b82-4760-bcf7-779244720f01" containerName="kubecfg-setup" Nov 21 14:19:33 crc kubenswrapper[4897]: E1121 14:19:33.673092 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c792eeb7-8d0e-44fc-a600-fe5c6e4f997e" containerName="pull" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.673098 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="c792eeb7-8d0e-44fc-a600-fe5c6e4f997e" containerName="pull" Nov 21 14:19:33 crc kubenswrapper[4897]: E1121 14:19:33.673110 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="222f65f2-0b82-4760-bcf7-779244720f01" containerName="ovn-controller" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.673115 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="222f65f2-0b82-4760-bcf7-779244720f01" containerName="ovn-controller" Nov 21 14:19:33 crc kubenswrapper[4897]: E1121 14:19:33.673123 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="222f65f2-0b82-4760-bcf7-779244720f01" containerName="ovn-acl-logging" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.673129 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="222f65f2-0b82-4760-bcf7-779244720f01" containerName="ovn-acl-logging" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.673225 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="222f65f2-0b82-4760-bcf7-779244720f01" containerName="ovnkube-controller" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.673233 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="222f65f2-0b82-4760-bcf7-779244720f01" containerName="ovnkube-controller" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.673239 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="222f65f2-0b82-4760-bcf7-779244720f01" containerName="kube-rbac-proxy-node" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.673245 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="222f65f2-0b82-4760-bcf7-779244720f01" containerName="sbdb" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.673251 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="c792eeb7-8d0e-44fc-a600-fe5c6e4f997e" containerName="extract" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.673258 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="222f65f2-0b82-4760-bcf7-779244720f01" containerName="ovn-acl-logging" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.673267 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="222f65f2-0b82-4760-bcf7-779244720f01" containerName="ovnkube-controller" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.673274 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="222f65f2-0b82-4760-bcf7-779244720f01" containerName="ovn-controller" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.673281 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="222f65f2-0b82-4760-bcf7-779244720f01" containerName="nbdb" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.673288 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="222f65f2-0b82-4760-bcf7-779244720f01" containerName="kube-rbac-proxy-ovn-metrics" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.673297 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="222f65f2-0b82-4760-bcf7-779244720f01" containerName="ovnkube-controller" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.673303 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="222f65f2-0b82-4760-bcf7-779244720f01" containerName="northd" Nov 21 14:19:33 crc kubenswrapper[4897]: E1121 14:19:33.673390 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="222f65f2-0b82-4760-bcf7-779244720f01" containerName="ovnkube-controller" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.673398 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="222f65f2-0b82-4760-bcf7-779244720f01" containerName="ovnkube-controller" Nov 21 14:19:33 crc kubenswrapper[4897]: E1121 14:19:33.673407 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="222f65f2-0b82-4760-bcf7-779244720f01" containerName="ovnkube-controller" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.673413 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="222f65f2-0b82-4760-bcf7-779244720f01" containerName="ovnkube-controller" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.673532 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="222f65f2-0b82-4760-bcf7-779244720f01" containerName="ovnkube-controller" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.675151 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-7d6sl" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.726926 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6j25v\" (UniqueName: \"kubernetes.io/projected/aee1eb81-536e-4dc3-be7f-f38a5b7efd83-kube-api-access-6j25v\") pod \"ovnkube-node-7d6sl\" (UID: \"aee1eb81-536e-4dc3-be7f-f38a5b7efd83\") " pod="openshift-ovn-kubernetes/ovnkube-node-7d6sl" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.726977 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/aee1eb81-536e-4dc3-be7f-f38a5b7efd83-run-openvswitch\") pod \"ovnkube-node-7d6sl\" (UID: \"aee1eb81-536e-4dc3-be7f-f38a5b7efd83\") " pod="openshift-ovn-kubernetes/ovnkube-node-7d6sl" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.727008 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/aee1eb81-536e-4dc3-be7f-f38a5b7efd83-ovnkube-config\") pod \"ovnkube-node-7d6sl\" (UID: \"aee1eb81-536e-4dc3-be7f-f38a5b7efd83\") " pod="openshift-ovn-kubernetes/ovnkube-node-7d6sl" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.727031 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/aee1eb81-536e-4dc3-be7f-f38a5b7efd83-host-run-netns\") pod \"ovnkube-node-7d6sl\" (UID: \"aee1eb81-536e-4dc3-be7f-f38a5b7efd83\") " pod="openshift-ovn-kubernetes/ovnkube-node-7d6sl" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.727144 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/aee1eb81-536e-4dc3-be7f-f38a5b7efd83-env-overrides\") pod \"ovnkube-node-7d6sl\" (UID: \"aee1eb81-536e-4dc3-be7f-f38a5b7efd83\") " pod="openshift-ovn-kubernetes/ovnkube-node-7d6sl" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.727198 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/aee1eb81-536e-4dc3-be7f-f38a5b7efd83-ovnkube-script-lib\") pod \"ovnkube-node-7d6sl\" (UID: \"aee1eb81-536e-4dc3-be7f-f38a5b7efd83\") " pod="openshift-ovn-kubernetes/ovnkube-node-7d6sl" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.727308 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/aee1eb81-536e-4dc3-be7f-f38a5b7efd83-host-cni-netd\") pod \"ovnkube-node-7d6sl\" (UID: \"aee1eb81-536e-4dc3-be7f-f38a5b7efd83\") " pod="openshift-ovn-kubernetes/ovnkube-node-7d6sl" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.727355 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/aee1eb81-536e-4dc3-be7f-f38a5b7efd83-systemd-units\") pod \"ovnkube-node-7d6sl\" (UID: \"aee1eb81-536e-4dc3-be7f-f38a5b7efd83\") " pod="openshift-ovn-kubernetes/ovnkube-node-7d6sl" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.727372 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/aee1eb81-536e-4dc3-be7f-f38a5b7efd83-node-log\") pod \"ovnkube-node-7d6sl\" (UID: \"aee1eb81-536e-4dc3-be7f-f38a5b7efd83\") " pod="openshift-ovn-kubernetes/ovnkube-node-7d6sl" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.727391 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/aee1eb81-536e-4dc3-be7f-f38a5b7efd83-ovn-node-metrics-cert\") pod \"ovnkube-node-7d6sl\" (UID: \"aee1eb81-536e-4dc3-be7f-f38a5b7efd83\") " pod="openshift-ovn-kubernetes/ovnkube-node-7d6sl" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.727427 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/aee1eb81-536e-4dc3-be7f-f38a5b7efd83-host-run-ovn-kubernetes\") pod \"ovnkube-node-7d6sl\" (UID: \"aee1eb81-536e-4dc3-be7f-f38a5b7efd83\") " pod="openshift-ovn-kubernetes/ovnkube-node-7d6sl" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.727469 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/aee1eb81-536e-4dc3-be7f-f38a5b7efd83-run-systemd\") pod \"ovnkube-node-7d6sl\" (UID: \"aee1eb81-536e-4dc3-be7f-f38a5b7efd83\") " pod="openshift-ovn-kubernetes/ovnkube-node-7d6sl" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.727497 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/aee1eb81-536e-4dc3-be7f-f38a5b7efd83-var-lib-openvswitch\") pod \"ovnkube-node-7d6sl\" (UID: \"aee1eb81-536e-4dc3-be7f-f38a5b7efd83\") " pod="openshift-ovn-kubernetes/ovnkube-node-7d6sl" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.727534 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/aee1eb81-536e-4dc3-be7f-f38a5b7efd83-log-socket\") pod \"ovnkube-node-7d6sl\" (UID: \"aee1eb81-536e-4dc3-be7f-f38a5b7efd83\") " pod="openshift-ovn-kubernetes/ovnkube-node-7d6sl" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.727556 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/aee1eb81-536e-4dc3-be7f-f38a5b7efd83-host-slash\") pod \"ovnkube-node-7d6sl\" (UID: \"aee1eb81-536e-4dc3-be7f-f38a5b7efd83\") " pod="openshift-ovn-kubernetes/ovnkube-node-7d6sl" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.727577 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/aee1eb81-536e-4dc3-be7f-f38a5b7efd83-host-kubelet\") pod \"ovnkube-node-7d6sl\" (UID: \"aee1eb81-536e-4dc3-be7f-f38a5b7efd83\") " pod="openshift-ovn-kubernetes/ovnkube-node-7d6sl" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.727596 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/aee1eb81-536e-4dc3-be7f-f38a5b7efd83-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-7d6sl\" (UID: \"aee1eb81-536e-4dc3-be7f-f38a5b7efd83\") " pod="openshift-ovn-kubernetes/ovnkube-node-7d6sl" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.727622 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/aee1eb81-536e-4dc3-be7f-f38a5b7efd83-etc-openvswitch\") pod \"ovnkube-node-7d6sl\" (UID: \"aee1eb81-536e-4dc3-be7f-f38a5b7efd83\") " pod="openshift-ovn-kubernetes/ovnkube-node-7d6sl" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.727639 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/aee1eb81-536e-4dc3-be7f-f38a5b7efd83-host-cni-bin\") pod \"ovnkube-node-7d6sl\" (UID: \"aee1eb81-536e-4dc3-be7f-f38a5b7efd83\") " pod="openshift-ovn-kubernetes/ovnkube-node-7d6sl" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.727653 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/aee1eb81-536e-4dc3-be7f-f38a5b7efd83-run-ovn\") pod \"ovnkube-node-7d6sl\" (UID: \"aee1eb81-536e-4dc3-be7f-f38a5b7efd83\") " pod="openshift-ovn-kubernetes/ovnkube-node-7d6sl" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.727719 4897 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/222f65f2-0b82-4760-bcf7-779244720f01-node-log\") on node \"crc\" DevicePath \"\"" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.727760 4897 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/222f65f2-0b82-4760-bcf7-779244720f01-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.727783 4897 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/222f65f2-0b82-4760-bcf7-779244720f01-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.727793 4897 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/222f65f2-0b82-4760-bcf7-779244720f01-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.727804 4897 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/222f65f2-0b82-4760-bcf7-779244720f01-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.727814 4897 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/222f65f2-0b82-4760-bcf7-779244720f01-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.727824 4897 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/222f65f2-0b82-4760-bcf7-779244720f01-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.727835 4897 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/222f65f2-0b82-4760-bcf7-779244720f01-run-systemd\") on node \"crc\" DevicePath \"\"" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.727843 4897 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/222f65f2-0b82-4760-bcf7-779244720f01-log-socket\") on node \"crc\" DevicePath \"\"" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.727852 4897 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/222f65f2-0b82-4760-bcf7-779244720f01-systemd-units\") on node \"crc\" DevicePath \"\"" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.727862 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mmvwn\" (UniqueName: \"kubernetes.io/projected/222f65f2-0b82-4760-bcf7-779244720f01-kube-api-access-mmvwn\") on node \"crc\" DevicePath \"\"" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.727871 4897 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/222f65f2-0b82-4760-bcf7-779244720f01-host-cni-bin\") on node \"crc\" DevicePath \"\"" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.727880 4897 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/222f65f2-0b82-4760-bcf7-779244720f01-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.727889 4897 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/222f65f2-0b82-4760-bcf7-779244720f01-run-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.829267 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/aee1eb81-536e-4dc3-be7f-f38a5b7efd83-var-lib-openvswitch\") pod \"ovnkube-node-7d6sl\" (UID: \"aee1eb81-536e-4dc3-be7f-f38a5b7efd83\") " pod="openshift-ovn-kubernetes/ovnkube-node-7d6sl" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.829308 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/aee1eb81-536e-4dc3-be7f-f38a5b7efd83-host-slash\") pod \"ovnkube-node-7d6sl\" (UID: \"aee1eb81-536e-4dc3-be7f-f38a5b7efd83\") " pod="openshift-ovn-kubernetes/ovnkube-node-7d6sl" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.829326 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/aee1eb81-536e-4dc3-be7f-f38a5b7efd83-log-socket\") pod \"ovnkube-node-7d6sl\" (UID: \"aee1eb81-536e-4dc3-be7f-f38a5b7efd83\") " pod="openshift-ovn-kubernetes/ovnkube-node-7d6sl" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.829351 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/aee1eb81-536e-4dc3-be7f-f38a5b7efd83-host-kubelet\") pod \"ovnkube-node-7d6sl\" (UID: \"aee1eb81-536e-4dc3-be7f-f38a5b7efd83\") " pod="openshift-ovn-kubernetes/ovnkube-node-7d6sl" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.829371 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/aee1eb81-536e-4dc3-be7f-f38a5b7efd83-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-7d6sl\" (UID: \"aee1eb81-536e-4dc3-be7f-f38a5b7efd83\") " pod="openshift-ovn-kubernetes/ovnkube-node-7d6sl" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.829396 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/aee1eb81-536e-4dc3-be7f-f38a5b7efd83-etc-openvswitch\") pod \"ovnkube-node-7d6sl\" (UID: \"aee1eb81-536e-4dc3-be7f-f38a5b7efd83\") " pod="openshift-ovn-kubernetes/ovnkube-node-7d6sl" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.829419 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/aee1eb81-536e-4dc3-be7f-f38a5b7efd83-host-cni-bin\") pod \"ovnkube-node-7d6sl\" (UID: \"aee1eb81-536e-4dc3-be7f-f38a5b7efd83\") " pod="openshift-ovn-kubernetes/ovnkube-node-7d6sl" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.829443 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/aee1eb81-536e-4dc3-be7f-f38a5b7efd83-run-ovn\") pod \"ovnkube-node-7d6sl\" (UID: \"aee1eb81-536e-4dc3-be7f-f38a5b7efd83\") " pod="openshift-ovn-kubernetes/ovnkube-node-7d6sl" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.829462 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6j25v\" (UniqueName: \"kubernetes.io/projected/aee1eb81-536e-4dc3-be7f-f38a5b7efd83-kube-api-access-6j25v\") pod \"ovnkube-node-7d6sl\" (UID: \"aee1eb81-536e-4dc3-be7f-f38a5b7efd83\") " pod="openshift-ovn-kubernetes/ovnkube-node-7d6sl" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.829478 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/aee1eb81-536e-4dc3-be7f-f38a5b7efd83-run-openvswitch\") pod \"ovnkube-node-7d6sl\" (UID: \"aee1eb81-536e-4dc3-be7f-f38a5b7efd83\") " pod="openshift-ovn-kubernetes/ovnkube-node-7d6sl" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.829520 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/aee1eb81-536e-4dc3-be7f-f38a5b7efd83-ovnkube-config\") pod \"ovnkube-node-7d6sl\" (UID: \"aee1eb81-536e-4dc3-be7f-f38a5b7efd83\") " pod="openshift-ovn-kubernetes/ovnkube-node-7d6sl" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.829535 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/aee1eb81-536e-4dc3-be7f-f38a5b7efd83-host-run-netns\") pod \"ovnkube-node-7d6sl\" (UID: \"aee1eb81-536e-4dc3-be7f-f38a5b7efd83\") " pod="openshift-ovn-kubernetes/ovnkube-node-7d6sl" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.829556 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/aee1eb81-536e-4dc3-be7f-f38a5b7efd83-env-overrides\") pod \"ovnkube-node-7d6sl\" (UID: \"aee1eb81-536e-4dc3-be7f-f38a5b7efd83\") " pod="openshift-ovn-kubernetes/ovnkube-node-7d6sl" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.829572 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/aee1eb81-536e-4dc3-be7f-f38a5b7efd83-ovnkube-script-lib\") pod \"ovnkube-node-7d6sl\" (UID: \"aee1eb81-536e-4dc3-be7f-f38a5b7efd83\") " pod="openshift-ovn-kubernetes/ovnkube-node-7d6sl" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.829596 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/aee1eb81-536e-4dc3-be7f-f38a5b7efd83-host-cni-netd\") pod \"ovnkube-node-7d6sl\" (UID: \"aee1eb81-536e-4dc3-be7f-f38a5b7efd83\") " pod="openshift-ovn-kubernetes/ovnkube-node-7d6sl" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.829612 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/aee1eb81-536e-4dc3-be7f-f38a5b7efd83-systemd-units\") pod \"ovnkube-node-7d6sl\" (UID: \"aee1eb81-536e-4dc3-be7f-f38a5b7efd83\") " pod="openshift-ovn-kubernetes/ovnkube-node-7d6sl" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.829625 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/aee1eb81-536e-4dc3-be7f-f38a5b7efd83-node-log\") pod \"ovnkube-node-7d6sl\" (UID: \"aee1eb81-536e-4dc3-be7f-f38a5b7efd83\") " pod="openshift-ovn-kubernetes/ovnkube-node-7d6sl" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.829639 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/aee1eb81-536e-4dc3-be7f-f38a5b7efd83-ovn-node-metrics-cert\") pod \"ovnkube-node-7d6sl\" (UID: \"aee1eb81-536e-4dc3-be7f-f38a5b7efd83\") " pod="openshift-ovn-kubernetes/ovnkube-node-7d6sl" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.829654 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/aee1eb81-536e-4dc3-be7f-f38a5b7efd83-host-run-ovn-kubernetes\") pod \"ovnkube-node-7d6sl\" (UID: \"aee1eb81-536e-4dc3-be7f-f38a5b7efd83\") " pod="openshift-ovn-kubernetes/ovnkube-node-7d6sl" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.829675 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/aee1eb81-536e-4dc3-be7f-f38a5b7efd83-run-systemd\") pod \"ovnkube-node-7d6sl\" (UID: \"aee1eb81-536e-4dc3-be7f-f38a5b7efd83\") " pod="openshift-ovn-kubernetes/ovnkube-node-7d6sl" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.829739 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/aee1eb81-536e-4dc3-be7f-f38a5b7efd83-run-systemd\") pod \"ovnkube-node-7d6sl\" (UID: \"aee1eb81-536e-4dc3-be7f-f38a5b7efd83\") " pod="openshift-ovn-kubernetes/ovnkube-node-7d6sl" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.829775 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/aee1eb81-536e-4dc3-be7f-f38a5b7efd83-var-lib-openvswitch\") pod \"ovnkube-node-7d6sl\" (UID: \"aee1eb81-536e-4dc3-be7f-f38a5b7efd83\") " pod="openshift-ovn-kubernetes/ovnkube-node-7d6sl" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.829798 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/aee1eb81-536e-4dc3-be7f-f38a5b7efd83-host-slash\") pod \"ovnkube-node-7d6sl\" (UID: \"aee1eb81-536e-4dc3-be7f-f38a5b7efd83\") " pod="openshift-ovn-kubernetes/ovnkube-node-7d6sl" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.829823 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/aee1eb81-536e-4dc3-be7f-f38a5b7efd83-log-socket\") pod \"ovnkube-node-7d6sl\" (UID: \"aee1eb81-536e-4dc3-be7f-f38a5b7efd83\") " pod="openshift-ovn-kubernetes/ovnkube-node-7d6sl" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.829850 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/aee1eb81-536e-4dc3-be7f-f38a5b7efd83-host-kubelet\") pod \"ovnkube-node-7d6sl\" (UID: \"aee1eb81-536e-4dc3-be7f-f38a5b7efd83\") " pod="openshift-ovn-kubernetes/ovnkube-node-7d6sl" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.829873 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/aee1eb81-536e-4dc3-be7f-f38a5b7efd83-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-7d6sl\" (UID: \"aee1eb81-536e-4dc3-be7f-f38a5b7efd83\") " pod="openshift-ovn-kubernetes/ovnkube-node-7d6sl" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.829895 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/aee1eb81-536e-4dc3-be7f-f38a5b7efd83-etc-openvswitch\") pod \"ovnkube-node-7d6sl\" (UID: \"aee1eb81-536e-4dc3-be7f-f38a5b7efd83\") " pod="openshift-ovn-kubernetes/ovnkube-node-7d6sl" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.829913 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/aee1eb81-536e-4dc3-be7f-f38a5b7efd83-host-cni-bin\") pod \"ovnkube-node-7d6sl\" (UID: \"aee1eb81-536e-4dc3-be7f-f38a5b7efd83\") " pod="openshift-ovn-kubernetes/ovnkube-node-7d6sl" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.829932 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/aee1eb81-536e-4dc3-be7f-f38a5b7efd83-run-ovn\") pod \"ovnkube-node-7d6sl\" (UID: \"aee1eb81-536e-4dc3-be7f-f38a5b7efd83\") " pod="openshift-ovn-kubernetes/ovnkube-node-7d6sl" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.830196 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/aee1eb81-536e-4dc3-be7f-f38a5b7efd83-run-openvswitch\") pod \"ovnkube-node-7d6sl\" (UID: \"aee1eb81-536e-4dc3-be7f-f38a5b7efd83\") " pod="openshift-ovn-kubernetes/ovnkube-node-7d6sl" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.830813 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/aee1eb81-536e-4dc3-be7f-f38a5b7efd83-ovnkube-config\") pod \"ovnkube-node-7d6sl\" (UID: \"aee1eb81-536e-4dc3-be7f-f38a5b7efd83\") " pod="openshift-ovn-kubernetes/ovnkube-node-7d6sl" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.830852 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/aee1eb81-536e-4dc3-be7f-f38a5b7efd83-host-run-netns\") pod \"ovnkube-node-7d6sl\" (UID: \"aee1eb81-536e-4dc3-be7f-f38a5b7efd83\") " pod="openshift-ovn-kubernetes/ovnkube-node-7d6sl" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.831125 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/aee1eb81-536e-4dc3-be7f-f38a5b7efd83-env-overrides\") pod \"ovnkube-node-7d6sl\" (UID: \"aee1eb81-536e-4dc3-be7f-f38a5b7efd83\") " pod="openshift-ovn-kubernetes/ovnkube-node-7d6sl" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.831524 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/aee1eb81-536e-4dc3-be7f-f38a5b7efd83-ovnkube-script-lib\") pod \"ovnkube-node-7d6sl\" (UID: \"aee1eb81-536e-4dc3-be7f-f38a5b7efd83\") " pod="openshift-ovn-kubernetes/ovnkube-node-7d6sl" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.831559 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/aee1eb81-536e-4dc3-be7f-f38a5b7efd83-host-cni-netd\") pod \"ovnkube-node-7d6sl\" (UID: \"aee1eb81-536e-4dc3-be7f-f38a5b7efd83\") " pod="openshift-ovn-kubernetes/ovnkube-node-7d6sl" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.831579 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/aee1eb81-536e-4dc3-be7f-f38a5b7efd83-systemd-units\") pod \"ovnkube-node-7d6sl\" (UID: \"aee1eb81-536e-4dc3-be7f-f38a5b7efd83\") " pod="openshift-ovn-kubernetes/ovnkube-node-7d6sl" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.831600 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/aee1eb81-536e-4dc3-be7f-f38a5b7efd83-node-log\") pod \"ovnkube-node-7d6sl\" (UID: \"aee1eb81-536e-4dc3-be7f-f38a5b7efd83\") " pod="openshift-ovn-kubernetes/ovnkube-node-7d6sl" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.831961 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/aee1eb81-536e-4dc3-be7f-f38a5b7efd83-host-run-ovn-kubernetes\") pod \"ovnkube-node-7d6sl\" (UID: \"aee1eb81-536e-4dc3-be7f-f38a5b7efd83\") " pod="openshift-ovn-kubernetes/ovnkube-node-7d6sl" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.835095 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/aee1eb81-536e-4dc3-be7f-f38a5b7efd83-ovn-node-metrics-cert\") pod \"ovnkube-node-7d6sl\" (UID: \"aee1eb81-536e-4dc3-be7f-f38a5b7efd83\") " pod="openshift-ovn-kubernetes/ovnkube-node-7d6sl" Nov 21 14:19:33 crc kubenswrapper[4897]: I1121 14:19:33.853356 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6j25v\" (UniqueName: \"kubernetes.io/projected/aee1eb81-536e-4dc3-be7f-f38a5b7efd83-kube-api-access-6j25v\") pod \"ovnkube-node-7d6sl\" (UID: \"aee1eb81-536e-4dc3-be7f-f38a5b7efd83\") " pod="openshift-ovn-kubernetes/ovnkube-node-7d6sl" Nov 21 14:19:34 crc kubenswrapper[4897]: I1121 14:19:34.006853 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-7d6sl" Nov 21 14:19:34 crc kubenswrapper[4897]: I1121 14:19:34.151644 4897 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-6ljgx_5f3b70b5-9294-4f69-9528-500d28f34c89/kube-multus/2.log" Nov 21 14:19:34 crc kubenswrapper[4897]: I1121 14:19:34.163303 4897 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tpfb6_222f65f2-0b82-4760-bcf7-779244720f01/ovn-acl-logging/0.log" Nov 21 14:19:34 crc kubenswrapper[4897]: I1121 14:19:34.163800 4897 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tpfb6_222f65f2-0b82-4760-bcf7-779244720f01/ovn-controller/0.log" Nov 21 14:19:34 crc kubenswrapper[4897]: I1121 14:19:34.164077 4897 generic.go:334] "Generic (PLEG): container finished" podID="222f65f2-0b82-4760-bcf7-779244720f01" containerID="ccd51aa0faf3b1327f207b355cddf86a7353c0051f7e5ce0396f3c4bc67e3b16" exitCode=0 Nov 21 14:19:34 crc kubenswrapper[4897]: I1121 14:19:34.164100 4897 generic.go:334] "Generic (PLEG): container finished" podID="222f65f2-0b82-4760-bcf7-779244720f01" containerID="92f1d932a83d65ba6a9a906cc489ff51e76547d07d42656630c59421f833c4db" exitCode=0 Nov 21 14:19:34 crc kubenswrapper[4897]: I1121 14:19:34.164163 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" event={"ID":"222f65f2-0b82-4760-bcf7-779244720f01","Type":"ContainerDied","Data":"ccd51aa0faf3b1327f207b355cddf86a7353c0051f7e5ce0396f3c4bc67e3b16"} Nov 21 14:19:34 crc kubenswrapper[4897]: I1121 14:19:34.164194 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" event={"ID":"222f65f2-0b82-4760-bcf7-779244720f01","Type":"ContainerDied","Data":"92f1d932a83d65ba6a9a906cc489ff51e76547d07d42656630c59421f833c4db"} Nov 21 14:19:34 crc kubenswrapper[4897]: I1121 14:19:34.164207 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" event={"ID":"222f65f2-0b82-4760-bcf7-779244720f01","Type":"ContainerDied","Data":"deecb27cbb58ca8e141468253196c2b73bb738df87fcd915dc32ef490981f458"} Nov 21 14:19:34 crc kubenswrapper[4897]: I1121 14:19:34.164226 4897 scope.go:117] "RemoveContainer" containerID="eaaa7662c416148afbfc746cc7b87fac845ce54740988de2c6d0af4b1a21b9e7" Nov 21 14:19:34 crc kubenswrapper[4897]: I1121 14:19:34.164230 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-tpfb6" Nov 21 14:19:34 crc kubenswrapper[4897]: I1121 14:19:34.166031 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7d6sl" event={"ID":"aee1eb81-536e-4dc3-be7f-f38a5b7efd83","Type":"ContainerStarted","Data":"3cbfb2f1687f144c2351411d6c0b4c77d822fda94c89c9e7bc97fda0c469eb0c"} Nov 21 14:19:34 crc kubenswrapper[4897]: I1121 14:19:34.166069 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7d6sl" event={"ID":"aee1eb81-536e-4dc3-be7f-f38a5b7efd83","Type":"ContainerStarted","Data":"8ff22817693b7fce221eca9630f5038452f636ba51626e87137c4b64bb1d08c5"} Nov 21 14:19:34 crc kubenswrapper[4897]: I1121 14:19:34.208002 4897 scope.go:117] "RemoveContainer" containerID="b486a86e0e97ab409e590f30a1bf5cb9f584a7c67f96ff1236e42ca4f8675b56" Nov 21 14:19:34 crc kubenswrapper[4897]: I1121 14:19:34.225470 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-tpfb6"] Nov 21 14:19:34 crc kubenswrapper[4897]: I1121 14:19:34.229090 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-tpfb6"] Nov 21 14:19:34 crc kubenswrapper[4897]: I1121 14:19:34.238321 4897 scope.go:117] "RemoveContainer" containerID="8971c38daad45bda3d61545f52120dad40992dbf2e1d6e1e776500ed58d73f36" Nov 21 14:19:34 crc kubenswrapper[4897]: I1121 14:19:34.251848 4897 scope.go:117] "RemoveContainer" containerID="cfee19f82e35869a641318a0c45bc7a2ae71d80039f5790e60e277be15c9d3ec" Nov 21 14:19:34 crc kubenswrapper[4897]: I1121 14:19:34.269454 4897 scope.go:117] "RemoveContainer" containerID="ccd51aa0faf3b1327f207b355cddf86a7353c0051f7e5ce0396f3c4bc67e3b16" Nov 21 14:19:34 crc kubenswrapper[4897]: I1121 14:19:34.295426 4897 scope.go:117] "RemoveContainer" containerID="92f1d932a83d65ba6a9a906cc489ff51e76547d07d42656630c59421f833c4db" Nov 21 14:19:34 crc kubenswrapper[4897]: I1121 14:19:34.310955 4897 scope.go:117] "RemoveContainer" containerID="ec279f2f558a8ac653006321b18cbaf13185876fe303c5cc1c65988aef09edbf" Nov 21 14:19:34 crc kubenswrapper[4897]: I1121 14:19:34.333829 4897 scope.go:117] "RemoveContainer" containerID="9de095f3c760b13d5197f624eac3806a1f4186577f2b95eb522532d453eb266e" Nov 21 14:19:34 crc kubenswrapper[4897]: I1121 14:19:34.352824 4897 scope.go:117] "RemoveContainer" containerID="7877db6f5c2bf0eb31218c64d67bca48f921ff7ac62b4e495f29024c086bf738" Nov 21 14:19:34 crc kubenswrapper[4897]: I1121 14:19:34.375860 4897 scope.go:117] "RemoveContainer" containerID="eaaa7662c416148afbfc746cc7b87fac845ce54740988de2c6d0af4b1a21b9e7" Nov 21 14:19:34 crc kubenswrapper[4897]: E1121 14:19:34.376260 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eaaa7662c416148afbfc746cc7b87fac845ce54740988de2c6d0af4b1a21b9e7\": container with ID starting with eaaa7662c416148afbfc746cc7b87fac845ce54740988de2c6d0af4b1a21b9e7 not found: ID does not exist" containerID="eaaa7662c416148afbfc746cc7b87fac845ce54740988de2c6d0af4b1a21b9e7" Nov 21 14:19:34 crc kubenswrapper[4897]: I1121 14:19:34.376289 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eaaa7662c416148afbfc746cc7b87fac845ce54740988de2c6d0af4b1a21b9e7"} err="failed to get container status \"eaaa7662c416148afbfc746cc7b87fac845ce54740988de2c6d0af4b1a21b9e7\": rpc error: code = NotFound desc = could not find container \"eaaa7662c416148afbfc746cc7b87fac845ce54740988de2c6d0af4b1a21b9e7\": container with ID starting with eaaa7662c416148afbfc746cc7b87fac845ce54740988de2c6d0af4b1a21b9e7 not found: ID does not exist" Nov 21 14:19:34 crc kubenswrapper[4897]: I1121 14:19:34.376310 4897 scope.go:117] "RemoveContainer" containerID="b486a86e0e97ab409e590f30a1bf5cb9f584a7c67f96ff1236e42ca4f8675b56" Nov 21 14:19:34 crc kubenswrapper[4897]: E1121 14:19:34.376562 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b486a86e0e97ab409e590f30a1bf5cb9f584a7c67f96ff1236e42ca4f8675b56\": container with ID starting with b486a86e0e97ab409e590f30a1bf5cb9f584a7c67f96ff1236e42ca4f8675b56 not found: ID does not exist" containerID="b486a86e0e97ab409e590f30a1bf5cb9f584a7c67f96ff1236e42ca4f8675b56" Nov 21 14:19:34 crc kubenswrapper[4897]: I1121 14:19:34.376584 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b486a86e0e97ab409e590f30a1bf5cb9f584a7c67f96ff1236e42ca4f8675b56"} err="failed to get container status \"b486a86e0e97ab409e590f30a1bf5cb9f584a7c67f96ff1236e42ca4f8675b56\": rpc error: code = NotFound desc = could not find container \"b486a86e0e97ab409e590f30a1bf5cb9f584a7c67f96ff1236e42ca4f8675b56\": container with ID starting with b486a86e0e97ab409e590f30a1bf5cb9f584a7c67f96ff1236e42ca4f8675b56 not found: ID does not exist" Nov 21 14:19:34 crc kubenswrapper[4897]: I1121 14:19:34.376598 4897 scope.go:117] "RemoveContainer" containerID="8971c38daad45bda3d61545f52120dad40992dbf2e1d6e1e776500ed58d73f36" Nov 21 14:19:34 crc kubenswrapper[4897]: E1121 14:19:34.376786 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8971c38daad45bda3d61545f52120dad40992dbf2e1d6e1e776500ed58d73f36\": container with ID starting with 8971c38daad45bda3d61545f52120dad40992dbf2e1d6e1e776500ed58d73f36 not found: ID does not exist" containerID="8971c38daad45bda3d61545f52120dad40992dbf2e1d6e1e776500ed58d73f36" Nov 21 14:19:34 crc kubenswrapper[4897]: I1121 14:19:34.376811 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8971c38daad45bda3d61545f52120dad40992dbf2e1d6e1e776500ed58d73f36"} err="failed to get container status \"8971c38daad45bda3d61545f52120dad40992dbf2e1d6e1e776500ed58d73f36\": rpc error: code = NotFound desc = could not find container \"8971c38daad45bda3d61545f52120dad40992dbf2e1d6e1e776500ed58d73f36\": container with ID starting with 8971c38daad45bda3d61545f52120dad40992dbf2e1d6e1e776500ed58d73f36 not found: ID does not exist" Nov 21 14:19:34 crc kubenswrapper[4897]: I1121 14:19:34.376824 4897 scope.go:117] "RemoveContainer" containerID="cfee19f82e35869a641318a0c45bc7a2ae71d80039f5790e60e277be15c9d3ec" Nov 21 14:19:34 crc kubenswrapper[4897]: E1121 14:19:34.376991 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cfee19f82e35869a641318a0c45bc7a2ae71d80039f5790e60e277be15c9d3ec\": container with ID starting with cfee19f82e35869a641318a0c45bc7a2ae71d80039f5790e60e277be15c9d3ec not found: ID does not exist" containerID="cfee19f82e35869a641318a0c45bc7a2ae71d80039f5790e60e277be15c9d3ec" Nov 21 14:19:34 crc kubenswrapper[4897]: I1121 14:19:34.377012 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cfee19f82e35869a641318a0c45bc7a2ae71d80039f5790e60e277be15c9d3ec"} err="failed to get container status \"cfee19f82e35869a641318a0c45bc7a2ae71d80039f5790e60e277be15c9d3ec\": rpc error: code = NotFound desc = could not find container \"cfee19f82e35869a641318a0c45bc7a2ae71d80039f5790e60e277be15c9d3ec\": container with ID starting with cfee19f82e35869a641318a0c45bc7a2ae71d80039f5790e60e277be15c9d3ec not found: ID does not exist" Nov 21 14:19:34 crc kubenswrapper[4897]: I1121 14:19:34.377025 4897 scope.go:117] "RemoveContainer" containerID="ccd51aa0faf3b1327f207b355cddf86a7353c0051f7e5ce0396f3c4bc67e3b16" Nov 21 14:19:34 crc kubenswrapper[4897]: E1121 14:19:34.377251 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ccd51aa0faf3b1327f207b355cddf86a7353c0051f7e5ce0396f3c4bc67e3b16\": container with ID starting with ccd51aa0faf3b1327f207b355cddf86a7353c0051f7e5ce0396f3c4bc67e3b16 not found: ID does not exist" containerID="ccd51aa0faf3b1327f207b355cddf86a7353c0051f7e5ce0396f3c4bc67e3b16" Nov 21 14:19:34 crc kubenswrapper[4897]: I1121 14:19:34.377273 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ccd51aa0faf3b1327f207b355cddf86a7353c0051f7e5ce0396f3c4bc67e3b16"} err="failed to get container status \"ccd51aa0faf3b1327f207b355cddf86a7353c0051f7e5ce0396f3c4bc67e3b16\": rpc error: code = NotFound desc = could not find container \"ccd51aa0faf3b1327f207b355cddf86a7353c0051f7e5ce0396f3c4bc67e3b16\": container with ID starting with ccd51aa0faf3b1327f207b355cddf86a7353c0051f7e5ce0396f3c4bc67e3b16 not found: ID does not exist" Nov 21 14:19:34 crc kubenswrapper[4897]: I1121 14:19:34.377288 4897 scope.go:117] "RemoveContainer" containerID="92f1d932a83d65ba6a9a906cc489ff51e76547d07d42656630c59421f833c4db" Nov 21 14:19:34 crc kubenswrapper[4897]: E1121 14:19:34.377456 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"92f1d932a83d65ba6a9a906cc489ff51e76547d07d42656630c59421f833c4db\": container with ID starting with 92f1d932a83d65ba6a9a906cc489ff51e76547d07d42656630c59421f833c4db not found: ID does not exist" containerID="92f1d932a83d65ba6a9a906cc489ff51e76547d07d42656630c59421f833c4db" Nov 21 14:19:34 crc kubenswrapper[4897]: I1121 14:19:34.377477 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"92f1d932a83d65ba6a9a906cc489ff51e76547d07d42656630c59421f833c4db"} err="failed to get container status \"92f1d932a83d65ba6a9a906cc489ff51e76547d07d42656630c59421f833c4db\": rpc error: code = NotFound desc = could not find container \"92f1d932a83d65ba6a9a906cc489ff51e76547d07d42656630c59421f833c4db\": container with ID starting with 92f1d932a83d65ba6a9a906cc489ff51e76547d07d42656630c59421f833c4db not found: ID does not exist" Nov 21 14:19:34 crc kubenswrapper[4897]: I1121 14:19:34.377489 4897 scope.go:117] "RemoveContainer" containerID="ec279f2f558a8ac653006321b18cbaf13185876fe303c5cc1c65988aef09edbf" Nov 21 14:19:34 crc kubenswrapper[4897]: E1121 14:19:34.377720 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec279f2f558a8ac653006321b18cbaf13185876fe303c5cc1c65988aef09edbf\": container with ID starting with ec279f2f558a8ac653006321b18cbaf13185876fe303c5cc1c65988aef09edbf not found: ID does not exist" containerID="ec279f2f558a8ac653006321b18cbaf13185876fe303c5cc1c65988aef09edbf" Nov 21 14:19:34 crc kubenswrapper[4897]: I1121 14:19:34.377747 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec279f2f558a8ac653006321b18cbaf13185876fe303c5cc1c65988aef09edbf"} err="failed to get container status \"ec279f2f558a8ac653006321b18cbaf13185876fe303c5cc1c65988aef09edbf\": rpc error: code = NotFound desc = could not find container \"ec279f2f558a8ac653006321b18cbaf13185876fe303c5cc1c65988aef09edbf\": container with ID starting with ec279f2f558a8ac653006321b18cbaf13185876fe303c5cc1c65988aef09edbf not found: ID does not exist" Nov 21 14:19:34 crc kubenswrapper[4897]: I1121 14:19:34.377760 4897 scope.go:117] "RemoveContainer" containerID="9de095f3c760b13d5197f624eac3806a1f4186577f2b95eb522532d453eb266e" Nov 21 14:19:34 crc kubenswrapper[4897]: E1121 14:19:34.378041 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9de095f3c760b13d5197f624eac3806a1f4186577f2b95eb522532d453eb266e\": container with ID starting with 9de095f3c760b13d5197f624eac3806a1f4186577f2b95eb522532d453eb266e not found: ID does not exist" containerID="9de095f3c760b13d5197f624eac3806a1f4186577f2b95eb522532d453eb266e" Nov 21 14:19:34 crc kubenswrapper[4897]: I1121 14:19:34.378111 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9de095f3c760b13d5197f624eac3806a1f4186577f2b95eb522532d453eb266e"} err="failed to get container status \"9de095f3c760b13d5197f624eac3806a1f4186577f2b95eb522532d453eb266e\": rpc error: code = NotFound desc = could not find container \"9de095f3c760b13d5197f624eac3806a1f4186577f2b95eb522532d453eb266e\": container with ID starting with 9de095f3c760b13d5197f624eac3806a1f4186577f2b95eb522532d453eb266e not found: ID does not exist" Nov 21 14:19:34 crc kubenswrapper[4897]: I1121 14:19:34.378166 4897 scope.go:117] "RemoveContainer" containerID="7877db6f5c2bf0eb31218c64d67bca48f921ff7ac62b4e495f29024c086bf738" Nov 21 14:19:34 crc kubenswrapper[4897]: E1121 14:19:34.378472 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7877db6f5c2bf0eb31218c64d67bca48f921ff7ac62b4e495f29024c086bf738\": container with ID starting with 7877db6f5c2bf0eb31218c64d67bca48f921ff7ac62b4e495f29024c086bf738 not found: ID does not exist" containerID="7877db6f5c2bf0eb31218c64d67bca48f921ff7ac62b4e495f29024c086bf738" Nov 21 14:19:34 crc kubenswrapper[4897]: I1121 14:19:34.378496 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7877db6f5c2bf0eb31218c64d67bca48f921ff7ac62b4e495f29024c086bf738"} err="failed to get container status \"7877db6f5c2bf0eb31218c64d67bca48f921ff7ac62b4e495f29024c086bf738\": rpc error: code = NotFound desc = could not find container \"7877db6f5c2bf0eb31218c64d67bca48f921ff7ac62b4e495f29024c086bf738\": container with ID starting with 7877db6f5c2bf0eb31218c64d67bca48f921ff7ac62b4e495f29024c086bf738 not found: ID does not exist" Nov 21 14:19:34 crc kubenswrapper[4897]: I1121 14:19:34.378527 4897 scope.go:117] "RemoveContainer" containerID="eaaa7662c416148afbfc746cc7b87fac845ce54740988de2c6d0af4b1a21b9e7" Nov 21 14:19:34 crc kubenswrapper[4897]: I1121 14:19:34.378692 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eaaa7662c416148afbfc746cc7b87fac845ce54740988de2c6d0af4b1a21b9e7"} err="failed to get container status \"eaaa7662c416148afbfc746cc7b87fac845ce54740988de2c6d0af4b1a21b9e7\": rpc error: code = NotFound desc = could not find container \"eaaa7662c416148afbfc746cc7b87fac845ce54740988de2c6d0af4b1a21b9e7\": container with ID starting with eaaa7662c416148afbfc746cc7b87fac845ce54740988de2c6d0af4b1a21b9e7 not found: ID does not exist" Nov 21 14:19:34 crc kubenswrapper[4897]: I1121 14:19:34.378713 4897 scope.go:117] "RemoveContainer" containerID="b486a86e0e97ab409e590f30a1bf5cb9f584a7c67f96ff1236e42ca4f8675b56" Nov 21 14:19:34 crc kubenswrapper[4897]: I1121 14:19:34.378871 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b486a86e0e97ab409e590f30a1bf5cb9f584a7c67f96ff1236e42ca4f8675b56"} err="failed to get container status \"b486a86e0e97ab409e590f30a1bf5cb9f584a7c67f96ff1236e42ca4f8675b56\": rpc error: code = NotFound desc = could not find container \"b486a86e0e97ab409e590f30a1bf5cb9f584a7c67f96ff1236e42ca4f8675b56\": container with ID starting with b486a86e0e97ab409e590f30a1bf5cb9f584a7c67f96ff1236e42ca4f8675b56 not found: ID does not exist" Nov 21 14:19:34 crc kubenswrapper[4897]: I1121 14:19:34.378891 4897 scope.go:117] "RemoveContainer" containerID="8971c38daad45bda3d61545f52120dad40992dbf2e1d6e1e776500ed58d73f36" Nov 21 14:19:34 crc kubenswrapper[4897]: I1121 14:19:34.379048 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8971c38daad45bda3d61545f52120dad40992dbf2e1d6e1e776500ed58d73f36"} err="failed to get container status \"8971c38daad45bda3d61545f52120dad40992dbf2e1d6e1e776500ed58d73f36\": rpc error: code = NotFound desc = could not find container \"8971c38daad45bda3d61545f52120dad40992dbf2e1d6e1e776500ed58d73f36\": container with ID starting with 8971c38daad45bda3d61545f52120dad40992dbf2e1d6e1e776500ed58d73f36 not found: ID does not exist" Nov 21 14:19:34 crc kubenswrapper[4897]: I1121 14:19:34.379067 4897 scope.go:117] "RemoveContainer" containerID="cfee19f82e35869a641318a0c45bc7a2ae71d80039f5790e60e277be15c9d3ec" Nov 21 14:19:34 crc kubenswrapper[4897]: I1121 14:19:34.379241 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cfee19f82e35869a641318a0c45bc7a2ae71d80039f5790e60e277be15c9d3ec"} err="failed to get container status \"cfee19f82e35869a641318a0c45bc7a2ae71d80039f5790e60e277be15c9d3ec\": rpc error: code = NotFound desc = could not find container \"cfee19f82e35869a641318a0c45bc7a2ae71d80039f5790e60e277be15c9d3ec\": container with ID starting with cfee19f82e35869a641318a0c45bc7a2ae71d80039f5790e60e277be15c9d3ec not found: ID does not exist" Nov 21 14:19:34 crc kubenswrapper[4897]: I1121 14:19:34.379262 4897 scope.go:117] "RemoveContainer" containerID="ccd51aa0faf3b1327f207b355cddf86a7353c0051f7e5ce0396f3c4bc67e3b16" Nov 21 14:19:34 crc kubenswrapper[4897]: I1121 14:19:34.379425 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ccd51aa0faf3b1327f207b355cddf86a7353c0051f7e5ce0396f3c4bc67e3b16"} err="failed to get container status \"ccd51aa0faf3b1327f207b355cddf86a7353c0051f7e5ce0396f3c4bc67e3b16\": rpc error: code = NotFound desc = could not find container \"ccd51aa0faf3b1327f207b355cddf86a7353c0051f7e5ce0396f3c4bc67e3b16\": container with ID starting with ccd51aa0faf3b1327f207b355cddf86a7353c0051f7e5ce0396f3c4bc67e3b16 not found: ID does not exist" Nov 21 14:19:34 crc kubenswrapper[4897]: I1121 14:19:34.379444 4897 scope.go:117] "RemoveContainer" containerID="92f1d932a83d65ba6a9a906cc489ff51e76547d07d42656630c59421f833c4db" Nov 21 14:19:34 crc kubenswrapper[4897]: I1121 14:19:34.379644 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"92f1d932a83d65ba6a9a906cc489ff51e76547d07d42656630c59421f833c4db"} err="failed to get container status \"92f1d932a83d65ba6a9a906cc489ff51e76547d07d42656630c59421f833c4db\": rpc error: code = NotFound desc = could not find container \"92f1d932a83d65ba6a9a906cc489ff51e76547d07d42656630c59421f833c4db\": container with ID starting with 92f1d932a83d65ba6a9a906cc489ff51e76547d07d42656630c59421f833c4db not found: ID does not exist" Nov 21 14:19:34 crc kubenswrapper[4897]: I1121 14:19:34.379671 4897 scope.go:117] "RemoveContainer" containerID="ec279f2f558a8ac653006321b18cbaf13185876fe303c5cc1c65988aef09edbf" Nov 21 14:19:34 crc kubenswrapper[4897]: I1121 14:19:34.379866 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec279f2f558a8ac653006321b18cbaf13185876fe303c5cc1c65988aef09edbf"} err="failed to get container status \"ec279f2f558a8ac653006321b18cbaf13185876fe303c5cc1c65988aef09edbf\": rpc error: code = NotFound desc = could not find container \"ec279f2f558a8ac653006321b18cbaf13185876fe303c5cc1c65988aef09edbf\": container with ID starting with ec279f2f558a8ac653006321b18cbaf13185876fe303c5cc1c65988aef09edbf not found: ID does not exist" Nov 21 14:19:34 crc kubenswrapper[4897]: I1121 14:19:34.379888 4897 scope.go:117] "RemoveContainer" containerID="9de095f3c760b13d5197f624eac3806a1f4186577f2b95eb522532d453eb266e" Nov 21 14:19:34 crc kubenswrapper[4897]: I1121 14:19:34.380060 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9de095f3c760b13d5197f624eac3806a1f4186577f2b95eb522532d453eb266e"} err="failed to get container status \"9de095f3c760b13d5197f624eac3806a1f4186577f2b95eb522532d453eb266e\": rpc error: code = NotFound desc = could not find container \"9de095f3c760b13d5197f624eac3806a1f4186577f2b95eb522532d453eb266e\": container with ID starting with 9de095f3c760b13d5197f624eac3806a1f4186577f2b95eb522532d453eb266e not found: ID does not exist" Nov 21 14:19:34 crc kubenswrapper[4897]: I1121 14:19:34.380080 4897 scope.go:117] "RemoveContainer" containerID="7877db6f5c2bf0eb31218c64d67bca48f921ff7ac62b4e495f29024c086bf738" Nov 21 14:19:34 crc kubenswrapper[4897]: I1121 14:19:34.380233 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7877db6f5c2bf0eb31218c64d67bca48f921ff7ac62b4e495f29024c086bf738"} err="failed to get container status \"7877db6f5c2bf0eb31218c64d67bca48f921ff7ac62b4e495f29024c086bf738\": rpc error: code = NotFound desc = could not find container \"7877db6f5c2bf0eb31218c64d67bca48f921ff7ac62b4e495f29024c086bf738\": container with ID starting with 7877db6f5c2bf0eb31218c64d67bca48f921ff7ac62b4e495f29024c086bf738 not found: ID does not exist" Nov 21 14:19:34 crc kubenswrapper[4897]: I1121 14:19:34.993479 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-s5lj9"] Nov 21 14:19:34 crc kubenswrapper[4897]: I1121 14:19:34.994537 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-s5lj9" Nov 21 14:19:34 crc kubenswrapper[4897]: I1121 14:19:34.996955 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-dockercfg-dgfdl" Nov 21 14:19:34 crc kubenswrapper[4897]: I1121 14:19:34.997176 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"kube-root-ca.crt" Nov 21 14:19:35 crc kubenswrapper[4897]: I1121 14:19:35.000006 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"openshift-service-ca.crt" Nov 21 14:19:35 crc kubenswrapper[4897]: I1121 14:19:35.046018 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-78b8q\" (UniqueName: \"kubernetes.io/projected/a001ae14-634e-4590-8177-fc7db62c3247-kube-api-access-78b8q\") pod \"obo-prometheus-operator-668cf9dfbb-s5lj9\" (UID: \"a001ae14-634e-4590-8177-fc7db62c3247\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-s5lj9" Nov 21 14:19:35 crc kubenswrapper[4897]: I1121 14:19:35.123212 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-6b8b5c98d-trqb7"] Nov 21 14:19:35 crc kubenswrapper[4897]: I1121 14:19:35.123910 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b8b5c98d-trqb7" Nov 21 14:19:35 crc kubenswrapper[4897]: I1121 14:19:35.128144 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-service-cert" Nov 21 14:19:35 crc kubenswrapper[4897]: I1121 14:19:35.128331 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-dockercfg-dllgh" Nov 21 14:19:35 crc kubenswrapper[4897]: I1121 14:19:35.139676 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-6b8b5c98d-lt42q"] Nov 21 14:19:35 crc kubenswrapper[4897]: I1121 14:19:35.140367 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b8b5c98d-lt42q" Nov 21 14:19:35 crc kubenswrapper[4897]: I1121 14:19:35.147840 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-78b8q\" (UniqueName: \"kubernetes.io/projected/a001ae14-634e-4590-8177-fc7db62c3247-kube-api-access-78b8q\") pod \"obo-prometheus-operator-668cf9dfbb-s5lj9\" (UID: \"a001ae14-634e-4590-8177-fc7db62c3247\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-s5lj9" Nov 21 14:19:35 crc kubenswrapper[4897]: I1121 14:19:35.177141 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-78b8q\" (UniqueName: \"kubernetes.io/projected/a001ae14-634e-4590-8177-fc7db62c3247-kube-api-access-78b8q\") pod \"obo-prometheus-operator-668cf9dfbb-s5lj9\" (UID: \"a001ae14-634e-4590-8177-fc7db62c3247\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-s5lj9" Nov 21 14:19:35 crc kubenswrapper[4897]: I1121 14:19:35.178794 4897 generic.go:334] "Generic (PLEG): container finished" podID="aee1eb81-536e-4dc3-be7f-f38a5b7efd83" containerID="3cbfb2f1687f144c2351411d6c0b4c77d822fda94c89c9e7bc97fda0c469eb0c" exitCode=0 Nov 21 14:19:35 crc kubenswrapper[4897]: I1121 14:19:35.178871 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7d6sl" event={"ID":"aee1eb81-536e-4dc3-be7f-f38a5b7efd83","Type":"ContainerDied","Data":"3cbfb2f1687f144c2351411d6c0b4c77d822fda94c89c9e7bc97fda0c469eb0c"} Nov 21 14:19:35 crc kubenswrapper[4897]: I1121 14:19:35.178898 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7d6sl" event={"ID":"aee1eb81-536e-4dc3-be7f-f38a5b7efd83","Type":"ContainerStarted","Data":"3d70a3bcc27a289a3475ba9b78050266940930ae65ac30959754eafe464a3ef6"} Nov 21 14:19:35 crc kubenswrapper[4897]: I1121 14:19:35.178911 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7d6sl" event={"ID":"aee1eb81-536e-4dc3-be7f-f38a5b7efd83","Type":"ContainerStarted","Data":"fdc83822d5816f5e935ec81716651be8e8aed7ca18ba7a716559a5ff15290f30"} Nov 21 14:19:35 crc kubenswrapper[4897]: I1121 14:19:35.178921 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7d6sl" event={"ID":"aee1eb81-536e-4dc3-be7f-f38a5b7efd83","Type":"ContainerStarted","Data":"6fab7c22b0255cb6bd6a7e555114da675ee5dded506160d3f243699db8ebe2b8"} Nov 21 14:19:35 crc kubenswrapper[4897]: I1121 14:19:35.178932 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7d6sl" event={"ID":"aee1eb81-536e-4dc3-be7f-f38a5b7efd83","Type":"ContainerStarted","Data":"6ff005c2c333f64375980ea95fd9feddd0f7123076a2c256122f889e372719a7"} Nov 21 14:19:35 crc kubenswrapper[4897]: I1121 14:19:35.178941 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7d6sl" event={"ID":"aee1eb81-536e-4dc3-be7f-f38a5b7efd83","Type":"ContainerStarted","Data":"e3f18bd8d7a2c35e0cbb56144493e29204d0918692cde7655b1467ad3d752057"} Nov 21 14:19:35 crc kubenswrapper[4897]: I1121 14:19:35.178951 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7d6sl" event={"ID":"aee1eb81-536e-4dc3-be7f-f38a5b7efd83","Type":"ContainerStarted","Data":"6987a647bb82fabeeea0472483f2875d78577629695c3a7b73916f8471696c9b"} Nov 21 14:19:35 crc kubenswrapper[4897]: I1121 14:19:35.248551 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/eeb48e8d-4f9c-41a3-97e3-a3d7c2833341-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-6b8b5c98d-lt42q\" (UID: \"eeb48e8d-4f9c-41a3-97e3-a3d7c2833341\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b8b5c98d-lt42q" Nov 21 14:19:35 crc kubenswrapper[4897]: I1121 14:19:35.248605 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/aebcbb07-a9ed-4904-b53f-cae07b85a101-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-6b8b5c98d-trqb7\" (UID: \"aebcbb07-a9ed-4904-b53f-cae07b85a101\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b8b5c98d-trqb7" Nov 21 14:19:35 crc kubenswrapper[4897]: I1121 14:19:35.248630 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/aebcbb07-a9ed-4904-b53f-cae07b85a101-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-6b8b5c98d-trqb7\" (UID: \"aebcbb07-a9ed-4904-b53f-cae07b85a101\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b8b5c98d-trqb7" Nov 21 14:19:35 crc kubenswrapper[4897]: I1121 14:19:35.248742 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/eeb48e8d-4f9c-41a3-97e3-a3d7c2833341-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-6b8b5c98d-lt42q\" (UID: \"eeb48e8d-4f9c-41a3-97e3-a3d7c2833341\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b8b5c98d-lt42q" Nov 21 14:19:35 crc kubenswrapper[4897]: I1121 14:19:35.308717 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-s5lj9" Nov 21 14:19:35 crc kubenswrapper[4897]: I1121 14:19:35.327120 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-5bpbf"] Nov 21 14:19:35 crc kubenswrapper[4897]: I1121 14:19:35.327933 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-5bpbf" Nov 21 14:19:35 crc kubenswrapper[4897]: I1121 14:19:35.330462 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-tls" Nov 21 14:19:35 crc kubenswrapper[4897]: E1121 14:19:35.334444 4897 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-s5lj9_openshift-operators_a001ae14-634e-4590-8177-fc7db62c3247_0(f11aba79c7a9dbfbf3ae31ce034a4a2f566967fa28e6c7cebe8f24c71ba736a3): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 21 14:19:35 crc kubenswrapper[4897]: E1121 14:19:35.334524 4897 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-s5lj9_openshift-operators_a001ae14-634e-4590-8177-fc7db62c3247_0(f11aba79c7a9dbfbf3ae31ce034a4a2f566967fa28e6c7cebe8f24c71ba736a3): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-s5lj9" Nov 21 14:19:35 crc kubenswrapper[4897]: E1121 14:19:35.334546 4897 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-s5lj9_openshift-operators_a001ae14-634e-4590-8177-fc7db62c3247_0(f11aba79c7a9dbfbf3ae31ce034a4a2f566967fa28e6c7cebe8f24c71ba736a3): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-s5lj9" Nov 21 14:19:35 crc kubenswrapper[4897]: E1121 14:19:35.334595 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-668cf9dfbb-s5lj9_openshift-operators(a001ae14-634e-4590-8177-fc7db62c3247)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-668cf9dfbb-s5lj9_openshift-operators(a001ae14-634e-4590-8177-fc7db62c3247)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-s5lj9_openshift-operators_a001ae14-634e-4590-8177-fc7db62c3247_0(f11aba79c7a9dbfbf3ae31ce034a4a2f566967fa28e6c7cebe8f24c71ba736a3): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-s5lj9" podUID="a001ae14-634e-4590-8177-fc7db62c3247" Nov 21 14:19:35 crc kubenswrapper[4897]: I1121 14:19:35.334705 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-sa-dockercfg-mqzqw" Nov 21 14:19:35 crc kubenswrapper[4897]: I1121 14:19:35.351181 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/eeb48e8d-4f9c-41a3-97e3-a3d7c2833341-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-6b8b5c98d-lt42q\" (UID: \"eeb48e8d-4f9c-41a3-97e3-a3d7c2833341\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b8b5c98d-lt42q" Nov 21 14:19:35 crc kubenswrapper[4897]: I1121 14:19:35.351244 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/eeb48e8d-4f9c-41a3-97e3-a3d7c2833341-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-6b8b5c98d-lt42q\" (UID: \"eeb48e8d-4f9c-41a3-97e3-a3d7c2833341\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b8b5c98d-lt42q" Nov 21 14:19:35 crc kubenswrapper[4897]: I1121 14:19:35.351273 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/aebcbb07-a9ed-4904-b53f-cae07b85a101-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-6b8b5c98d-trqb7\" (UID: \"aebcbb07-a9ed-4904-b53f-cae07b85a101\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b8b5c98d-trqb7" Nov 21 14:19:35 crc kubenswrapper[4897]: I1121 14:19:35.351298 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/aebcbb07-a9ed-4904-b53f-cae07b85a101-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-6b8b5c98d-trqb7\" (UID: \"aebcbb07-a9ed-4904-b53f-cae07b85a101\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b8b5c98d-trqb7" Nov 21 14:19:35 crc kubenswrapper[4897]: I1121 14:19:35.362190 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/eeb48e8d-4f9c-41a3-97e3-a3d7c2833341-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-6b8b5c98d-lt42q\" (UID: \"eeb48e8d-4f9c-41a3-97e3-a3d7c2833341\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b8b5c98d-lt42q" Nov 21 14:19:35 crc kubenswrapper[4897]: I1121 14:19:35.364111 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/aebcbb07-a9ed-4904-b53f-cae07b85a101-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-6b8b5c98d-trqb7\" (UID: \"aebcbb07-a9ed-4904-b53f-cae07b85a101\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b8b5c98d-trqb7" Nov 21 14:19:35 crc kubenswrapper[4897]: I1121 14:19:35.365337 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/eeb48e8d-4f9c-41a3-97e3-a3d7c2833341-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-6b8b5c98d-lt42q\" (UID: \"eeb48e8d-4f9c-41a3-97e3-a3d7c2833341\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b8b5c98d-lt42q" Nov 21 14:19:35 crc kubenswrapper[4897]: I1121 14:19:35.366154 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/aebcbb07-a9ed-4904-b53f-cae07b85a101-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-6b8b5c98d-trqb7\" (UID: \"aebcbb07-a9ed-4904-b53f-cae07b85a101\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b8b5c98d-trqb7" Nov 21 14:19:35 crc kubenswrapper[4897]: I1121 14:19:35.437247 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b8b5c98d-trqb7" Nov 21 14:19:35 crc kubenswrapper[4897]: I1121 14:19:35.453644 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b8b5c98d-lt42q" Nov 21 14:19:35 crc kubenswrapper[4897]: I1121 14:19:35.454092 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9fdxh\" (UniqueName: \"kubernetes.io/projected/7ae559e4-6f4b-47ee-bf32-8ce2d7eb9741-kube-api-access-9fdxh\") pod \"observability-operator-d8bb48f5d-5bpbf\" (UID: \"7ae559e4-6f4b-47ee-bf32-8ce2d7eb9741\") " pod="openshift-operators/observability-operator-d8bb48f5d-5bpbf" Nov 21 14:19:35 crc kubenswrapper[4897]: I1121 14:19:35.454223 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/7ae559e4-6f4b-47ee-bf32-8ce2d7eb9741-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-5bpbf\" (UID: \"7ae559e4-6f4b-47ee-bf32-8ce2d7eb9741\") " pod="openshift-operators/observability-operator-d8bb48f5d-5bpbf" Nov 21 14:19:35 crc kubenswrapper[4897]: E1121 14:19:35.476204 4897 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6b8b5c98d-trqb7_openshift-operators_aebcbb07-a9ed-4904-b53f-cae07b85a101_0(a45e7e88ccc639f38a2518ea7b74ba5675c6e14f38aaf0975d56a583d03d13bf): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 21 14:19:35 crc kubenswrapper[4897]: E1121 14:19:35.476269 4897 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6b8b5c98d-trqb7_openshift-operators_aebcbb07-a9ed-4904-b53f-cae07b85a101_0(a45e7e88ccc639f38a2518ea7b74ba5675c6e14f38aaf0975d56a583d03d13bf): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b8b5c98d-trqb7" Nov 21 14:19:35 crc kubenswrapper[4897]: E1121 14:19:35.476299 4897 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6b8b5c98d-trqb7_openshift-operators_aebcbb07-a9ed-4904-b53f-cae07b85a101_0(a45e7e88ccc639f38a2518ea7b74ba5675c6e14f38aaf0975d56a583d03d13bf): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b8b5c98d-trqb7" Nov 21 14:19:35 crc kubenswrapper[4897]: E1121 14:19:35.476354 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-6b8b5c98d-trqb7_openshift-operators(aebcbb07-a9ed-4904-b53f-cae07b85a101)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-6b8b5c98d-trqb7_openshift-operators(aebcbb07-a9ed-4904-b53f-cae07b85a101)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6b8b5c98d-trqb7_openshift-operators_aebcbb07-a9ed-4904-b53f-cae07b85a101_0(a45e7e88ccc639f38a2518ea7b74ba5675c6e14f38aaf0975d56a583d03d13bf): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b8b5c98d-trqb7" podUID="aebcbb07-a9ed-4904-b53f-cae07b85a101" Nov 21 14:19:35 crc kubenswrapper[4897]: E1121 14:19:35.490807 4897 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6b8b5c98d-lt42q_openshift-operators_eeb48e8d-4f9c-41a3-97e3-a3d7c2833341_0(d07da5bcff950b6154b1e99edaaba721ab58d0690f383a291508f6106f44ba24): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 21 14:19:35 crc kubenswrapper[4897]: E1121 14:19:35.490878 4897 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6b8b5c98d-lt42q_openshift-operators_eeb48e8d-4f9c-41a3-97e3-a3d7c2833341_0(d07da5bcff950b6154b1e99edaaba721ab58d0690f383a291508f6106f44ba24): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b8b5c98d-lt42q" Nov 21 14:19:35 crc kubenswrapper[4897]: E1121 14:19:35.490898 4897 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6b8b5c98d-lt42q_openshift-operators_eeb48e8d-4f9c-41a3-97e3-a3d7c2833341_0(d07da5bcff950b6154b1e99edaaba721ab58d0690f383a291508f6106f44ba24): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b8b5c98d-lt42q" Nov 21 14:19:35 crc kubenswrapper[4897]: E1121 14:19:35.490947 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-6b8b5c98d-lt42q_openshift-operators(eeb48e8d-4f9c-41a3-97e3-a3d7c2833341)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-6b8b5c98d-lt42q_openshift-operators(eeb48e8d-4f9c-41a3-97e3-a3d7c2833341)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6b8b5c98d-lt42q_openshift-operators_eeb48e8d-4f9c-41a3-97e3-a3d7c2833341_0(d07da5bcff950b6154b1e99edaaba721ab58d0690f383a291508f6106f44ba24): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b8b5c98d-lt42q" podUID="eeb48e8d-4f9c-41a3-97e3-a3d7c2833341" Nov 21 14:19:35 crc kubenswrapper[4897]: I1121 14:19:35.532492 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/perses-operator-5446b9c989-q8rbz"] Nov 21 14:19:35 crc kubenswrapper[4897]: I1121 14:19:35.533451 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-q8rbz" Nov 21 14:19:35 crc kubenswrapper[4897]: I1121 14:19:35.535077 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"perses-operator-dockercfg-2qvg7" Nov 21 14:19:35 crc kubenswrapper[4897]: I1121 14:19:35.555778 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/7ae559e4-6f4b-47ee-bf32-8ce2d7eb9741-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-5bpbf\" (UID: \"7ae559e4-6f4b-47ee-bf32-8ce2d7eb9741\") " pod="openshift-operators/observability-operator-d8bb48f5d-5bpbf" Nov 21 14:19:35 crc kubenswrapper[4897]: I1121 14:19:35.555865 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9fdxh\" (UniqueName: \"kubernetes.io/projected/7ae559e4-6f4b-47ee-bf32-8ce2d7eb9741-kube-api-access-9fdxh\") pod \"observability-operator-d8bb48f5d-5bpbf\" (UID: \"7ae559e4-6f4b-47ee-bf32-8ce2d7eb9741\") " pod="openshift-operators/observability-operator-d8bb48f5d-5bpbf" Nov 21 14:19:35 crc kubenswrapper[4897]: I1121 14:19:35.561111 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/7ae559e4-6f4b-47ee-bf32-8ce2d7eb9741-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-5bpbf\" (UID: \"7ae559e4-6f4b-47ee-bf32-8ce2d7eb9741\") " pod="openshift-operators/observability-operator-d8bb48f5d-5bpbf" Nov 21 14:19:35 crc kubenswrapper[4897]: I1121 14:19:35.573705 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9fdxh\" (UniqueName: \"kubernetes.io/projected/7ae559e4-6f4b-47ee-bf32-8ce2d7eb9741-kube-api-access-9fdxh\") pod \"observability-operator-d8bb48f5d-5bpbf\" (UID: \"7ae559e4-6f4b-47ee-bf32-8ce2d7eb9741\") " pod="openshift-operators/observability-operator-d8bb48f5d-5bpbf" Nov 21 14:19:35 crc kubenswrapper[4897]: I1121 14:19:35.647698 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-5bpbf" Nov 21 14:19:35 crc kubenswrapper[4897]: I1121 14:19:35.657362 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/aa332b81-d1b0-4d8f-bbf1-8ad49072470e-openshift-service-ca\") pod \"perses-operator-5446b9c989-q8rbz\" (UID: \"aa332b81-d1b0-4d8f-bbf1-8ad49072470e\") " pod="openshift-operators/perses-operator-5446b9c989-q8rbz" Nov 21 14:19:35 crc kubenswrapper[4897]: I1121 14:19:35.657460 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-frn9t\" (UniqueName: \"kubernetes.io/projected/aa332b81-d1b0-4d8f-bbf1-8ad49072470e-kube-api-access-frn9t\") pod \"perses-operator-5446b9c989-q8rbz\" (UID: \"aa332b81-d1b0-4d8f-bbf1-8ad49072470e\") " pod="openshift-operators/perses-operator-5446b9c989-q8rbz" Nov 21 14:19:35 crc kubenswrapper[4897]: E1121 14:19:35.667204 4897 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-5bpbf_openshift-operators_7ae559e4-6f4b-47ee-bf32-8ce2d7eb9741_0(cecbe071b0da573e7c19ce2cff69cc06aa92675543fae3588492cc6eaf6c7b89): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 21 14:19:35 crc kubenswrapper[4897]: E1121 14:19:35.667247 4897 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-5bpbf_openshift-operators_7ae559e4-6f4b-47ee-bf32-8ce2d7eb9741_0(cecbe071b0da573e7c19ce2cff69cc06aa92675543fae3588492cc6eaf6c7b89): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-d8bb48f5d-5bpbf" Nov 21 14:19:35 crc kubenswrapper[4897]: E1121 14:19:35.667269 4897 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-5bpbf_openshift-operators_7ae559e4-6f4b-47ee-bf32-8ce2d7eb9741_0(cecbe071b0da573e7c19ce2cff69cc06aa92675543fae3588492cc6eaf6c7b89): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-d8bb48f5d-5bpbf" Nov 21 14:19:35 crc kubenswrapper[4897]: E1121 14:19:35.667306 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"observability-operator-d8bb48f5d-5bpbf_openshift-operators(7ae559e4-6f4b-47ee-bf32-8ce2d7eb9741)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"observability-operator-d8bb48f5d-5bpbf_openshift-operators(7ae559e4-6f4b-47ee-bf32-8ce2d7eb9741)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-5bpbf_openshift-operators_7ae559e4-6f4b-47ee-bf32-8ce2d7eb9741_0(cecbe071b0da573e7c19ce2cff69cc06aa92675543fae3588492cc6eaf6c7b89): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/observability-operator-d8bb48f5d-5bpbf" podUID="7ae559e4-6f4b-47ee-bf32-8ce2d7eb9741" Nov 21 14:19:35 crc kubenswrapper[4897]: I1121 14:19:35.759223 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-frn9t\" (UniqueName: \"kubernetes.io/projected/aa332b81-d1b0-4d8f-bbf1-8ad49072470e-kube-api-access-frn9t\") pod \"perses-operator-5446b9c989-q8rbz\" (UID: \"aa332b81-d1b0-4d8f-bbf1-8ad49072470e\") " pod="openshift-operators/perses-operator-5446b9c989-q8rbz" Nov 21 14:19:35 crc kubenswrapper[4897]: I1121 14:19:35.759303 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/aa332b81-d1b0-4d8f-bbf1-8ad49072470e-openshift-service-ca\") pod \"perses-operator-5446b9c989-q8rbz\" (UID: \"aa332b81-d1b0-4d8f-bbf1-8ad49072470e\") " pod="openshift-operators/perses-operator-5446b9c989-q8rbz" Nov 21 14:19:35 crc kubenswrapper[4897]: I1121 14:19:35.760164 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/aa332b81-d1b0-4d8f-bbf1-8ad49072470e-openshift-service-ca\") pod \"perses-operator-5446b9c989-q8rbz\" (UID: \"aa332b81-d1b0-4d8f-bbf1-8ad49072470e\") " pod="openshift-operators/perses-operator-5446b9c989-q8rbz" Nov 21 14:19:35 crc kubenswrapper[4897]: I1121 14:19:35.779537 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-frn9t\" (UniqueName: \"kubernetes.io/projected/aa332b81-d1b0-4d8f-bbf1-8ad49072470e-kube-api-access-frn9t\") pod \"perses-operator-5446b9c989-q8rbz\" (UID: \"aa332b81-d1b0-4d8f-bbf1-8ad49072470e\") " pod="openshift-operators/perses-operator-5446b9c989-q8rbz" Nov 21 14:19:35 crc kubenswrapper[4897]: I1121 14:19:35.849258 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-q8rbz" Nov 21 14:19:35 crc kubenswrapper[4897]: E1121 14:19:35.875627 4897 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-q8rbz_openshift-operators_aa332b81-d1b0-4d8f-bbf1-8ad49072470e_0(01eb943ff123ad680ae730cdb29aa8de98b1bc272a5cac738b304516bebe3494): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 21 14:19:35 crc kubenswrapper[4897]: E1121 14:19:35.875692 4897 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-q8rbz_openshift-operators_aa332b81-d1b0-4d8f-bbf1-8ad49072470e_0(01eb943ff123ad680ae730cdb29aa8de98b1bc272a5cac738b304516bebe3494): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5446b9c989-q8rbz" Nov 21 14:19:35 crc kubenswrapper[4897]: E1121 14:19:35.875712 4897 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-q8rbz_openshift-operators_aa332b81-d1b0-4d8f-bbf1-8ad49072470e_0(01eb943ff123ad680ae730cdb29aa8de98b1bc272a5cac738b304516bebe3494): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5446b9c989-q8rbz" Nov 21 14:19:35 crc kubenswrapper[4897]: E1121 14:19:35.875766 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"perses-operator-5446b9c989-q8rbz_openshift-operators(aa332b81-d1b0-4d8f-bbf1-8ad49072470e)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"perses-operator-5446b9c989-q8rbz_openshift-operators(aa332b81-d1b0-4d8f-bbf1-8ad49072470e)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-q8rbz_openshift-operators_aa332b81-d1b0-4d8f-bbf1-8ad49072470e_0(01eb943ff123ad680ae730cdb29aa8de98b1bc272a5cac738b304516bebe3494): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/perses-operator-5446b9c989-q8rbz" podUID="aa332b81-d1b0-4d8f-bbf1-8ad49072470e" Nov 21 14:19:36 crc kubenswrapper[4897]: I1121 14:19:36.097184 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="222f65f2-0b82-4760-bcf7-779244720f01" path="/var/lib/kubelet/pods/222f65f2-0b82-4760-bcf7-779244720f01/volumes" Nov 21 14:19:38 crc kubenswrapper[4897]: I1121 14:19:38.201902 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7d6sl" event={"ID":"aee1eb81-536e-4dc3-be7f-f38a5b7efd83","Type":"ContainerStarted","Data":"d62989198b0646c5eefe1247a4ff8b883bc591005b9a8f79a8ced6f02660aeaa"} Nov 21 14:19:40 crc kubenswrapper[4897]: I1121 14:19:40.215174 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7d6sl" event={"ID":"aee1eb81-536e-4dc3-be7f-f38a5b7efd83","Type":"ContainerStarted","Data":"c23e43a357c13307625236d59eda0b3a6f8e3021a7d7cc98ae69da361a04604b"} Nov 21 14:19:40 crc kubenswrapper[4897]: I1121 14:19:40.215708 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-7d6sl" Nov 21 14:19:40 crc kubenswrapper[4897]: I1121 14:19:40.215738 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-7d6sl" Nov 21 14:19:40 crc kubenswrapper[4897]: I1121 14:19:40.215746 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-7d6sl" Nov 21 14:19:40 crc kubenswrapper[4897]: I1121 14:19:40.255981 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-7d6sl" Nov 21 14:19:40 crc kubenswrapper[4897]: I1121 14:19:40.257425 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-7d6sl" Nov 21 14:19:40 crc kubenswrapper[4897]: I1121 14:19:40.261358 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-7d6sl" podStartSLOduration=7.261350044 podStartE2EDuration="7.261350044s" podCreationTimestamp="2025-11-21 14:19:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:19:40.260324816 +0000 UTC m=+657.544918291" watchObservedRunningTime="2025-11-21 14:19:40.261350044 +0000 UTC m=+657.545943519" Nov 21 14:19:40 crc kubenswrapper[4897]: I1121 14:19:40.595204 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-6b8b5c98d-trqb7"] Nov 21 14:19:40 crc kubenswrapper[4897]: I1121 14:19:40.595362 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b8b5c98d-trqb7" Nov 21 14:19:40 crc kubenswrapper[4897]: I1121 14:19:40.595882 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b8b5c98d-trqb7" Nov 21 14:19:40 crc kubenswrapper[4897]: I1121 14:19:40.600079 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-6b8b5c98d-lt42q"] Nov 21 14:19:40 crc kubenswrapper[4897]: I1121 14:19:40.600224 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b8b5c98d-lt42q" Nov 21 14:19:40 crc kubenswrapper[4897]: I1121 14:19:40.600819 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b8b5c98d-lt42q" Nov 21 14:19:40 crc kubenswrapper[4897]: I1121 14:19:40.606162 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5446b9c989-q8rbz"] Nov 21 14:19:40 crc kubenswrapper[4897]: I1121 14:19:40.606295 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-q8rbz" Nov 21 14:19:40 crc kubenswrapper[4897]: I1121 14:19:40.606785 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-q8rbz" Nov 21 14:19:40 crc kubenswrapper[4897]: I1121 14:19:40.640602 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-s5lj9"] Nov 21 14:19:40 crc kubenswrapper[4897]: I1121 14:19:40.640706 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-s5lj9" Nov 21 14:19:40 crc kubenswrapper[4897]: I1121 14:19:40.641133 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-s5lj9" Nov 21 14:19:40 crc kubenswrapper[4897]: E1121 14:19:40.641142 4897 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6b8b5c98d-trqb7_openshift-operators_aebcbb07-a9ed-4904-b53f-cae07b85a101_0(40bb7501503bb2031aea238ea365a8053fc4e02acdc2f854b6f603ebc7455d90): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 21 14:19:40 crc kubenswrapper[4897]: E1121 14:19:40.641203 4897 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6b8b5c98d-trqb7_openshift-operators_aebcbb07-a9ed-4904-b53f-cae07b85a101_0(40bb7501503bb2031aea238ea365a8053fc4e02acdc2f854b6f603ebc7455d90): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b8b5c98d-trqb7" Nov 21 14:19:40 crc kubenswrapper[4897]: E1121 14:19:40.641240 4897 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6b8b5c98d-trqb7_openshift-operators_aebcbb07-a9ed-4904-b53f-cae07b85a101_0(40bb7501503bb2031aea238ea365a8053fc4e02acdc2f854b6f603ebc7455d90): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b8b5c98d-trqb7" Nov 21 14:19:40 crc kubenswrapper[4897]: E1121 14:19:40.641316 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-6b8b5c98d-trqb7_openshift-operators(aebcbb07-a9ed-4904-b53f-cae07b85a101)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-6b8b5c98d-trqb7_openshift-operators(aebcbb07-a9ed-4904-b53f-cae07b85a101)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6b8b5c98d-trqb7_openshift-operators_aebcbb07-a9ed-4904-b53f-cae07b85a101_0(40bb7501503bb2031aea238ea365a8053fc4e02acdc2f854b6f603ebc7455d90): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b8b5c98d-trqb7" podUID="aebcbb07-a9ed-4904-b53f-cae07b85a101" Nov 21 14:19:40 crc kubenswrapper[4897]: I1121 14:19:40.644209 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-5bpbf"] Nov 21 14:19:40 crc kubenswrapper[4897]: I1121 14:19:40.644288 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-5bpbf" Nov 21 14:19:40 crc kubenswrapper[4897]: I1121 14:19:40.644559 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-5bpbf" Nov 21 14:19:40 crc kubenswrapper[4897]: E1121 14:19:40.653071 4897 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6b8b5c98d-lt42q_openshift-operators_eeb48e8d-4f9c-41a3-97e3-a3d7c2833341_0(b163bc8ae8d88190631bf72dc0d62c82dd39fcc6fa1f1180765ab4777a3667c5): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 21 14:19:40 crc kubenswrapper[4897]: E1121 14:19:40.653258 4897 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6b8b5c98d-lt42q_openshift-operators_eeb48e8d-4f9c-41a3-97e3-a3d7c2833341_0(b163bc8ae8d88190631bf72dc0d62c82dd39fcc6fa1f1180765ab4777a3667c5): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b8b5c98d-lt42q" Nov 21 14:19:40 crc kubenswrapper[4897]: E1121 14:19:40.653284 4897 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6b8b5c98d-lt42q_openshift-operators_eeb48e8d-4f9c-41a3-97e3-a3d7c2833341_0(b163bc8ae8d88190631bf72dc0d62c82dd39fcc6fa1f1180765ab4777a3667c5): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b8b5c98d-lt42q" Nov 21 14:19:40 crc kubenswrapper[4897]: E1121 14:19:40.653334 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-6b8b5c98d-lt42q_openshift-operators(eeb48e8d-4f9c-41a3-97e3-a3d7c2833341)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-6b8b5c98d-lt42q_openshift-operators(eeb48e8d-4f9c-41a3-97e3-a3d7c2833341)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6b8b5c98d-lt42q_openshift-operators_eeb48e8d-4f9c-41a3-97e3-a3d7c2833341_0(b163bc8ae8d88190631bf72dc0d62c82dd39fcc6fa1f1180765ab4777a3667c5): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b8b5c98d-lt42q" podUID="eeb48e8d-4f9c-41a3-97e3-a3d7c2833341" Nov 21 14:19:40 crc kubenswrapper[4897]: E1121 14:19:40.679656 4897 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-q8rbz_openshift-operators_aa332b81-d1b0-4d8f-bbf1-8ad49072470e_0(5e0e7f9e17102ce43e2170525b865f711cbb8623456ffd0308a334bddb99103c): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 21 14:19:40 crc kubenswrapper[4897]: E1121 14:19:40.679754 4897 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-q8rbz_openshift-operators_aa332b81-d1b0-4d8f-bbf1-8ad49072470e_0(5e0e7f9e17102ce43e2170525b865f711cbb8623456ffd0308a334bddb99103c): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5446b9c989-q8rbz" Nov 21 14:19:40 crc kubenswrapper[4897]: E1121 14:19:40.679788 4897 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-q8rbz_openshift-operators_aa332b81-d1b0-4d8f-bbf1-8ad49072470e_0(5e0e7f9e17102ce43e2170525b865f711cbb8623456ffd0308a334bddb99103c): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5446b9c989-q8rbz" Nov 21 14:19:40 crc kubenswrapper[4897]: E1121 14:19:40.679863 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"perses-operator-5446b9c989-q8rbz_openshift-operators(aa332b81-d1b0-4d8f-bbf1-8ad49072470e)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"perses-operator-5446b9c989-q8rbz_openshift-operators(aa332b81-d1b0-4d8f-bbf1-8ad49072470e)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-q8rbz_openshift-operators_aa332b81-d1b0-4d8f-bbf1-8ad49072470e_0(5e0e7f9e17102ce43e2170525b865f711cbb8623456ffd0308a334bddb99103c): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/perses-operator-5446b9c989-q8rbz" podUID="aa332b81-d1b0-4d8f-bbf1-8ad49072470e" Nov 21 14:19:40 crc kubenswrapper[4897]: E1121 14:19:40.694264 4897 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-s5lj9_openshift-operators_a001ae14-634e-4590-8177-fc7db62c3247_0(6d876b0001aa0a4f22ea1cb59863850b8bc14ba3d29d35fd530e204eb678d279): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 21 14:19:40 crc kubenswrapper[4897]: E1121 14:19:40.694319 4897 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-s5lj9_openshift-operators_a001ae14-634e-4590-8177-fc7db62c3247_0(6d876b0001aa0a4f22ea1cb59863850b8bc14ba3d29d35fd530e204eb678d279): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-s5lj9" Nov 21 14:19:40 crc kubenswrapper[4897]: E1121 14:19:40.694340 4897 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-s5lj9_openshift-operators_a001ae14-634e-4590-8177-fc7db62c3247_0(6d876b0001aa0a4f22ea1cb59863850b8bc14ba3d29d35fd530e204eb678d279): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-s5lj9" Nov 21 14:19:40 crc kubenswrapper[4897]: E1121 14:19:40.694388 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-668cf9dfbb-s5lj9_openshift-operators(a001ae14-634e-4590-8177-fc7db62c3247)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-668cf9dfbb-s5lj9_openshift-operators(a001ae14-634e-4590-8177-fc7db62c3247)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-s5lj9_openshift-operators_a001ae14-634e-4590-8177-fc7db62c3247_0(6d876b0001aa0a4f22ea1cb59863850b8bc14ba3d29d35fd530e204eb678d279): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-s5lj9" podUID="a001ae14-634e-4590-8177-fc7db62c3247" Nov 21 14:19:40 crc kubenswrapper[4897]: E1121 14:19:40.698872 4897 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-5bpbf_openshift-operators_7ae559e4-6f4b-47ee-bf32-8ce2d7eb9741_0(5b0d0736b242de5a564fb5b4e2f1e46aa0d6826913221bf810ad840241af94ce): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 21 14:19:40 crc kubenswrapper[4897]: E1121 14:19:40.698975 4897 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-5bpbf_openshift-operators_7ae559e4-6f4b-47ee-bf32-8ce2d7eb9741_0(5b0d0736b242de5a564fb5b4e2f1e46aa0d6826913221bf810ad840241af94ce): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-d8bb48f5d-5bpbf" Nov 21 14:19:40 crc kubenswrapper[4897]: E1121 14:19:40.699011 4897 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-5bpbf_openshift-operators_7ae559e4-6f4b-47ee-bf32-8ce2d7eb9741_0(5b0d0736b242de5a564fb5b4e2f1e46aa0d6826913221bf810ad840241af94ce): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-d8bb48f5d-5bpbf" Nov 21 14:19:40 crc kubenswrapper[4897]: E1121 14:19:40.699080 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"observability-operator-d8bb48f5d-5bpbf_openshift-operators(7ae559e4-6f4b-47ee-bf32-8ce2d7eb9741)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"observability-operator-d8bb48f5d-5bpbf_openshift-operators(7ae559e4-6f4b-47ee-bf32-8ce2d7eb9741)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-5bpbf_openshift-operators_7ae559e4-6f4b-47ee-bf32-8ce2d7eb9741_0(5b0d0736b242de5a564fb5b4e2f1e46aa0d6826913221bf810ad840241af94ce): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/observability-operator-d8bb48f5d-5bpbf" podUID="7ae559e4-6f4b-47ee-bf32-8ce2d7eb9741" Nov 21 14:19:45 crc kubenswrapper[4897]: I1121 14:19:45.088973 4897 scope.go:117] "RemoveContainer" containerID="382da75081f6e2afec6e1492c40398e4b099071e26f7162a289311aedd0d7777" Nov 21 14:19:45 crc kubenswrapper[4897]: E1121 14:19:45.089706 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-6ljgx_openshift-multus(5f3b70b5-9294-4f69-9528-500d28f34c89)\"" pod="openshift-multus/multus-6ljgx" podUID="5f3b70b5-9294-4f69-9528-500d28f34c89" Nov 21 14:19:52 crc kubenswrapper[4897]: I1121 14:19:52.088085 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-5bpbf" Nov 21 14:19:52 crc kubenswrapper[4897]: I1121 14:19:52.089222 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-5bpbf" Nov 21 14:19:52 crc kubenswrapper[4897]: E1121 14:19:52.119691 4897 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-5bpbf_openshift-operators_7ae559e4-6f4b-47ee-bf32-8ce2d7eb9741_0(590a04ea4eee74b31d701dff721df02f76ad39b5e230cbc38c32359d66d38cb5): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 21 14:19:52 crc kubenswrapper[4897]: E1121 14:19:52.119756 4897 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-5bpbf_openshift-operators_7ae559e4-6f4b-47ee-bf32-8ce2d7eb9741_0(590a04ea4eee74b31d701dff721df02f76ad39b5e230cbc38c32359d66d38cb5): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-d8bb48f5d-5bpbf" Nov 21 14:19:52 crc kubenswrapper[4897]: E1121 14:19:52.119778 4897 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-5bpbf_openshift-operators_7ae559e4-6f4b-47ee-bf32-8ce2d7eb9741_0(590a04ea4eee74b31d701dff721df02f76ad39b5e230cbc38c32359d66d38cb5): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-d8bb48f5d-5bpbf" Nov 21 14:19:52 crc kubenswrapper[4897]: E1121 14:19:52.119821 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"observability-operator-d8bb48f5d-5bpbf_openshift-operators(7ae559e4-6f4b-47ee-bf32-8ce2d7eb9741)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"observability-operator-d8bb48f5d-5bpbf_openshift-operators(7ae559e4-6f4b-47ee-bf32-8ce2d7eb9741)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-5bpbf_openshift-operators_7ae559e4-6f4b-47ee-bf32-8ce2d7eb9741_0(590a04ea4eee74b31d701dff721df02f76ad39b5e230cbc38c32359d66d38cb5): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/observability-operator-d8bb48f5d-5bpbf" podUID="7ae559e4-6f4b-47ee-bf32-8ce2d7eb9741" Nov 21 14:19:53 crc kubenswrapper[4897]: I1121 14:19:53.088621 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-q8rbz" Nov 21 14:19:53 crc kubenswrapper[4897]: I1121 14:19:53.089599 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-q8rbz" Nov 21 14:19:53 crc kubenswrapper[4897]: E1121 14:19:53.113296 4897 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-q8rbz_openshift-operators_aa332b81-d1b0-4d8f-bbf1-8ad49072470e_0(b7e21e4e0df0f668bee8b80ea1e020776f4b8c3a605b64f29c44e2d3a43cbf79): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 21 14:19:53 crc kubenswrapper[4897]: E1121 14:19:53.113417 4897 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-q8rbz_openshift-operators_aa332b81-d1b0-4d8f-bbf1-8ad49072470e_0(b7e21e4e0df0f668bee8b80ea1e020776f4b8c3a605b64f29c44e2d3a43cbf79): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5446b9c989-q8rbz" Nov 21 14:19:53 crc kubenswrapper[4897]: E1121 14:19:53.113448 4897 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-q8rbz_openshift-operators_aa332b81-d1b0-4d8f-bbf1-8ad49072470e_0(b7e21e4e0df0f668bee8b80ea1e020776f4b8c3a605b64f29c44e2d3a43cbf79): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5446b9c989-q8rbz" Nov 21 14:19:53 crc kubenswrapper[4897]: E1121 14:19:53.113532 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"perses-operator-5446b9c989-q8rbz_openshift-operators(aa332b81-d1b0-4d8f-bbf1-8ad49072470e)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"perses-operator-5446b9c989-q8rbz_openshift-operators(aa332b81-d1b0-4d8f-bbf1-8ad49072470e)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-q8rbz_openshift-operators_aa332b81-d1b0-4d8f-bbf1-8ad49072470e_0(b7e21e4e0df0f668bee8b80ea1e020776f4b8c3a605b64f29c44e2d3a43cbf79): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/perses-operator-5446b9c989-q8rbz" podUID="aa332b81-d1b0-4d8f-bbf1-8ad49072470e" Nov 21 14:19:54 crc kubenswrapper[4897]: I1121 14:19:54.087884 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-s5lj9" Nov 21 14:19:54 crc kubenswrapper[4897]: I1121 14:19:54.092480 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-s5lj9" Nov 21 14:19:54 crc kubenswrapper[4897]: E1121 14:19:54.112322 4897 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-s5lj9_openshift-operators_a001ae14-634e-4590-8177-fc7db62c3247_0(35679a9f2d1a2353a1a28ce7cda0e9b02e0ff44adb28ea87084ec004320daa5c): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 21 14:19:54 crc kubenswrapper[4897]: E1121 14:19:54.112396 4897 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-s5lj9_openshift-operators_a001ae14-634e-4590-8177-fc7db62c3247_0(35679a9f2d1a2353a1a28ce7cda0e9b02e0ff44adb28ea87084ec004320daa5c): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-s5lj9" Nov 21 14:19:54 crc kubenswrapper[4897]: E1121 14:19:54.112429 4897 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-s5lj9_openshift-operators_a001ae14-634e-4590-8177-fc7db62c3247_0(35679a9f2d1a2353a1a28ce7cda0e9b02e0ff44adb28ea87084ec004320daa5c): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-s5lj9" Nov 21 14:19:54 crc kubenswrapper[4897]: E1121 14:19:54.112492 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-668cf9dfbb-s5lj9_openshift-operators(a001ae14-634e-4590-8177-fc7db62c3247)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-668cf9dfbb-s5lj9_openshift-operators(a001ae14-634e-4590-8177-fc7db62c3247)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-s5lj9_openshift-operators_a001ae14-634e-4590-8177-fc7db62c3247_0(35679a9f2d1a2353a1a28ce7cda0e9b02e0ff44adb28ea87084ec004320daa5c): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-s5lj9" podUID="a001ae14-634e-4590-8177-fc7db62c3247" Nov 21 14:19:55 crc kubenswrapper[4897]: I1121 14:19:55.088593 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b8b5c98d-trqb7" Nov 21 14:19:55 crc kubenswrapper[4897]: I1121 14:19:55.088616 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b8b5c98d-lt42q" Nov 21 14:19:55 crc kubenswrapper[4897]: I1121 14:19:55.089099 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b8b5c98d-lt42q" Nov 21 14:19:55 crc kubenswrapper[4897]: I1121 14:19:55.089658 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b8b5c98d-trqb7" Nov 21 14:19:55 crc kubenswrapper[4897]: E1121 14:19:55.125282 4897 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6b8b5c98d-lt42q_openshift-operators_eeb48e8d-4f9c-41a3-97e3-a3d7c2833341_0(39e25af06a18be3c2e1767bcd7d98115fb17cbf6ab9f207088010a61c9ee9735): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 21 14:19:55 crc kubenswrapper[4897]: E1121 14:19:55.125356 4897 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6b8b5c98d-lt42q_openshift-operators_eeb48e8d-4f9c-41a3-97e3-a3d7c2833341_0(39e25af06a18be3c2e1767bcd7d98115fb17cbf6ab9f207088010a61c9ee9735): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b8b5c98d-lt42q" Nov 21 14:19:55 crc kubenswrapper[4897]: E1121 14:19:55.125385 4897 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6b8b5c98d-lt42q_openshift-operators_eeb48e8d-4f9c-41a3-97e3-a3d7c2833341_0(39e25af06a18be3c2e1767bcd7d98115fb17cbf6ab9f207088010a61c9ee9735): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b8b5c98d-lt42q" Nov 21 14:19:55 crc kubenswrapper[4897]: E1121 14:19:55.125440 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-6b8b5c98d-lt42q_openshift-operators(eeb48e8d-4f9c-41a3-97e3-a3d7c2833341)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-6b8b5c98d-lt42q_openshift-operators(eeb48e8d-4f9c-41a3-97e3-a3d7c2833341)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6b8b5c98d-lt42q_openshift-operators_eeb48e8d-4f9c-41a3-97e3-a3d7c2833341_0(39e25af06a18be3c2e1767bcd7d98115fb17cbf6ab9f207088010a61c9ee9735): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b8b5c98d-lt42q" podUID="eeb48e8d-4f9c-41a3-97e3-a3d7c2833341" Nov 21 14:19:55 crc kubenswrapper[4897]: E1121 14:19:55.126353 4897 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6b8b5c98d-trqb7_openshift-operators_aebcbb07-a9ed-4904-b53f-cae07b85a101_0(841c97ec2d6aa8fbc05d4d59f46b5d5e47d1f5f41a527ea4d8bad839920db857): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 21 14:19:55 crc kubenswrapper[4897]: E1121 14:19:55.126426 4897 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6b8b5c98d-trqb7_openshift-operators_aebcbb07-a9ed-4904-b53f-cae07b85a101_0(841c97ec2d6aa8fbc05d4d59f46b5d5e47d1f5f41a527ea4d8bad839920db857): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b8b5c98d-trqb7" Nov 21 14:19:55 crc kubenswrapper[4897]: E1121 14:19:55.126453 4897 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6b8b5c98d-trqb7_openshift-operators_aebcbb07-a9ed-4904-b53f-cae07b85a101_0(841c97ec2d6aa8fbc05d4d59f46b5d5e47d1f5f41a527ea4d8bad839920db857): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b8b5c98d-trqb7" Nov 21 14:19:55 crc kubenswrapper[4897]: E1121 14:19:55.126522 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-6b8b5c98d-trqb7_openshift-operators(aebcbb07-a9ed-4904-b53f-cae07b85a101)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-6b8b5c98d-trqb7_openshift-operators(aebcbb07-a9ed-4904-b53f-cae07b85a101)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6b8b5c98d-trqb7_openshift-operators_aebcbb07-a9ed-4904-b53f-cae07b85a101_0(841c97ec2d6aa8fbc05d4d59f46b5d5e47d1f5f41a527ea4d8bad839920db857): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b8b5c98d-trqb7" podUID="aebcbb07-a9ed-4904-b53f-cae07b85a101" Nov 21 14:19:57 crc kubenswrapper[4897]: I1121 14:19:57.089025 4897 scope.go:117] "RemoveContainer" containerID="382da75081f6e2afec6e1492c40398e4b099071e26f7162a289311aedd0d7777" Nov 21 14:19:58 crc kubenswrapper[4897]: I1121 14:19:58.310710 4897 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-6ljgx_5f3b70b5-9294-4f69-9528-500d28f34c89/kube-multus/2.log" Nov 21 14:19:58 crc kubenswrapper[4897]: I1121 14:19:58.311041 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-6ljgx" event={"ID":"5f3b70b5-9294-4f69-9528-500d28f34c89","Type":"ContainerStarted","Data":"49e7ca58174e430ad97032f89c6d426277e181ef9903760a785e09dcd60158ca"} Nov 21 14:20:04 crc kubenswrapper[4897]: I1121 14:20:04.065314 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-7d6sl" Nov 21 14:20:04 crc kubenswrapper[4897]: I1121 14:20:04.088232 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-q8rbz" Nov 21 14:20:04 crc kubenswrapper[4897]: I1121 14:20:04.088272 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-5bpbf" Nov 21 14:20:04 crc kubenswrapper[4897]: I1121 14:20:04.088856 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-q8rbz" Nov 21 14:20:04 crc kubenswrapper[4897]: I1121 14:20:04.089279 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-5bpbf" Nov 21 14:20:04 crc kubenswrapper[4897]: I1121 14:20:04.358978 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-5bpbf"] Nov 21 14:20:04 crc kubenswrapper[4897]: I1121 14:20:04.398636 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5446b9c989-q8rbz"] Nov 21 14:20:05 crc kubenswrapper[4897]: I1121 14:20:05.368165 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-d8bb48f5d-5bpbf" event={"ID":"7ae559e4-6f4b-47ee-bf32-8ce2d7eb9741","Type":"ContainerStarted","Data":"f3fd77f0d92476df0544c981f7176e82b3919833744a623cbc5856c1574cb883"} Nov 21 14:20:05 crc kubenswrapper[4897]: I1121 14:20:05.369552 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5446b9c989-q8rbz" event={"ID":"aa332b81-d1b0-4d8f-bbf1-8ad49072470e","Type":"ContainerStarted","Data":"691fbe3afcb7209b4486f58e9909b5138ff0579d38d83dd154648dfd726ae7ad"} Nov 21 14:20:06 crc kubenswrapper[4897]: I1121 14:20:06.088492 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b8b5c98d-trqb7" Nov 21 14:20:06 crc kubenswrapper[4897]: I1121 14:20:06.089367 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b8b5c98d-trqb7" Nov 21 14:20:06 crc kubenswrapper[4897]: I1121 14:20:06.495052 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-6b8b5c98d-trqb7"] Nov 21 14:20:06 crc kubenswrapper[4897]: W1121 14:20:06.499659 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaebcbb07_a9ed_4904_b53f_cae07b85a101.slice/crio-e0ab4326fb050b1dffb2f60e4ac5b91e30298a367bc45727f775226a2a5e5e51 WatchSource:0}: Error finding container e0ab4326fb050b1dffb2f60e4ac5b91e30298a367bc45727f775226a2a5e5e51: Status 404 returned error can't find the container with id e0ab4326fb050b1dffb2f60e4ac5b91e30298a367bc45727f775226a2a5e5e51 Nov 21 14:20:07 crc kubenswrapper[4897]: I1121 14:20:07.088655 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b8b5c98d-lt42q" Nov 21 14:20:07 crc kubenswrapper[4897]: I1121 14:20:07.089448 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b8b5c98d-lt42q" Nov 21 14:20:07 crc kubenswrapper[4897]: I1121 14:20:07.384498 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b8b5c98d-trqb7" event={"ID":"aebcbb07-a9ed-4904-b53f-cae07b85a101","Type":"ContainerStarted","Data":"e0ab4326fb050b1dffb2f60e4ac5b91e30298a367bc45727f775226a2a5e5e51"} Nov 21 14:20:08 crc kubenswrapper[4897]: I1121 14:20:08.088982 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-s5lj9" Nov 21 14:20:08 crc kubenswrapper[4897]: I1121 14:20:08.089817 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-s5lj9" Nov 21 14:20:12 crc kubenswrapper[4897]: I1121 14:20:12.915432 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-s5lj9"] Nov 21 14:20:12 crc kubenswrapper[4897]: W1121 14:20:12.917731 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda001ae14_634e_4590_8177_fc7db62c3247.slice/crio-bf67b79a597e254552d3177fc4fdd84b53b213d6f56f7e9092d76b27384ec8ee WatchSource:0}: Error finding container bf67b79a597e254552d3177fc4fdd84b53b213d6f56f7e9092d76b27384ec8ee: Status 404 returned error can't find the container with id bf67b79a597e254552d3177fc4fdd84b53b213d6f56f7e9092d76b27384ec8ee Nov 21 14:20:12 crc kubenswrapper[4897]: I1121 14:20:12.968007 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-6b8b5c98d-lt42q"] Nov 21 14:20:12 crc kubenswrapper[4897]: W1121 14:20:12.974966 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeeb48e8d_4f9c_41a3_97e3_a3d7c2833341.slice/crio-409ca47219ed8a1c14ed0d451cb0f6084ef7db69125863bd736927422c104c53 WatchSource:0}: Error finding container 409ca47219ed8a1c14ed0d451cb0f6084ef7db69125863bd736927422c104c53: Status 404 returned error can't find the container with id 409ca47219ed8a1c14ed0d451cb0f6084ef7db69125863bd736927422c104c53 Nov 21 14:20:13 crc kubenswrapper[4897]: I1121 14:20:13.423842 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b8b5c98d-trqb7" event={"ID":"aebcbb07-a9ed-4904-b53f-cae07b85a101","Type":"ContainerStarted","Data":"8443a175fa155a6a17c8bbb26821018e161eedb98620cd6ef2a753d09232ff4f"} Nov 21 14:20:13 crc kubenswrapper[4897]: I1121 14:20:13.425792 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-d8bb48f5d-5bpbf" event={"ID":"7ae559e4-6f4b-47ee-bf32-8ce2d7eb9741","Type":"ContainerStarted","Data":"56201688de4f974b7090856e96732bb58e251c273ec6a68eb809b4300a65ff3c"} Nov 21 14:20:13 crc kubenswrapper[4897]: I1121 14:20:13.425967 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/observability-operator-d8bb48f5d-5bpbf" Nov 21 14:20:13 crc kubenswrapper[4897]: I1121 14:20:13.436466 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-d8bb48f5d-5bpbf" Nov 21 14:20:13 crc kubenswrapper[4897]: I1121 14:20:13.437066 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-s5lj9" event={"ID":"a001ae14-634e-4590-8177-fc7db62c3247","Type":"ContainerStarted","Data":"bf67b79a597e254552d3177fc4fdd84b53b213d6f56f7e9092d76b27384ec8ee"} Nov 21 14:20:13 crc kubenswrapper[4897]: I1121 14:20:13.440032 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b8b5c98d-lt42q" event={"ID":"eeb48e8d-4f9c-41a3-97e3-a3d7c2833341","Type":"ContainerStarted","Data":"c400c7c6ba31d9fbf580fc94e240a0e7eea59e0ae5ff6a83f69c8e7fc5757d9c"} Nov 21 14:20:13 crc kubenswrapper[4897]: I1121 14:20:13.440080 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b8b5c98d-lt42q" event={"ID":"eeb48e8d-4f9c-41a3-97e3-a3d7c2833341","Type":"ContainerStarted","Data":"409ca47219ed8a1c14ed0d451cb0f6084ef7db69125863bd736927422c104c53"} Nov 21 14:20:13 crc kubenswrapper[4897]: I1121 14:20:13.443273 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5446b9c989-q8rbz" event={"ID":"aa332b81-d1b0-4d8f-bbf1-8ad49072470e","Type":"ContainerStarted","Data":"e01e729525dbff76bc614d7145520af69a3eadd69690b1938b836bbf9eaa303b"} Nov 21 14:20:13 crc kubenswrapper[4897]: I1121 14:20:13.443700 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/perses-operator-5446b9c989-q8rbz" Nov 21 14:20:13 crc kubenswrapper[4897]: I1121 14:20:13.453517 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b8b5c98d-trqb7" podStartSLOduration=32.301821703 podStartE2EDuration="38.453482491s" podCreationTimestamp="2025-11-21 14:19:35 +0000 UTC" firstStartedPulling="2025-11-21 14:20:06.501680387 +0000 UTC m=+683.786273862" lastFinishedPulling="2025-11-21 14:20:12.653341175 +0000 UTC m=+689.937934650" observedRunningTime="2025-11-21 14:20:13.45267168 +0000 UTC m=+690.737265165" watchObservedRunningTime="2025-11-21 14:20:13.453482491 +0000 UTC m=+690.738075966" Nov 21 14:20:13 crc kubenswrapper[4897]: I1121 14:20:13.491885 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/perses-operator-5446b9c989-q8rbz" podStartSLOduration=30.382060822 podStartE2EDuration="38.491867065s" podCreationTimestamp="2025-11-21 14:19:35 +0000 UTC" firstStartedPulling="2025-11-21 14:20:04.411142298 +0000 UTC m=+681.695735773" lastFinishedPulling="2025-11-21 14:20:12.520948501 +0000 UTC m=+689.805542016" observedRunningTime="2025-11-21 14:20:13.478116155 +0000 UTC m=+690.762709660" watchObservedRunningTime="2025-11-21 14:20:13.491867065 +0000 UTC m=+690.776460540" Nov 21 14:20:13 crc kubenswrapper[4897]: I1121 14:20:13.514680 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6b8b5c98d-lt42q" podStartSLOduration=38.514661308 podStartE2EDuration="38.514661308s" podCreationTimestamp="2025-11-21 14:19:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:20:13.509556251 +0000 UTC m=+690.794149726" watchObservedRunningTime="2025-11-21 14:20:13.514661308 +0000 UTC m=+690.799254803" Nov 21 14:20:13 crc kubenswrapper[4897]: I1121 14:20:13.542963 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-operator-d8bb48f5d-5bpbf" podStartSLOduration=30.390009605 podStartE2EDuration="38.542944599s" podCreationTimestamp="2025-11-21 14:19:35 +0000 UTC" firstStartedPulling="2025-11-21 14:20:04.3681185 +0000 UTC m=+681.652711975" lastFinishedPulling="2025-11-21 14:20:12.521053464 +0000 UTC m=+689.805646969" observedRunningTime="2025-11-21 14:20:13.538090309 +0000 UTC m=+690.822683804" watchObservedRunningTime="2025-11-21 14:20:13.542944599 +0000 UTC m=+690.827538074" Nov 21 14:20:16 crc kubenswrapper[4897]: I1121 14:20:16.468718 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-s5lj9" event={"ID":"a001ae14-634e-4590-8177-fc7db62c3247","Type":"ContainerStarted","Data":"a8ccaa368dae63d5be9eea20862898f045179b19c738ebe59925ef206b685266"} Nov 21 14:20:16 crc kubenswrapper[4897]: I1121 14:20:16.496913 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-s5lj9" podStartSLOduration=40.15667906 podStartE2EDuration="42.496895268s" podCreationTimestamp="2025-11-21 14:19:34 +0000 UTC" firstStartedPulling="2025-11-21 14:20:12.920819805 +0000 UTC m=+690.205413280" lastFinishedPulling="2025-11-21 14:20:15.261035983 +0000 UTC m=+692.545629488" observedRunningTime="2025-11-21 14:20:16.490418273 +0000 UTC m=+693.775011768" watchObservedRunningTime="2025-11-21 14:20:16.496895268 +0000 UTC m=+693.781488733" Nov 21 14:20:19 crc kubenswrapper[4897]: I1121 14:20:19.509996 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-msgn9"] Nov 21 14:20:19 crc kubenswrapper[4897]: I1121 14:20:19.511366 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-msgn9" Nov 21 14:20:19 crc kubenswrapper[4897]: I1121 14:20:19.518313 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Nov 21 14:20:19 crc kubenswrapper[4897]: I1121 14:20:19.525542 4897 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-wkqwl" Nov 21 14:20:19 crc kubenswrapper[4897]: I1121 14:20:19.525870 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Nov 21 14:20:19 crc kubenswrapper[4897]: I1121 14:20:19.527829 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-dxlpp"] Nov 21 14:20:19 crc kubenswrapper[4897]: I1121 14:20:19.534843 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-dxlpp" Nov 21 14:20:19 crc kubenswrapper[4897]: I1121 14:20:19.537145 4897 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-cn67q" Nov 21 14:20:19 crc kubenswrapper[4897]: I1121 14:20:19.538486 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-msgn9"] Nov 21 14:20:19 crc kubenswrapper[4897]: I1121 14:20:19.548457 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-dxlpp"] Nov 21 14:20:19 crc kubenswrapper[4897]: I1121 14:20:19.553223 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-2px7w"] Nov 21 14:20:19 crc kubenswrapper[4897]: I1121 14:20:19.555745 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-2px7w" Nov 21 14:20:19 crc kubenswrapper[4897]: I1121 14:20:19.556742 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-2px7w"] Nov 21 14:20:19 crc kubenswrapper[4897]: I1121 14:20:19.558004 4897 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-cpz44" Nov 21 14:20:19 crc kubenswrapper[4897]: I1121 14:20:19.631907 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tnhc5\" (UniqueName: \"kubernetes.io/projected/8249503c-57a3-4bf9-89da-cfb084c8136d-kube-api-access-tnhc5\") pod \"cert-manager-cainjector-7f985d654d-msgn9\" (UID: \"8249503c-57a3-4bf9-89da-cfb084c8136d\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-msgn9" Nov 21 14:20:19 crc kubenswrapper[4897]: I1121 14:20:19.631979 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9xkdj\" (UniqueName: \"kubernetes.io/projected/4dc91e18-7227-40bf-b774-4f6af8e121d5-kube-api-access-9xkdj\") pod \"cert-manager-5b446d88c5-dxlpp\" (UID: \"4dc91e18-7227-40bf-b774-4f6af8e121d5\") " pod="cert-manager/cert-manager-5b446d88c5-dxlpp" Nov 21 14:20:19 crc kubenswrapper[4897]: I1121 14:20:19.733980 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tnhc5\" (UniqueName: \"kubernetes.io/projected/8249503c-57a3-4bf9-89da-cfb084c8136d-kube-api-access-tnhc5\") pod \"cert-manager-cainjector-7f985d654d-msgn9\" (UID: \"8249503c-57a3-4bf9-89da-cfb084c8136d\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-msgn9" Nov 21 14:20:19 crc kubenswrapper[4897]: I1121 14:20:19.734093 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9xkdj\" (UniqueName: \"kubernetes.io/projected/4dc91e18-7227-40bf-b774-4f6af8e121d5-kube-api-access-9xkdj\") pod \"cert-manager-5b446d88c5-dxlpp\" (UID: \"4dc91e18-7227-40bf-b774-4f6af8e121d5\") " pod="cert-manager/cert-manager-5b446d88c5-dxlpp" Nov 21 14:20:19 crc kubenswrapper[4897]: I1121 14:20:19.734191 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n8kt4\" (UniqueName: \"kubernetes.io/projected/1952ec90-d336-432a-8726-b79b839a884c-kube-api-access-n8kt4\") pod \"cert-manager-webhook-5655c58dd6-2px7w\" (UID: \"1952ec90-d336-432a-8726-b79b839a884c\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-2px7w" Nov 21 14:20:19 crc kubenswrapper[4897]: I1121 14:20:19.753422 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9xkdj\" (UniqueName: \"kubernetes.io/projected/4dc91e18-7227-40bf-b774-4f6af8e121d5-kube-api-access-9xkdj\") pod \"cert-manager-5b446d88c5-dxlpp\" (UID: \"4dc91e18-7227-40bf-b774-4f6af8e121d5\") " pod="cert-manager/cert-manager-5b446d88c5-dxlpp" Nov 21 14:20:19 crc kubenswrapper[4897]: I1121 14:20:19.753428 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tnhc5\" (UniqueName: \"kubernetes.io/projected/8249503c-57a3-4bf9-89da-cfb084c8136d-kube-api-access-tnhc5\") pod \"cert-manager-cainjector-7f985d654d-msgn9\" (UID: \"8249503c-57a3-4bf9-89da-cfb084c8136d\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-msgn9" Nov 21 14:20:19 crc kubenswrapper[4897]: I1121 14:20:19.841197 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n8kt4\" (UniqueName: \"kubernetes.io/projected/1952ec90-d336-432a-8726-b79b839a884c-kube-api-access-n8kt4\") pod \"cert-manager-webhook-5655c58dd6-2px7w\" (UID: \"1952ec90-d336-432a-8726-b79b839a884c\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-2px7w" Nov 21 14:20:19 crc kubenswrapper[4897]: I1121 14:20:19.841722 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-msgn9" Nov 21 14:20:19 crc kubenswrapper[4897]: I1121 14:20:19.855974 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-dxlpp" Nov 21 14:20:19 crc kubenswrapper[4897]: I1121 14:20:19.861872 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n8kt4\" (UniqueName: \"kubernetes.io/projected/1952ec90-d336-432a-8726-b79b839a884c-kube-api-access-n8kt4\") pod \"cert-manager-webhook-5655c58dd6-2px7w\" (UID: \"1952ec90-d336-432a-8726-b79b839a884c\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-2px7w" Nov 21 14:20:19 crc kubenswrapper[4897]: I1121 14:20:19.874249 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-2px7w" Nov 21 14:20:20 crc kubenswrapper[4897]: I1121 14:20:20.338549 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-msgn9"] Nov 21 14:20:20 crc kubenswrapper[4897]: I1121 14:20:20.414386 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-dxlpp"] Nov 21 14:20:20 crc kubenswrapper[4897]: W1121 14:20:20.418778 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1952ec90_d336_432a_8726_b79b839a884c.slice/crio-230ff89f1c9ad8e431f908cd23bfa3e7b9e8344aad9f50874866794d322d96b8 WatchSource:0}: Error finding container 230ff89f1c9ad8e431f908cd23bfa3e7b9e8344aad9f50874866794d322d96b8: Status 404 returned error can't find the container with id 230ff89f1c9ad8e431f908cd23bfa3e7b9e8344aad9f50874866794d322d96b8 Nov 21 14:20:20 crc kubenswrapper[4897]: I1121 14:20:20.424819 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-2px7w"] Nov 21 14:20:20 crc kubenswrapper[4897]: I1121 14:20:20.498180 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-msgn9" event={"ID":"8249503c-57a3-4bf9-89da-cfb084c8136d","Type":"ContainerStarted","Data":"afbf3fb26ef8afe251a23e1bb058fa194ec87a3c6880f2558cc40092e8fb7fa5"} Nov 21 14:20:20 crc kubenswrapper[4897]: I1121 14:20:20.499393 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-2px7w" event={"ID":"1952ec90-d336-432a-8726-b79b839a884c","Type":"ContainerStarted","Data":"230ff89f1c9ad8e431f908cd23bfa3e7b9e8344aad9f50874866794d322d96b8"} Nov 21 14:20:20 crc kubenswrapper[4897]: I1121 14:20:20.500585 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-dxlpp" event={"ID":"4dc91e18-7227-40bf-b774-4f6af8e121d5","Type":"ContainerStarted","Data":"4c8834590ab69d55950efb7efbb48c3281c153ad88c9fac888126ccc969177fc"} Nov 21 14:20:24 crc kubenswrapper[4897]: I1121 14:20:24.530494 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-2px7w" event={"ID":"1952ec90-d336-432a-8726-b79b839a884c","Type":"ContainerStarted","Data":"f50bc68ce9acf4c5b262bdc0c4a244c8ce11b09753f461715ca289daf60d528a"} Nov 21 14:20:24 crc kubenswrapper[4897]: I1121 14:20:24.532191 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-2px7w" Nov 21 14:20:24 crc kubenswrapper[4897]: I1121 14:20:24.532544 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-dxlpp" event={"ID":"4dc91e18-7227-40bf-b774-4f6af8e121d5","Type":"ContainerStarted","Data":"f7da8cb8e00704fd98c84fb416619e304771648808d49d08476096a3e95e6db3"} Nov 21 14:20:24 crc kubenswrapper[4897]: I1121 14:20:24.533977 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-msgn9" event={"ID":"8249503c-57a3-4bf9-89da-cfb084c8136d","Type":"ContainerStarted","Data":"75b81e7bd73a1c45f7782eb16b4a99c04867a7288f82c436de4fa6445a9243e0"} Nov 21 14:20:24 crc kubenswrapper[4897]: I1121 14:20:24.547828 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-2px7w" podStartSLOduration=1.862028949 podStartE2EDuration="5.547807555s" podCreationTimestamp="2025-11-21 14:20:19 +0000 UTC" firstStartedPulling="2025-11-21 14:20:20.421786969 +0000 UTC m=+697.706380444" lastFinishedPulling="2025-11-21 14:20:24.107565565 +0000 UTC m=+701.392159050" observedRunningTime="2025-11-21 14:20:24.544668221 +0000 UTC m=+701.829261706" watchObservedRunningTime="2025-11-21 14:20:24.547807555 +0000 UTC m=+701.832401030" Nov 21 14:20:24 crc kubenswrapper[4897]: I1121 14:20:24.556336 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-msgn9" podStartSLOduration=1.85575614 podStartE2EDuration="5.556313524s" podCreationTimestamp="2025-11-21 14:20:19 +0000 UTC" firstStartedPulling="2025-11-21 14:20:20.346738079 +0000 UTC m=+697.631331554" lastFinishedPulling="2025-11-21 14:20:24.047295463 +0000 UTC m=+701.331888938" observedRunningTime="2025-11-21 14:20:24.555610845 +0000 UTC m=+701.840204330" watchObservedRunningTime="2025-11-21 14:20:24.556313524 +0000 UTC m=+701.840906999" Nov 21 14:20:24 crc kubenswrapper[4897]: I1121 14:20:24.567558 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-dxlpp" podStartSLOduration=1.921663194 podStartE2EDuration="5.567538816s" podCreationTimestamp="2025-11-21 14:20:19 +0000 UTC" firstStartedPulling="2025-11-21 14:20:20.401417711 +0000 UTC m=+697.686011216" lastFinishedPulling="2025-11-21 14:20:24.047293353 +0000 UTC m=+701.331886838" observedRunningTime="2025-11-21 14:20:24.566334974 +0000 UTC m=+701.850928459" watchObservedRunningTime="2025-11-21 14:20:24.567538816 +0000 UTC m=+701.852132291" Nov 21 14:20:25 crc kubenswrapper[4897]: I1121 14:20:25.853146 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/perses-operator-5446b9c989-q8rbz" Nov 21 14:20:29 crc kubenswrapper[4897]: I1121 14:20:29.876851 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-2px7w" Nov 21 14:20:53 crc kubenswrapper[4897]: I1121 14:20:53.786227 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fknv8t"] Nov 21 14:20:53 crc kubenswrapper[4897]: I1121 14:20:53.788392 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fknv8t" Nov 21 14:20:53 crc kubenswrapper[4897]: I1121 14:20:53.793562 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 21 14:20:53 crc kubenswrapper[4897]: I1121 14:20:53.801465 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fknv8t"] Nov 21 14:20:53 crc kubenswrapper[4897]: I1121 14:20:53.881047 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1832f955-8108-4635-b24e-5988ca72ca44-bundle\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fknv8t\" (UID: \"1832f955-8108-4635-b24e-5988ca72ca44\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fknv8t" Nov 21 14:20:53 crc kubenswrapper[4897]: I1121 14:20:53.881113 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qfk9c\" (UniqueName: \"kubernetes.io/projected/1832f955-8108-4635-b24e-5988ca72ca44-kube-api-access-qfk9c\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fknv8t\" (UID: \"1832f955-8108-4635-b24e-5988ca72ca44\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fknv8t" Nov 21 14:20:53 crc kubenswrapper[4897]: I1121 14:20:53.881236 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1832f955-8108-4635-b24e-5988ca72ca44-util\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fknv8t\" (UID: \"1832f955-8108-4635-b24e-5988ca72ca44\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fknv8t" Nov 21 14:20:53 crc kubenswrapper[4897]: I1121 14:20:53.974647 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb825xdk"] Nov 21 14:20:53 crc kubenswrapper[4897]: I1121 14:20:53.975841 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb825xdk" Nov 21 14:20:53 crc kubenswrapper[4897]: I1121 14:20:53.982959 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qfk9c\" (UniqueName: \"kubernetes.io/projected/1832f955-8108-4635-b24e-5988ca72ca44-kube-api-access-qfk9c\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fknv8t\" (UID: \"1832f955-8108-4635-b24e-5988ca72ca44\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fknv8t" Nov 21 14:20:53 crc kubenswrapper[4897]: I1121 14:20:53.983176 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1832f955-8108-4635-b24e-5988ca72ca44-util\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fknv8t\" (UID: \"1832f955-8108-4635-b24e-5988ca72ca44\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fknv8t" Nov 21 14:20:53 crc kubenswrapper[4897]: I1121 14:20:53.983257 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1832f955-8108-4635-b24e-5988ca72ca44-bundle\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fknv8t\" (UID: \"1832f955-8108-4635-b24e-5988ca72ca44\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fknv8t" Nov 21 14:20:53 crc kubenswrapper[4897]: I1121 14:20:53.983810 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1832f955-8108-4635-b24e-5988ca72ca44-util\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fknv8t\" (UID: \"1832f955-8108-4635-b24e-5988ca72ca44\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fknv8t" Nov 21 14:20:53 crc kubenswrapper[4897]: I1121 14:20:53.983998 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1832f955-8108-4635-b24e-5988ca72ca44-bundle\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fknv8t\" (UID: \"1832f955-8108-4635-b24e-5988ca72ca44\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fknv8t" Nov 21 14:20:53 crc kubenswrapper[4897]: I1121 14:20:53.992972 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb825xdk"] Nov 21 14:20:54 crc kubenswrapper[4897]: I1121 14:20:54.007450 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qfk9c\" (UniqueName: \"kubernetes.io/projected/1832f955-8108-4635-b24e-5988ca72ca44-kube-api-access-qfk9c\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fknv8t\" (UID: \"1832f955-8108-4635-b24e-5988ca72ca44\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fknv8t" Nov 21 14:20:54 crc kubenswrapper[4897]: I1121 14:20:54.084465 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c56fa316-41a4-4425-9fe2-e21746f35514-util\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb825xdk\" (UID: \"c56fa316-41a4-4425-9fe2-e21746f35514\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb825xdk" Nov 21 14:20:54 crc kubenswrapper[4897]: I1121 14:20:54.084807 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dxm9b\" (UniqueName: \"kubernetes.io/projected/c56fa316-41a4-4425-9fe2-e21746f35514-kube-api-access-dxm9b\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb825xdk\" (UID: \"c56fa316-41a4-4425-9fe2-e21746f35514\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb825xdk" Nov 21 14:20:54 crc kubenswrapper[4897]: I1121 14:20:54.085008 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c56fa316-41a4-4425-9fe2-e21746f35514-bundle\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb825xdk\" (UID: \"c56fa316-41a4-4425-9fe2-e21746f35514\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb825xdk" Nov 21 14:20:54 crc kubenswrapper[4897]: I1121 14:20:54.105663 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fknv8t" Nov 21 14:20:54 crc kubenswrapper[4897]: I1121 14:20:54.186560 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c56fa316-41a4-4425-9fe2-e21746f35514-util\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb825xdk\" (UID: \"c56fa316-41a4-4425-9fe2-e21746f35514\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb825xdk" Nov 21 14:20:54 crc kubenswrapper[4897]: I1121 14:20:54.186850 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dxm9b\" (UniqueName: \"kubernetes.io/projected/c56fa316-41a4-4425-9fe2-e21746f35514-kube-api-access-dxm9b\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb825xdk\" (UID: \"c56fa316-41a4-4425-9fe2-e21746f35514\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb825xdk" Nov 21 14:20:54 crc kubenswrapper[4897]: I1121 14:20:54.186897 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c56fa316-41a4-4425-9fe2-e21746f35514-bundle\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb825xdk\" (UID: \"c56fa316-41a4-4425-9fe2-e21746f35514\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb825xdk" Nov 21 14:20:54 crc kubenswrapper[4897]: I1121 14:20:54.187768 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c56fa316-41a4-4425-9fe2-e21746f35514-bundle\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb825xdk\" (UID: \"c56fa316-41a4-4425-9fe2-e21746f35514\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb825xdk" Nov 21 14:20:54 crc kubenswrapper[4897]: I1121 14:20:54.188044 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c56fa316-41a4-4425-9fe2-e21746f35514-util\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb825xdk\" (UID: \"c56fa316-41a4-4425-9fe2-e21746f35514\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb825xdk" Nov 21 14:20:54 crc kubenswrapper[4897]: I1121 14:20:54.231565 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dxm9b\" (UniqueName: \"kubernetes.io/projected/c56fa316-41a4-4425-9fe2-e21746f35514-kube-api-access-dxm9b\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb825xdk\" (UID: \"c56fa316-41a4-4425-9fe2-e21746f35514\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb825xdk" Nov 21 14:20:54 crc kubenswrapper[4897]: I1121 14:20:54.293914 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb825xdk" Nov 21 14:20:54 crc kubenswrapper[4897]: I1121 14:20:54.518916 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb825xdk"] Nov 21 14:20:54 crc kubenswrapper[4897]: I1121 14:20:54.695586 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fknv8t"] Nov 21 14:20:54 crc kubenswrapper[4897]: W1121 14:20:54.701016 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1832f955_8108_4635_b24e_5988ca72ca44.slice/crio-cda606220039976c5bb35824dd64b88ce09ffba5d0fe3b5c54af9ae87ab87b8a WatchSource:0}: Error finding container cda606220039976c5bb35824dd64b88ce09ffba5d0fe3b5c54af9ae87ab87b8a: Status 404 returned error can't find the container with id cda606220039976c5bb35824dd64b88ce09ffba5d0fe3b5c54af9ae87ab87b8a Nov 21 14:20:54 crc kubenswrapper[4897]: I1121 14:20:54.758859 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fknv8t" event={"ID":"1832f955-8108-4635-b24e-5988ca72ca44","Type":"ContainerStarted","Data":"cda606220039976c5bb35824dd64b88ce09ffba5d0fe3b5c54af9ae87ab87b8a"} Nov 21 14:20:54 crc kubenswrapper[4897]: I1121 14:20:54.760211 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb825xdk" event={"ID":"c56fa316-41a4-4425-9fe2-e21746f35514","Type":"ContainerStarted","Data":"80ca6d1929b4808da2e8f44b1d665b3be662021375bb36ede0a606786108b2dc"} Nov 21 14:20:55 crc kubenswrapper[4897]: I1121 14:20:55.771328 4897 generic.go:334] "Generic (PLEG): container finished" podID="c56fa316-41a4-4425-9fe2-e21746f35514" containerID="8846d9feb62d2ef69a2ae49c6ce50a401b72c0a483875c4105bc5cb2e513d407" exitCode=0 Nov 21 14:20:55 crc kubenswrapper[4897]: I1121 14:20:55.771396 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb825xdk" event={"ID":"c56fa316-41a4-4425-9fe2-e21746f35514","Type":"ContainerDied","Data":"8846d9feb62d2ef69a2ae49c6ce50a401b72c0a483875c4105bc5cb2e513d407"} Nov 21 14:20:55 crc kubenswrapper[4897]: I1121 14:20:55.775802 4897 generic.go:334] "Generic (PLEG): container finished" podID="1832f955-8108-4635-b24e-5988ca72ca44" containerID="002f09e16a65a75decb763489d41efb18541818478da37dfced29dcf59dd6696" exitCode=0 Nov 21 14:20:55 crc kubenswrapper[4897]: I1121 14:20:55.775858 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fknv8t" event={"ID":"1832f955-8108-4635-b24e-5988ca72ca44","Type":"ContainerDied","Data":"002f09e16a65a75decb763489d41efb18541818478da37dfced29dcf59dd6696"} Nov 21 14:20:57 crc kubenswrapper[4897]: I1121 14:20:57.791878 4897 generic.go:334] "Generic (PLEG): container finished" podID="c56fa316-41a4-4425-9fe2-e21746f35514" containerID="c9be501e50341697d61f35f9da613bfdaff8500da61b6938afe9ab8aa63254e1" exitCode=0 Nov 21 14:20:57 crc kubenswrapper[4897]: I1121 14:20:57.791959 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb825xdk" event={"ID":"c56fa316-41a4-4425-9fe2-e21746f35514","Type":"ContainerDied","Data":"c9be501e50341697d61f35f9da613bfdaff8500da61b6938afe9ab8aa63254e1"} Nov 21 14:20:57 crc kubenswrapper[4897]: I1121 14:20:57.801863 4897 generic.go:334] "Generic (PLEG): container finished" podID="1832f955-8108-4635-b24e-5988ca72ca44" containerID="d66eae2ec09fe563363ee97c60332dd8664ddb2214c3709b53fa85a835963cf1" exitCode=0 Nov 21 14:20:57 crc kubenswrapper[4897]: I1121 14:20:57.801931 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fknv8t" event={"ID":"1832f955-8108-4635-b24e-5988ca72ca44","Type":"ContainerDied","Data":"d66eae2ec09fe563363ee97c60332dd8664ddb2214c3709b53fa85a835963cf1"} Nov 21 14:20:58 crc kubenswrapper[4897]: I1121 14:20:58.844025 4897 generic.go:334] "Generic (PLEG): container finished" podID="1832f955-8108-4635-b24e-5988ca72ca44" containerID="c5414990012c4c136130e9a3411ed9beafe3d1153c0c991dd9f4250706123ca2" exitCode=0 Nov 21 14:20:58 crc kubenswrapper[4897]: I1121 14:20:58.844244 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fknv8t" event={"ID":"1832f955-8108-4635-b24e-5988ca72ca44","Type":"ContainerDied","Data":"c5414990012c4c136130e9a3411ed9beafe3d1153c0c991dd9f4250706123ca2"} Nov 21 14:20:58 crc kubenswrapper[4897]: I1121 14:20:58.865709 4897 generic.go:334] "Generic (PLEG): container finished" podID="c56fa316-41a4-4425-9fe2-e21746f35514" containerID="7d431c4efb46343fb8db4f663969da67978c028fd40445420b24f44afc7806cd" exitCode=0 Nov 21 14:20:58 crc kubenswrapper[4897]: I1121 14:20:58.865770 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb825xdk" event={"ID":"c56fa316-41a4-4425-9fe2-e21746f35514","Type":"ContainerDied","Data":"7d431c4efb46343fb8db4f663969da67978c028fd40445420b24f44afc7806cd"} Nov 21 14:21:00 crc kubenswrapper[4897]: I1121 14:21:00.200444 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb825xdk" Nov 21 14:21:00 crc kubenswrapper[4897]: I1121 14:21:00.207638 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fknv8t" Nov 21 14:21:00 crc kubenswrapper[4897]: I1121 14:21:00.285716 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dxm9b\" (UniqueName: \"kubernetes.io/projected/c56fa316-41a4-4425-9fe2-e21746f35514-kube-api-access-dxm9b\") pod \"c56fa316-41a4-4425-9fe2-e21746f35514\" (UID: \"c56fa316-41a4-4425-9fe2-e21746f35514\") " Nov 21 14:21:00 crc kubenswrapper[4897]: I1121 14:21:00.286217 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c56fa316-41a4-4425-9fe2-e21746f35514-bundle\") pod \"c56fa316-41a4-4425-9fe2-e21746f35514\" (UID: \"c56fa316-41a4-4425-9fe2-e21746f35514\") " Nov 21 14:21:00 crc kubenswrapper[4897]: I1121 14:21:00.286280 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c56fa316-41a4-4425-9fe2-e21746f35514-util\") pod \"c56fa316-41a4-4425-9fe2-e21746f35514\" (UID: \"c56fa316-41a4-4425-9fe2-e21746f35514\") " Nov 21 14:21:00 crc kubenswrapper[4897]: I1121 14:21:00.287550 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c56fa316-41a4-4425-9fe2-e21746f35514-bundle" (OuterVolumeSpecName: "bundle") pod "c56fa316-41a4-4425-9fe2-e21746f35514" (UID: "c56fa316-41a4-4425-9fe2-e21746f35514"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:21:00 crc kubenswrapper[4897]: I1121 14:21:00.294644 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c56fa316-41a4-4425-9fe2-e21746f35514-kube-api-access-dxm9b" (OuterVolumeSpecName: "kube-api-access-dxm9b") pod "c56fa316-41a4-4425-9fe2-e21746f35514" (UID: "c56fa316-41a4-4425-9fe2-e21746f35514"). InnerVolumeSpecName "kube-api-access-dxm9b". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:21:00 crc kubenswrapper[4897]: I1121 14:21:00.318381 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c56fa316-41a4-4425-9fe2-e21746f35514-util" (OuterVolumeSpecName: "util") pod "c56fa316-41a4-4425-9fe2-e21746f35514" (UID: "c56fa316-41a4-4425-9fe2-e21746f35514"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:21:00 crc kubenswrapper[4897]: I1121 14:21:00.388063 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1832f955-8108-4635-b24e-5988ca72ca44-util\") pod \"1832f955-8108-4635-b24e-5988ca72ca44\" (UID: \"1832f955-8108-4635-b24e-5988ca72ca44\") " Nov 21 14:21:00 crc kubenswrapper[4897]: I1121 14:21:00.388204 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1832f955-8108-4635-b24e-5988ca72ca44-bundle\") pod \"1832f955-8108-4635-b24e-5988ca72ca44\" (UID: \"1832f955-8108-4635-b24e-5988ca72ca44\") " Nov 21 14:21:00 crc kubenswrapper[4897]: I1121 14:21:00.388224 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qfk9c\" (UniqueName: \"kubernetes.io/projected/1832f955-8108-4635-b24e-5988ca72ca44-kube-api-access-qfk9c\") pod \"1832f955-8108-4635-b24e-5988ca72ca44\" (UID: \"1832f955-8108-4635-b24e-5988ca72ca44\") " Nov 21 14:21:00 crc kubenswrapper[4897]: I1121 14:21:00.388459 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dxm9b\" (UniqueName: \"kubernetes.io/projected/c56fa316-41a4-4425-9fe2-e21746f35514-kube-api-access-dxm9b\") on node \"crc\" DevicePath \"\"" Nov 21 14:21:00 crc kubenswrapper[4897]: I1121 14:21:00.388472 4897 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c56fa316-41a4-4425-9fe2-e21746f35514-bundle\") on node \"crc\" DevicePath \"\"" Nov 21 14:21:00 crc kubenswrapper[4897]: I1121 14:21:00.388481 4897 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c56fa316-41a4-4425-9fe2-e21746f35514-util\") on node \"crc\" DevicePath \"\"" Nov 21 14:21:00 crc kubenswrapper[4897]: I1121 14:21:00.391095 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1832f955-8108-4635-b24e-5988ca72ca44-bundle" (OuterVolumeSpecName: "bundle") pod "1832f955-8108-4635-b24e-5988ca72ca44" (UID: "1832f955-8108-4635-b24e-5988ca72ca44"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:21:00 crc kubenswrapper[4897]: I1121 14:21:00.392561 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1832f955-8108-4635-b24e-5988ca72ca44-kube-api-access-qfk9c" (OuterVolumeSpecName: "kube-api-access-qfk9c") pod "1832f955-8108-4635-b24e-5988ca72ca44" (UID: "1832f955-8108-4635-b24e-5988ca72ca44"). InnerVolumeSpecName "kube-api-access-qfk9c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:21:00 crc kubenswrapper[4897]: I1121 14:21:00.490012 4897 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1832f955-8108-4635-b24e-5988ca72ca44-bundle\") on node \"crc\" DevicePath \"\"" Nov 21 14:21:00 crc kubenswrapper[4897]: I1121 14:21:00.490052 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qfk9c\" (UniqueName: \"kubernetes.io/projected/1832f955-8108-4635-b24e-5988ca72ca44-kube-api-access-qfk9c\") on node \"crc\" DevicePath \"\"" Nov 21 14:21:00 crc kubenswrapper[4897]: I1121 14:21:00.837186 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1832f955-8108-4635-b24e-5988ca72ca44-util" (OuterVolumeSpecName: "util") pod "1832f955-8108-4635-b24e-5988ca72ca44" (UID: "1832f955-8108-4635-b24e-5988ca72ca44"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:21:00 crc kubenswrapper[4897]: I1121 14:21:00.887926 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb825xdk" Nov 21 14:21:00 crc kubenswrapper[4897]: I1121 14:21:00.887922 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb825xdk" event={"ID":"c56fa316-41a4-4425-9fe2-e21746f35514","Type":"ContainerDied","Data":"80ca6d1929b4808da2e8f44b1d665b3be662021375bb36ede0a606786108b2dc"} Nov 21 14:21:00 crc kubenswrapper[4897]: I1121 14:21:00.888376 4897 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="80ca6d1929b4808da2e8f44b1d665b3be662021375bb36ede0a606786108b2dc" Nov 21 14:21:00 crc kubenswrapper[4897]: I1121 14:21:00.894836 4897 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1832f955-8108-4635-b24e-5988ca72ca44-util\") on node \"crc\" DevicePath \"\"" Nov 21 14:21:00 crc kubenswrapper[4897]: I1121 14:21:00.894839 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fknv8t" event={"ID":"1832f955-8108-4635-b24e-5988ca72ca44","Type":"ContainerDied","Data":"cda606220039976c5bb35824dd64b88ce09ffba5d0fe3b5c54af9ae87ab87b8a"} Nov 21 14:21:00 crc kubenswrapper[4897]: I1121 14:21:00.894910 4897 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cda606220039976c5bb35824dd64b88ce09ffba5d0fe3b5c54af9ae87ab87b8a" Nov 21 14:21:00 crc kubenswrapper[4897]: I1121 14:21:00.894938 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fknv8t" Nov 21 14:21:04 crc kubenswrapper[4897]: I1121 14:21:04.370956 4897 patch_prober.go:28] interesting pod/machine-config-daemon-krv5b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 21 14:21:04 crc kubenswrapper[4897]: I1121 14:21:04.371363 4897 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 21 14:21:10 crc kubenswrapper[4897]: I1121 14:21:10.938818 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators-redhat/loki-operator-controller-manager-757bf8868c-7gvjj"] Nov 21 14:21:10 crc kubenswrapper[4897]: E1121 14:21:10.939673 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1832f955-8108-4635-b24e-5988ca72ca44" containerName="pull" Nov 21 14:21:10 crc kubenswrapper[4897]: I1121 14:21:10.939695 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="1832f955-8108-4635-b24e-5988ca72ca44" containerName="pull" Nov 21 14:21:10 crc kubenswrapper[4897]: E1121 14:21:10.939716 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c56fa316-41a4-4425-9fe2-e21746f35514" containerName="pull" Nov 21 14:21:10 crc kubenswrapper[4897]: I1121 14:21:10.939727 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="c56fa316-41a4-4425-9fe2-e21746f35514" containerName="pull" Nov 21 14:21:10 crc kubenswrapper[4897]: E1121 14:21:10.939738 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c56fa316-41a4-4425-9fe2-e21746f35514" containerName="util" Nov 21 14:21:10 crc kubenswrapper[4897]: I1121 14:21:10.939750 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="c56fa316-41a4-4425-9fe2-e21746f35514" containerName="util" Nov 21 14:21:10 crc kubenswrapper[4897]: E1121 14:21:10.939763 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1832f955-8108-4635-b24e-5988ca72ca44" containerName="extract" Nov 21 14:21:10 crc kubenswrapper[4897]: I1121 14:21:10.939773 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="1832f955-8108-4635-b24e-5988ca72ca44" containerName="extract" Nov 21 14:21:10 crc kubenswrapper[4897]: E1121 14:21:10.939799 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1832f955-8108-4635-b24e-5988ca72ca44" containerName="util" Nov 21 14:21:10 crc kubenswrapper[4897]: I1121 14:21:10.939812 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="1832f955-8108-4635-b24e-5988ca72ca44" containerName="util" Nov 21 14:21:10 crc kubenswrapper[4897]: E1121 14:21:10.939832 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c56fa316-41a4-4425-9fe2-e21746f35514" containerName="extract" Nov 21 14:21:10 crc kubenswrapper[4897]: I1121 14:21:10.939843 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="c56fa316-41a4-4425-9fe2-e21746f35514" containerName="extract" Nov 21 14:21:10 crc kubenswrapper[4897]: I1121 14:21:10.940026 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="c56fa316-41a4-4425-9fe2-e21746f35514" containerName="extract" Nov 21 14:21:10 crc kubenswrapper[4897]: I1121 14:21:10.940040 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="1832f955-8108-4635-b24e-5988ca72ca44" containerName="extract" Nov 21 14:21:10 crc kubenswrapper[4897]: I1121 14:21:10.941208 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators-redhat/loki-operator-controller-manager-757bf8868c-7gvjj" Nov 21 14:21:10 crc kubenswrapper[4897]: I1121 14:21:10.946251 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators-redhat"/"openshift-service-ca.crt" Nov 21 14:21:10 crc kubenswrapper[4897]: I1121 14:21:10.946490 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators-redhat"/"loki-operator-manager-config" Nov 21 14:21:10 crc kubenswrapper[4897]: I1121 14:21:10.946751 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators-redhat"/"loki-operator-metrics" Nov 21 14:21:10 crc kubenswrapper[4897]: I1121 14:21:10.947129 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators-redhat"/"loki-operator-controller-manager-service-cert" Nov 21 14:21:10 crc kubenswrapper[4897]: I1121 14:21:10.947267 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators-redhat"/"loki-operator-controller-manager-dockercfg-57qpr" Nov 21 14:21:10 crc kubenswrapper[4897]: I1121 14:21:10.947381 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators-redhat"/"kube-root-ca.crt" Nov 21 14:21:10 crc kubenswrapper[4897]: I1121 14:21:10.970330 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators-redhat/loki-operator-controller-manager-757bf8868c-7gvjj"] Nov 21 14:21:11 crc kubenswrapper[4897]: I1121 14:21:11.040739 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/053c0db4-be6c-4590-bf73-ac68334f67df-webhook-cert\") pod \"loki-operator-controller-manager-757bf8868c-7gvjj\" (UID: \"053c0db4-be6c-4590-bf73-ac68334f67df\") " pod="openshift-operators-redhat/loki-operator-controller-manager-757bf8868c-7gvjj" Nov 21 14:21:11 crc kubenswrapper[4897]: I1121 14:21:11.040813 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"loki-operator-metrics-cert\" (UniqueName: \"kubernetes.io/secret/053c0db4-be6c-4590-bf73-ac68334f67df-loki-operator-metrics-cert\") pod \"loki-operator-controller-manager-757bf8868c-7gvjj\" (UID: \"053c0db4-be6c-4590-bf73-ac68334f67df\") " pod="openshift-operators-redhat/loki-operator-controller-manager-757bf8868c-7gvjj" Nov 21 14:21:11 crc kubenswrapper[4897]: I1121 14:21:11.040971 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manager-config\" (UniqueName: \"kubernetes.io/configmap/053c0db4-be6c-4590-bf73-ac68334f67df-manager-config\") pod \"loki-operator-controller-manager-757bf8868c-7gvjj\" (UID: \"053c0db4-be6c-4590-bf73-ac68334f67df\") " pod="openshift-operators-redhat/loki-operator-controller-manager-757bf8868c-7gvjj" Nov 21 14:21:11 crc kubenswrapper[4897]: I1121 14:21:11.041006 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5mmkh\" (UniqueName: \"kubernetes.io/projected/053c0db4-be6c-4590-bf73-ac68334f67df-kube-api-access-5mmkh\") pod \"loki-operator-controller-manager-757bf8868c-7gvjj\" (UID: \"053c0db4-be6c-4590-bf73-ac68334f67df\") " pod="openshift-operators-redhat/loki-operator-controller-manager-757bf8868c-7gvjj" Nov 21 14:21:11 crc kubenswrapper[4897]: I1121 14:21:11.041060 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/053c0db4-be6c-4590-bf73-ac68334f67df-apiservice-cert\") pod \"loki-operator-controller-manager-757bf8868c-7gvjj\" (UID: \"053c0db4-be6c-4590-bf73-ac68334f67df\") " pod="openshift-operators-redhat/loki-operator-controller-manager-757bf8868c-7gvjj" Nov 21 14:21:11 crc kubenswrapper[4897]: I1121 14:21:11.142531 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/053c0db4-be6c-4590-bf73-ac68334f67df-webhook-cert\") pod \"loki-operator-controller-manager-757bf8868c-7gvjj\" (UID: \"053c0db4-be6c-4590-bf73-ac68334f67df\") " pod="openshift-operators-redhat/loki-operator-controller-manager-757bf8868c-7gvjj" Nov 21 14:21:11 crc kubenswrapper[4897]: I1121 14:21:11.142919 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"loki-operator-metrics-cert\" (UniqueName: \"kubernetes.io/secret/053c0db4-be6c-4590-bf73-ac68334f67df-loki-operator-metrics-cert\") pod \"loki-operator-controller-manager-757bf8868c-7gvjj\" (UID: \"053c0db4-be6c-4590-bf73-ac68334f67df\") " pod="openshift-operators-redhat/loki-operator-controller-manager-757bf8868c-7gvjj" Nov 21 14:21:11 crc kubenswrapper[4897]: I1121 14:21:11.142969 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manager-config\" (UniqueName: \"kubernetes.io/configmap/053c0db4-be6c-4590-bf73-ac68334f67df-manager-config\") pod \"loki-operator-controller-manager-757bf8868c-7gvjj\" (UID: \"053c0db4-be6c-4590-bf73-ac68334f67df\") " pod="openshift-operators-redhat/loki-operator-controller-manager-757bf8868c-7gvjj" Nov 21 14:21:11 crc kubenswrapper[4897]: I1121 14:21:11.142985 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5mmkh\" (UniqueName: \"kubernetes.io/projected/053c0db4-be6c-4590-bf73-ac68334f67df-kube-api-access-5mmkh\") pod \"loki-operator-controller-manager-757bf8868c-7gvjj\" (UID: \"053c0db4-be6c-4590-bf73-ac68334f67df\") " pod="openshift-operators-redhat/loki-operator-controller-manager-757bf8868c-7gvjj" Nov 21 14:21:11 crc kubenswrapper[4897]: I1121 14:21:11.143012 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/053c0db4-be6c-4590-bf73-ac68334f67df-apiservice-cert\") pod \"loki-operator-controller-manager-757bf8868c-7gvjj\" (UID: \"053c0db4-be6c-4590-bf73-ac68334f67df\") " pod="openshift-operators-redhat/loki-operator-controller-manager-757bf8868c-7gvjj" Nov 21 14:21:11 crc kubenswrapper[4897]: I1121 14:21:11.144585 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manager-config\" (UniqueName: \"kubernetes.io/configmap/053c0db4-be6c-4590-bf73-ac68334f67df-manager-config\") pod \"loki-operator-controller-manager-757bf8868c-7gvjj\" (UID: \"053c0db4-be6c-4590-bf73-ac68334f67df\") " pod="openshift-operators-redhat/loki-operator-controller-manager-757bf8868c-7gvjj" Nov 21 14:21:11 crc kubenswrapper[4897]: I1121 14:21:11.162362 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"loki-operator-metrics-cert\" (UniqueName: \"kubernetes.io/secret/053c0db4-be6c-4590-bf73-ac68334f67df-loki-operator-metrics-cert\") pod \"loki-operator-controller-manager-757bf8868c-7gvjj\" (UID: \"053c0db4-be6c-4590-bf73-ac68334f67df\") " pod="openshift-operators-redhat/loki-operator-controller-manager-757bf8868c-7gvjj" Nov 21 14:21:11 crc kubenswrapper[4897]: I1121 14:21:11.164027 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/053c0db4-be6c-4590-bf73-ac68334f67df-apiservice-cert\") pod \"loki-operator-controller-manager-757bf8868c-7gvjj\" (UID: \"053c0db4-be6c-4590-bf73-ac68334f67df\") " pod="openshift-operators-redhat/loki-operator-controller-manager-757bf8868c-7gvjj" Nov 21 14:21:11 crc kubenswrapper[4897]: I1121 14:21:11.166065 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/053c0db4-be6c-4590-bf73-ac68334f67df-webhook-cert\") pod \"loki-operator-controller-manager-757bf8868c-7gvjj\" (UID: \"053c0db4-be6c-4590-bf73-ac68334f67df\") " pod="openshift-operators-redhat/loki-operator-controller-manager-757bf8868c-7gvjj" Nov 21 14:21:11 crc kubenswrapper[4897]: I1121 14:21:11.185859 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5mmkh\" (UniqueName: \"kubernetes.io/projected/053c0db4-be6c-4590-bf73-ac68334f67df-kube-api-access-5mmkh\") pod \"loki-operator-controller-manager-757bf8868c-7gvjj\" (UID: \"053c0db4-be6c-4590-bf73-ac68334f67df\") " pod="openshift-operators-redhat/loki-operator-controller-manager-757bf8868c-7gvjj" Nov 21 14:21:11 crc kubenswrapper[4897]: I1121 14:21:11.258155 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators-redhat/loki-operator-controller-manager-757bf8868c-7gvjj" Nov 21 14:21:11 crc kubenswrapper[4897]: I1121 14:21:11.720819 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators-redhat/loki-operator-controller-manager-757bf8868c-7gvjj"] Nov 21 14:21:11 crc kubenswrapper[4897]: W1121 14:21:11.727708 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod053c0db4_be6c_4590_bf73_ac68334f67df.slice/crio-0d06732007fde9db8f85faf0f6b7011c5e6dffcb508f469d156e1ec34d6ed26c WatchSource:0}: Error finding container 0d06732007fde9db8f85faf0f6b7011c5e6dffcb508f469d156e1ec34d6ed26c: Status 404 returned error can't find the container with id 0d06732007fde9db8f85faf0f6b7011c5e6dffcb508f469d156e1ec34d6ed26c Nov 21 14:21:11 crc kubenswrapper[4897]: I1121 14:21:11.967099 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators-redhat/loki-operator-controller-manager-757bf8868c-7gvjj" event={"ID":"053c0db4-be6c-4590-bf73-ac68334f67df","Type":"ContainerStarted","Data":"0d06732007fde9db8f85faf0f6b7011c5e6dffcb508f469d156e1ec34d6ed26c"} Nov 21 14:21:14 crc kubenswrapper[4897]: I1121 14:21:14.175680 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/cluster-logging-operator-ff9846bd-q9vjt"] Nov 21 14:21:14 crc kubenswrapper[4897]: I1121 14:21:14.176707 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/cluster-logging-operator-ff9846bd-q9vjt" Nov 21 14:21:14 crc kubenswrapper[4897]: I1121 14:21:14.178723 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"kube-root-ca.crt" Nov 21 14:21:14 crc kubenswrapper[4897]: I1121 14:21:14.180123 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"openshift-service-ca.crt" Nov 21 14:21:14 crc kubenswrapper[4897]: I1121 14:21:14.181262 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"cluster-logging-operator-dockercfg-bpx45" Nov 21 14:21:14 crc kubenswrapper[4897]: I1121 14:21:14.188076 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/cluster-logging-operator-ff9846bd-q9vjt"] Nov 21 14:21:14 crc kubenswrapper[4897]: I1121 14:21:14.293723 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xvh4c\" (UniqueName: \"kubernetes.io/projected/fcc70cd4-b166-456e-8a57-ea6515decc68-kube-api-access-xvh4c\") pod \"cluster-logging-operator-ff9846bd-q9vjt\" (UID: \"fcc70cd4-b166-456e-8a57-ea6515decc68\") " pod="openshift-logging/cluster-logging-operator-ff9846bd-q9vjt" Nov 21 14:21:14 crc kubenswrapper[4897]: I1121 14:21:14.395222 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xvh4c\" (UniqueName: \"kubernetes.io/projected/fcc70cd4-b166-456e-8a57-ea6515decc68-kube-api-access-xvh4c\") pod \"cluster-logging-operator-ff9846bd-q9vjt\" (UID: \"fcc70cd4-b166-456e-8a57-ea6515decc68\") " pod="openshift-logging/cluster-logging-operator-ff9846bd-q9vjt" Nov 21 14:21:14 crc kubenswrapper[4897]: I1121 14:21:14.417604 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xvh4c\" (UniqueName: \"kubernetes.io/projected/fcc70cd4-b166-456e-8a57-ea6515decc68-kube-api-access-xvh4c\") pod \"cluster-logging-operator-ff9846bd-q9vjt\" (UID: \"fcc70cd4-b166-456e-8a57-ea6515decc68\") " pod="openshift-logging/cluster-logging-operator-ff9846bd-q9vjt" Nov 21 14:21:14 crc kubenswrapper[4897]: I1121 14:21:14.498861 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/cluster-logging-operator-ff9846bd-q9vjt" Nov 21 14:21:16 crc kubenswrapper[4897]: I1121 14:21:16.797978 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/cluster-logging-operator-ff9846bd-q9vjt"] Nov 21 14:21:16 crc kubenswrapper[4897]: W1121 14:21:16.804694 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfcc70cd4_b166_456e_8a57_ea6515decc68.slice/crio-3064e51ad6e0a240882d3f63da8d08f40d7d4a33a41e384287f4f6e334e9d75f WatchSource:0}: Error finding container 3064e51ad6e0a240882d3f63da8d08f40d7d4a33a41e384287f4f6e334e9d75f: Status 404 returned error can't find the container with id 3064e51ad6e0a240882d3f63da8d08f40d7d4a33a41e384287f4f6e334e9d75f Nov 21 14:21:17 crc kubenswrapper[4897]: I1121 14:21:17.012474 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators-redhat/loki-operator-controller-manager-757bf8868c-7gvjj" event={"ID":"053c0db4-be6c-4590-bf73-ac68334f67df","Type":"ContainerStarted","Data":"977dd90366ba71bd412a99909cbbbc7b4698060de9556a91e89e1019f011c48c"} Nov 21 14:21:17 crc kubenswrapper[4897]: I1121 14:21:17.013696 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/cluster-logging-operator-ff9846bd-q9vjt" event={"ID":"fcc70cd4-b166-456e-8a57-ea6515decc68","Type":"ContainerStarted","Data":"3064e51ad6e0a240882d3f63da8d08f40d7d4a33a41e384287f4f6e334e9d75f"} Nov 21 14:21:17 crc kubenswrapper[4897]: I1121 14:21:17.268917 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-rnnlh"] Nov 21 14:21:17 crc kubenswrapper[4897]: I1121 14:21:17.269448 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-rnnlh" podUID="4bf1a18e-ec09-4f71-b6a1-645698067dc1" containerName="controller-manager" containerID="cri-o://fa8cb63bd527ad7c0cc4785027de1ca3217b1c4c8d640a8757cd888455fe1d99" gracePeriod=30 Nov 21 14:21:17 crc kubenswrapper[4897]: I1121 14:21:17.347535 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-lclf5"] Nov 21 14:21:17 crc kubenswrapper[4897]: I1121 14:21:17.347804 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lclf5" podUID="21295e15-1d49-4261-9970-fb61c1d4854c" containerName="route-controller-manager" containerID="cri-o://86eaed5522081a2fff93adb245c025c95e874ea66e172ecaed1b5770a821a7bc" gracePeriod=30 Nov 21 14:21:17 crc kubenswrapper[4897]: I1121 14:21:17.660905 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-rnnlh" Nov 21 14:21:17 crc kubenswrapper[4897]: I1121 14:21:17.724123 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lclf5" Nov 21 14:21:17 crc kubenswrapper[4897]: I1121 14:21:17.778219 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4bf1a18e-ec09-4f71-b6a1-645698067dc1-serving-cert\") pod \"4bf1a18e-ec09-4f71-b6a1-645698067dc1\" (UID: \"4bf1a18e-ec09-4f71-b6a1-645698067dc1\") " Nov 21 14:21:17 crc kubenswrapper[4897]: I1121 14:21:17.778388 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4bf1a18e-ec09-4f71-b6a1-645698067dc1-proxy-ca-bundles\") pod \"4bf1a18e-ec09-4f71-b6a1-645698067dc1\" (UID: \"4bf1a18e-ec09-4f71-b6a1-645698067dc1\") " Nov 21 14:21:17 crc kubenswrapper[4897]: I1121 14:21:17.778531 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4bf1a18e-ec09-4f71-b6a1-645698067dc1-client-ca\") pod \"4bf1a18e-ec09-4f71-b6a1-645698067dc1\" (UID: \"4bf1a18e-ec09-4f71-b6a1-645698067dc1\") " Nov 21 14:21:17 crc kubenswrapper[4897]: I1121 14:21:17.778595 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n9k72\" (UniqueName: \"kubernetes.io/projected/4bf1a18e-ec09-4f71-b6a1-645698067dc1-kube-api-access-n9k72\") pod \"4bf1a18e-ec09-4f71-b6a1-645698067dc1\" (UID: \"4bf1a18e-ec09-4f71-b6a1-645698067dc1\") " Nov 21 14:21:17 crc kubenswrapper[4897]: I1121 14:21:17.778680 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4bf1a18e-ec09-4f71-b6a1-645698067dc1-config\") pod \"4bf1a18e-ec09-4f71-b6a1-645698067dc1\" (UID: \"4bf1a18e-ec09-4f71-b6a1-645698067dc1\") " Nov 21 14:21:17 crc kubenswrapper[4897]: I1121 14:21:17.779207 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bf1a18e-ec09-4f71-b6a1-645698067dc1-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "4bf1a18e-ec09-4f71-b6a1-645698067dc1" (UID: "4bf1a18e-ec09-4f71-b6a1-645698067dc1"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:21:17 crc kubenswrapper[4897]: I1121 14:21:17.779714 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bf1a18e-ec09-4f71-b6a1-645698067dc1-client-ca" (OuterVolumeSpecName: "client-ca") pod "4bf1a18e-ec09-4f71-b6a1-645698067dc1" (UID: "4bf1a18e-ec09-4f71-b6a1-645698067dc1"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:21:17 crc kubenswrapper[4897]: I1121 14:21:17.779739 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bf1a18e-ec09-4f71-b6a1-645698067dc1-config" (OuterVolumeSpecName: "config") pod "4bf1a18e-ec09-4f71-b6a1-645698067dc1" (UID: "4bf1a18e-ec09-4f71-b6a1-645698067dc1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:21:17 crc kubenswrapper[4897]: I1121 14:21:17.785139 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bf1a18e-ec09-4f71-b6a1-645698067dc1-kube-api-access-n9k72" (OuterVolumeSpecName: "kube-api-access-n9k72") pod "4bf1a18e-ec09-4f71-b6a1-645698067dc1" (UID: "4bf1a18e-ec09-4f71-b6a1-645698067dc1"). InnerVolumeSpecName "kube-api-access-n9k72". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:21:17 crc kubenswrapper[4897]: I1121 14:21:17.785666 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4bf1a18e-ec09-4f71-b6a1-645698067dc1-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "4bf1a18e-ec09-4f71-b6a1-645698067dc1" (UID: "4bf1a18e-ec09-4f71-b6a1-645698067dc1"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:21:17 crc kubenswrapper[4897]: I1121 14:21:17.879842 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21295e15-1d49-4261-9970-fb61c1d4854c-config\") pod \"21295e15-1d49-4261-9970-fb61c1d4854c\" (UID: \"21295e15-1d49-4261-9970-fb61c1d4854c\") " Nov 21 14:21:17 crc kubenswrapper[4897]: I1121 14:21:17.879942 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/21295e15-1d49-4261-9970-fb61c1d4854c-serving-cert\") pod \"21295e15-1d49-4261-9970-fb61c1d4854c\" (UID: \"21295e15-1d49-4261-9970-fb61c1d4854c\") " Nov 21 14:21:17 crc kubenswrapper[4897]: I1121 14:21:17.880007 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/21295e15-1d49-4261-9970-fb61c1d4854c-client-ca\") pod \"21295e15-1d49-4261-9970-fb61c1d4854c\" (UID: \"21295e15-1d49-4261-9970-fb61c1d4854c\") " Nov 21 14:21:17 crc kubenswrapper[4897]: I1121 14:21:17.880030 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mtpkb\" (UniqueName: \"kubernetes.io/projected/21295e15-1d49-4261-9970-fb61c1d4854c-kube-api-access-mtpkb\") pod \"21295e15-1d49-4261-9970-fb61c1d4854c\" (UID: \"21295e15-1d49-4261-9970-fb61c1d4854c\") " Nov 21 14:21:17 crc kubenswrapper[4897]: I1121 14:21:17.880360 4897 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4bf1a18e-ec09-4f71-b6a1-645698067dc1-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 21 14:21:17 crc kubenswrapper[4897]: I1121 14:21:17.880387 4897 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4bf1a18e-ec09-4f71-b6a1-645698067dc1-client-ca\") on node \"crc\" DevicePath \"\"" Nov 21 14:21:17 crc kubenswrapper[4897]: I1121 14:21:17.880396 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n9k72\" (UniqueName: \"kubernetes.io/projected/4bf1a18e-ec09-4f71-b6a1-645698067dc1-kube-api-access-n9k72\") on node \"crc\" DevicePath \"\"" Nov 21 14:21:17 crc kubenswrapper[4897]: I1121 14:21:17.880405 4897 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4bf1a18e-ec09-4f71-b6a1-645698067dc1-config\") on node \"crc\" DevicePath \"\"" Nov 21 14:21:17 crc kubenswrapper[4897]: I1121 14:21:17.880414 4897 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4bf1a18e-ec09-4f71-b6a1-645698067dc1-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 21 14:21:17 crc kubenswrapper[4897]: I1121 14:21:17.881714 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/21295e15-1d49-4261-9970-fb61c1d4854c-config" (OuterVolumeSpecName: "config") pod "21295e15-1d49-4261-9970-fb61c1d4854c" (UID: "21295e15-1d49-4261-9970-fb61c1d4854c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:21:17 crc kubenswrapper[4897]: I1121 14:21:17.881734 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/21295e15-1d49-4261-9970-fb61c1d4854c-client-ca" (OuterVolumeSpecName: "client-ca") pod "21295e15-1d49-4261-9970-fb61c1d4854c" (UID: "21295e15-1d49-4261-9970-fb61c1d4854c"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:21:17 crc kubenswrapper[4897]: I1121 14:21:17.889064 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21295e15-1d49-4261-9970-fb61c1d4854c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "21295e15-1d49-4261-9970-fb61c1d4854c" (UID: "21295e15-1d49-4261-9970-fb61c1d4854c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:21:17 crc kubenswrapper[4897]: I1121 14:21:17.891852 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/21295e15-1d49-4261-9970-fb61c1d4854c-kube-api-access-mtpkb" (OuterVolumeSpecName: "kube-api-access-mtpkb") pod "21295e15-1d49-4261-9970-fb61c1d4854c" (UID: "21295e15-1d49-4261-9970-fb61c1d4854c"). InnerVolumeSpecName "kube-api-access-mtpkb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:21:17 crc kubenswrapper[4897]: I1121 14:21:17.981734 4897 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/21295e15-1d49-4261-9970-fb61c1d4854c-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 21 14:21:17 crc kubenswrapper[4897]: I1121 14:21:17.981767 4897 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/21295e15-1d49-4261-9970-fb61c1d4854c-client-ca\") on node \"crc\" DevicePath \"\"" Nov 21 14:21:17 crc kubenswrapper[4897]: I1121 14:21:17.981776 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mtpkb\" (UniqueName: \"kubernetes.io/projected/21295e15-1d49-4261-9970-fb61c1d4854c-kube-api-access-mtpkb\") on node \"crc\" DevicePath \"\"" Nov 21 14:21:17 crc kubenswrapper[4897]: I1121 14:21:17.981786 4897 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21295e15-1d49-4261-9970-fb61c1d4854c-config\") on node \"crc\" DevicePath \"\"" Nov 21 14:21:18 crc kubenswrapper[4897]: I1121 14:21:18.027447 4897 generic.go:334] "Generic (PLEG): container finished" podID="21295e15-1d49-4261-9970-fb61c1d4854c" containerID="86eaed5522081a2fff93adb245c025c95e874ea66e172ecaed1b5770a821a7bc" exitCode=0 Nov 21 14:21:18 crc kubenswrapper[4897]: I1121 14:21:18.027528 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lclf5" Nov 21 14:21:18 crc kubenswrapper[4897]: I1121 14:21:18.027556 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lclf5" event={"ID":"21295e15-1d49-4261-9970-fb61c1d4854c","Type":"ContainerDied","Data":"86eaed5522081a2fff93adb245c025c95e874ea66e172ecaed1b5770a821a7bc"} Nov 21 14:21:18 crc kubenswrapper[4897]: I1121 14:21:18.027593 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lclf5" event={"ID":"21295e15-1d49-4261-9970-fb61c1d4854c","Type":"ContainerDied","Data":"3b8df1d580432d827538c096ed3b5f0a33e8962c87d588205ff955abcc71e1a4"} Nov 21 14:21:18 crc kubenswrapper[4897]: I1121 14:21:18.027615 4897 scope.go:117] "RemoveContainer" containerID="86eaed5522081a2fff93adb245c025c95e874ea66e172ecaed1b5770a821a7bc" Nov 21 14:21:18 crc kubenswrapper[4897]: I1121 14:21:18.037817 4897 generic.go:334] "Generic (PLEG): container finished" podID="4bf1a18e-ec09-4f71-b6a1-645698067dc1" containerID="fa8cb63bd527ad7c0cc4785027de1ca3217b1c4c8d640a8757cd888455fe1d99" exitCode=0 Nov 21 14:21:18 crc kubenswrapper[4897]: I1121 14:21:18.037867 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-rnnlh" event={"ID":"4bf1a18e-ec09-4f71-b6a1-645698067dc1","Type":"ContainerDied","Data":"fa8cb63bd527ad7c0cc4785027de1ca3217b1c4c8d640a8757cd888455fe1d99"} Nov 21 14:21:18 crc kubenswrapper[4897]: I1121 14:21:18.037895 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-rnnlh" event={"ID":"4bf1a18e-ec09-4f71-b6a1-645698067dc1","Type":"ContainerDied","Data":"bd87699e03c4d7fba1fe4336158f39a61691180e715516516c484eb3d45e40fd"} Nov 21 14:21:18 crc kubenswrapper[4897]: I1121 14:21:18.037979 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-rnnlh" Nov 21 14:21:18 crc kubenswrapper[4897]: I1121 14:21:18.047171 4897 scope.go:117] "RemoveContainer" containerID="86eaed5522081a2fff93adb245c025c95e874ea66e172ecaed1b5770a821a7bc" Nov 21 14:21:18 crc kubenswrapper[4897]: E1121 14:21:18.053912 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"86eaed5522081a2fff93adb245c025c95e874ea66e172ecaed1b5770a821a7bc\": container with ID starting with 86eaed5522081a2fff93adb245c025c95e874ea66e172ecaed1b5770a821a7bc not found: ID does not exist" containerID="86eaed5522081a2fff93adb245c025c95e874ea66e172ecaed1b5770a821a7bc" Nov 21 14:21:18 crc kubenswrapper[4897]: I1121 14:21:18.054188 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"86eaed5522081a2fff93adb245c025c95e874ea66e172ecaed1b5770a821a7bc"} err="failed to get container status \"86eaed5522081a2fff93adb245c025c95e874ea66e172ecaed1b5770a821a7bc\": rpc error: code = NotFound desc = could not find container \"86eaed5522081a2fff93adb245c025c95e874ea66e172ecaed1b5770a821a7bc\": container with ID starting with 86eaed5522081a2fff93adb245c025c95e874ea66e172ecaed1b5770a821a7bc not found: ID does not exist" Nov 21 14:21:18 crc kubenswrapper[4897]: I1121 14:21:18.054278 4897 scope.go:117] "RemoveContainer" containerID="fa8cb63bd527ad7c0cc4785027de1ca3217b1c4c8d640a8757cd888455fe1d99" Nov 21 14:21:18 crc kubenswrapper[4897]: I1121 14:21:18.070093 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-lclf5"] Nov 21 14:21:18 crc kubenswrapper[4897]: I1121 14:21:18.074334 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-lclf5"] Nov 21 14:21:18 crc kubenswrapper[4897]: I1121 14:21:18.080888 4897 scope.go:117] "RemoveContainer" containerID="fa8cb63bd527ad7c0cc4785027de1ca3217b1c4c8d640a8757cd888455fe1d99" Nov 21 14:21:18 crc kubenswrapper[4897]: I1121 14:21:18.081091 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-rnnlh"] Nov 21 14:21:18 crc kubenswrapper[4897]: E1121 14:21:18.084280 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fa8cb63bd527ad7c0cc4785027de1ca3217b1c4c8d640a8757cd888455fe1d99\": container with ID starting with fa8cb63bd527ad7c0cc4785027de1ca3217b1c4c8d640a8757cd888455fe1d99 not found: ID does not exist" containerID="fa8cb63bd527ad7c0cc4785027de1ca3217b1c4c8d640a8757cd888455fe1d99" Nov 21 14:21:18 crc kubenswrapper[4897]: I1121 14:21:18.084380 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fa8cb63bd527ad7c0cc4785027de1ca3217b1c4c8d640a8757cd888455fe1d99"} err="failed to get container status \"fa8cb63bd527ad7c0cc4785027de1ca3217b1c4c8d640a8757cd888455fe1d99\": rpc error: code = NotFound desc = could not find container \"fa8cb63bd527ad7c0cc4785027de1ca3217b1c4c8d640a8757cd888455fe1d99\": container with ID starting with fa8cb63bd527ad7c0cc4785027de1ca3217b1c4c8d640a8757cd888455fe1d99 not found: ID does not exist" Nov 21 14:21:18 crc kubenswrapper[4897]: I1121 14:21:18.085087 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-rnnlh"] Nov 21 14:21:18 crc kubenswrapper[4897]: I1121 14:21:18.101147 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="21295e15-1d49-4261-9970-fb61c1d4854c" path="/var/lib/kubelet/pods/21295e15-1d49-4261-9970-fb61c1d4854c/volumes" Nov 21 14:21:18 crc kubenswrapper[4897]: I1121 14:21:18.101834 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bf1a18e-ec09-4f71-b6a1-645698067dc1" path="/var/lib/kubelet/pods/4bf1a18e-ec09-4f71-b6a1-645698067dc1/volumes" Nov 21 14:21:18 crc kubenswrapper[4897]: I1121 14:21:18.834652 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-7d97b76f9c-rfh2w"] Nov 21 14:21:18 crc kubenswrapper[4897]: E1121 14:21:18.834912 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21295e15-1d49-4261-9970-fb61c1d4854c" containerName="route-controller-manager" Nov 21 14:21:18 crc kubenswrapper[4897]: I1121 14:21:18.834931 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="21295e15-1d49-4261-9970-fb61c1d4854c" containerName="route-controller-manager" Nov 21 14:21:18 crc kubenswrapper[4897]: E1121 14:21:18.834945 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4bf1a18e-ec09-4f71-b6a1-645698067dc1" containerName="controller-manager" Nov 21 14:21:18 crc kubenswrapper[4897]: I1121 14:21:18.834951 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="4bf1a18e-ec09-4f71-b6a1-645698067dc1" containerName="controller-manager" Nov 21 14:21:18 crc kubenswrapper[4897]: I1121 14:21:18.835071 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="21295e15-1d49-4261-9970-fb61c1d4854c" containerName="route-controller-manager" Nov 21 14:21:18 crc kubenswrapper[4897]: I1121 14:21:18.835097 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="4bf1a18e-ec09-4f71-b6a1-645698067dc1" containerName="controller-manager" Nov 21 14:21:18 crc kubenswrapper[4897]: I1121 14:21:18.835483 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7d97b76f9c-rfh2w" Nov 21 14:21:18 crc kubenswrapper[4897]: I1121 14:21:18.838869 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-78bb484b55-w5p5w"] Nov 21 14:21:18 crc kubenswrapper[4897]: I1121 14:21:18.839739 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-78bb484b55-w5p5w" Nov 21 14:21:18 crc kubenswrapper[4897]: I1121 14:21:18.844118 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 21 14:21:18 crc kubenswrapper[4897]: I1121 14:21:18.846782 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 21 14:21:18 crc kubenswrapper[4897]: I1121 14:21:18.846984 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 21 14:21:18 crc kubenswrapper[4897]: I1121 14:21:18.847010 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 21 14:21:18 crc kubenswrapper[4897]: I1121 14:21:18.847192 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 21 14:21:18 crc kubenswrapper[4897]: I1121 14:21:18.847210 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 21 14:21:18 crc kubenswrapper[4897]: I1121 14:21:18.847250 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 21 14:21:18 crc kubenswrapper[4897]: I1121 14:21:18.847306 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 21 14:21:18 crc kubenswrapper[4897]: I1121 14:21:18.847347 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 21 14:21:18 crc kubenswrapper[4897]: I1121 14:21:18.847391 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 21 14:21:18 crc kubenswrapper[4897]: I1121 14:21:18.847420 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 21 14:21:18 crc kubenswrapper[4897]: I1121 14:21:18.847877 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 21 14:21:18 crc kubenswrapper[4897]: I1121 14:21:18.848121 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 21 14:21:18 crc kubenswrapper[4897]: I1121 14:21:18.860771 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-78bb484b55-w5p5w"] Nov 21 14:21:18 crc kubenswrapper[4897]: I1121 14:21:18.863493 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7d97b76f9c-rfh2w"] Nov 21 14:21:19 crc kubenswrapper[4897]: I1121 14:21:19.002369 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c9098754-c488-47f0-b965-33a24293d259-client-ca\") pod \"controller-manager-7d97b76f9c-rfh2w\" (UID: \"c9098754-c488-47f0-b965-33a24293d259\") " pod="openshift-controller-manager/controller-manager-7d97b76f9c-rfh2w" Nov 21 14:21:19 crc kubenswrapper[4897]: I1121 14:21:19.002448 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9098754-c488-47f0-b965-33a24293d259-config\") pod \"controller-manager-7d97b76f9c-rfh2w\" (UID: \"c9098754-c488-47f0-b965-33a24293d259\") " pod="openshift-controller-manager/controller-manager-7d97b76f9c-rfh2w" Nov 21 14:21:19 crc kubenswrapper[4897]: I1121 14:21:19.002490 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fbbc7ea9-6d1b-485f-a0eb-c59ab28c77a3-serving-cert\") pod \"route-controller-manager-78bb484b55-w5p5w\" (UID: \"fbbc7ea9-6d1b-485f-a0eb-c59ab28c77a3\") " pod="openshift-route-controller-manager/route-controller-manager-78bb484b55-w5p5w" Nov 21 14:21:19 crc kubenswrapper[4897]: I1121 14:21:19.002547 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c9098754-c488-47f0-b965-33a24293d259-proxy-ca-bundles\") pod \"controller-manager-7d97b76f9c-rfh2w\" (UID: \"c9098754-c488-47f0-b965-33a24293d259\") " pod="openshift-controller-manager/controller-manager-7d97b76f9c-rfh2w" Nov 21 14:21:19 crc kubenswrapper[4897]: I1121 14:21:19.002570 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-767vd\" (UniqueName: \"kubernetes.io/projected/c9098754-c488-47f0-b965-33a24293d259-kube-api-access-767vd\") pod \"controller-manager-7d97b76f9c-rfh2w\" (UID: \"c9098754-c488-47f0-b965-33a24293d259\") " pod="openshift-controller-manager/controller-manager-7d97b76f9c-rfh2w" Nov 21 14:21:19 crc kubenswrapper[4897]: I1121 14:21:19.002594 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fbbc7ea9-6d1b-485f-a0eb-c59ab28c77a3-config\") pod \"route-controller-manager-78bb484b55-w5p5w\" (UID: \"fbbc7ea9-6d1b-485f-a0eb-c59ab28c77a3\") " pod="openshift-route-controller-manager/route-controller-manager-78bb484b55-w5p5w" Nov 21 14:21:19 crc kubenswrapper[4897]: I1121 14:21:19.002617 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fbbc7ea9-6d1b-485f-a0eb-c59ab28c77a3-client-ca\") pod \"route-controller-manager-78bb484b55-w5p5w\" (UID: \"fbbc7ea9-6d1b-485f-a0eb-c59ab28c77a3\") " pod="openshift-route-controller-manager/route-controller-manager-78bb484b55-w5p5w" Nov 21 14:21:19 crc kubenswrapper[4897]: I1121 14:21:19.002713 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kzn66\" (UniqueName: \"kubernetes.io/projected/fbbc7ea9-6d1b-485f-a0eb-c59ab28c77a3-kube-api-access-kzn66\") pod \"route-controller-manager-78bb484b55-w5p5w\" (UID: \"fbbc7ea9-6d1b-485f-a0eb-c59ab28c77a3\") " pod="openshift-route-controller-manager/route-controller-manager-78bb484b55-w5p5w" Nov 21 14:21:19 crc kubenswrapper[4897]: I1121 14:21:19.002748 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c9098754-c488-47f0-b965-33a24293d259-serving-cert\") pod \"controller-manager-7d97b76f9c-rfh2w\" (UID: \"c9098754-c488-47f0-b965-33a24293d259\") " pod="openshift-controller-manager/controller-manager-7d97b76f9c-rfh2w" Nov 21 14:21:19 crc kubenswrapper[4897]: I1121 14:21:19.104204 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c9098754-c488-47f0-b965-33a24293d259-client-ca\") pod \"controller-manager-7d97b76f9c-rfh2w\" (UID: \"c9098754-c488-47f0-b965-33a24293d259\") " pod="openshift-controller-manager/controller-manager-7d97b76f9c-rfh2w" Nov 21 14:21:19 crc kubenswrapper[4897]: I1121 14:21:19.104275 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9098754-c488-47f0-b965-33a24293d259-config\") pod \"controller-manager-7d97b76f9c-rfh2w\" (UID: \"c9098754-c488-47f0-b965-33a24293d259\") " pod="openshift-controller-manager/controller-manager-7d97b76f9c-rfh2w" Nov 21 14:21:19 crc kubenswrapper[4897]: I1121 14:21:19.104311 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fbbc7ea9-6d1b-485f-a0eb-c59ab28c77a3-serving-cert\") pod \"route-controller-manager-78bb484b55-w5p5w\" (UID: \"fbbc7ea9-6d1b-485f-a0eb-c59ab28c77a3\") " pod="openshift-route-controller-manager/route-controller-manager-78bb484b55-w5p5w" Nov 21 14:21:19 crc kubenswrapper[4897]: I1121 14:21:19.104348 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c9098754-c488-47f0-b965-33a24293d259-proxy-ca-bundles\") pod \"controller-manager-7d97b76f9c-rfh2w\" (UID: \"c9098754-c488-47f0-b965-33a24293d259\") " pod="openshift-controller-manager/controller-manager-7d97b76f9c-rfh2w" Nov 21 14:21:19 crc kubenswrapper[4897]: I1121 14:21:19.104370 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-767vd\" (UniqueName: \"kubernetes.io/projected/c9098754-c488-47f0-b965-33a24293d259-kube-api-access-767vd\") pod \"controller-manager-7d97b76f9c-rfh2w\" (UID: \"c9098754-c488-47f0-b965-33a24293d259\") " pod="openshift-controller-manager/controller-manager-7d97b76f9c-rfh2w" Nov 21 14:21:19 crc kubenswrapper[4897]: I1121 14:21:19.104394 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fbbc7ea9-6d1b-485f-a0eb-c59ab28c77a3-config\") pod \"route-controller-manager-78bb484b55-w5p5w\" (UID: \"fbbc7ea9-6d1b-485f-a0eb-c59ab28c77a3\") " pod="openshift-route-controller-manager/route-controller-manager-78bb484b55-w5p5w" Nov 21 14:21:19 crc kubenswrapper[4897]: I1121 14:21:19.104416 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fbbc7ea9-6d1b-485f-a0eb-c59ab28c77a3-client-ca\") pod \"route-controller-manager-78bb484b55-w5p5w\" (UID: \"fbbc7ea9-6d1b-485f-a0eb-c59ab28c77a3\") " pod="openshift-route-controller-manager/route-controller-manager-78bb484b55-w5p5w" Nov 21 14:21:19 crc kubenswrapper[4897]: I1121 14:21:19.104517 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kzn66\" (UniqueName: \"kubernetes.io/projected/fbbc7ea9-6d1b-485f-a0eb-c59ab28c77a3-kube-api-access-kzn66\") pod \"route-controller-manager-78bb484b55-w5p5w\" (UID: \"fbbc7ea9-6d1b-485f-a0eb-c59ab28c77a3\") " pod="openshift-route-controller-manager/route-controller-manager-78bb484b55-w5p5w" Nov 21 14:21:19 crc kubenswrapper[4897]: I1121 14:21:19.104548 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c9098754-c488-47f0-b965-33a24293d259-serving-cert\") pod \"controller-manager-7d97b76f9c-rfh2w\" (UID: \"c9098754-c488-47f0-b965-33a24293d259\") " pod="openshift-controller-manager/controller-manager-7d97b76f9c-rfh2w" Nov 21 14:21:19 crc kubenswrapper[4897]: I1121 14:21:19.105864 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c9098754-c488-47f0-b965-33a24293d259-proxy-ca-bundles\") pod \"controller-manager-7d97b76f9c-rfh2w\" (UID: \"c9098754-c488-47f0-b965-33a24293d259\") " pod="openshift-controller-manager/controller-manager-7d97b76f9c-rfh2w" Nov 21 14:21:19 crc kubenswrapper[4897]: I1121 14:21:19.106541 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c9098754-c488-47f0-b965-33a24293d259-client-ca\") pod \"controller-manager-7d97b76f9c-rfh2w\" (UID: \"c9098754-c488-47f0-b965-33a24293d259\") " pod="openshift-controller-manager/controller-manager-7d97b76f9c-rfh2w" Nov 21 14:21:19 crc kubenswrapper[4897]: I1121 14:21:19.108295 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fbbc7ea9-6d1b-485f-a0eb-c59ab28c77a3-client-ca\") pod \"route-controller-manager-78bb484b55-w5p5w\" (UID: \"fbbc7ea9-6d1b-485f-a0eb-c59ab28c77a3\") " pod="openshift-route-controller-manager/route-controller-manager-78bb484b55-w5p5w" Nov 21 14:21:19 crc kubenswrapper[4897]: I1121 14:21:19.108298 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fbbc7ea9-6d1b-485f-a0eb-c59ab28c77a3-config\") pod \"route-controller-manager-78bb484b55-w5p5w\" (UID: \"fbbc7ea9-6d1b-485f-a0eb-c59ab28c77a3\") " pod="openshift-route-controller-manager/route-controller-manager-78bb484b55-w5p5w" Nov 21 14:21:19 crc kubenswrapper[4897]: I1121 14:21:19.108462 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9098754-c488-47f0-b965-33a24293d259-config\") pod \"controller-manager-7d97b76f9c-rfh2w\" (UID: \"c9098754-c488-47f0-b965-33a24293d259\") " pod="openshift-controller-manager/controller-manager-7d97b76f9c-rfh2w" Nov 21 14:21:19 crc kubenswrapper[4897]: I1121 14:21:19.112718 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fbbc7ea9-6d1b-485f-a0eb-c59ab28c77a3-serving-cert\") pod \"route-controller-manager-78bb484b55-w5p5w\" (UID: \"fbbc7ea9-6d1b-485f-a0eb-c59ab28c77a3\") " pod="openshift-route-controller-manager/route-controller-manager-78bb484b55-w5p5w" Nov 21 14:21:19 crc kubenswrapper[4897]: I1121 14:21:19.113235 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c9098754-c488-47f0-b965-33a24293d259-serving-cert\") pod \"controller-manager-7d97b76f9c-rfh2w\" (UID: \"c9098754-c488-47f0-b965-33a24293d259\") " pod="openshift-controller-manager/controller-manager-7d97b76f9c-rfh2w" Nov 21 14:21:19 crc kubenswrapper[4897]: I1121 14:21:19.125035 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-767vd\" (UniqueName: \"kubernetes.io/projected/c9098754-c488-47f0-b965-33a24293d259-kube-api-access-767vd\") pod \"controller-manager-7d97b76f9c-rfh2w\" (UID: \"c9098754-c488-47f0-b965-33a24293d259\") " pod="openshift-controller-manager/controller-manager-7d97b76f9c-rfh2w" Nov 21 14:21:19 crc kubenswrapper[4897]: I1121 14:21:19.128260 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kzn66\" (UniqueName: \"kubernetes.io/projected/fbbc7ea9-6d1b-485f-a0eb-c59ab28c77a3-kube-api-access-kzn66\") pod \"route-controller-manager-78bb484b55-w5p5w\" (UID: \"fbbc7ea9-6d1b-485f-a0eb-c59ab28c77a3\") " pod="openshift-route-controller-manager/route-controller-manager-78bb484b55-w5p5w" Nov 21 14:21:19 crc kubenswrapper[4897]: I1121 14:21:19.157774 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7d97b76f9c-rfh2w" Nov 21 14:21:19 crc kubenswrapper[4897]: I1121 14:21:19.167751 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-78bb484b55-w5p5w" Nov 21 14:21:19 crc kubenswrapper[4897]: I1121 14:21:19.649475 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-78bb484b55-w5p5w"] Nov 21 14:21:19 crc kubenswrapper[4897]: I1121 14:21:19.663338 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7d97b76f9c-rfh2w"] Nov 21 14:21:19 crc kubenswrapper[4897]: W1121 14:21:19.666924 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfbbc7ea9_6d1b_485f_a0eb_c59ab28c77a3.slice/crio-bd37cdbb807e3b816bd02e5be2728241f0d141f14036d8a329df4253f0a9e3ee WatchSource:0}: Error finding container bd37cdbb807e3b816bd02e5be2728241f0d141f14036d8a329df4253f0a9e3ee: Status 404 returned error can't find the container with id bd37cdbb807e3b816bd02e5be2728241f0d141f14036d8a329df4253f0a9e3ee Nov 21 14:21:19 crc kubenswrapper[4897]: W1121 14:21:19.669432 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc9098754_c488_47f0_b965_33a24293d259.slice/crio-44f2da6d4661e5514fc8c831451b16ebb11c14372ba3b3f22dfca3d0004ce214 WatchSource:0}: Error finding container 44f2da6d4661e5514fc8c831451b16ebb11c14372ba3b3f22dfca3d0004ce214: Status 404 returned error can't find the container with id 44f2da6d4661e5514fc8c831451b16ebb11c14372ba3b3f22dfca3d0004ce214 Nov 21 14:21:20 crc kubenswrapper[4897]: I1121 14:21:20.057544 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7d97b76f9c-rfh2w" event={"ID":"c9098754-c488-47f0-b965-33a24293d259","Type":"ContainerStarted","Data":"17443585c0878e046d7d426e679991cc52003eeaf0f6748ea627c3285bbbf576"} Nov 21 14:21:20 crc kubenswrapper[4897]: I1121 14:21:20.057615 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7d97b76f9c-rfh2w" event={"ID":"c9098754-c488-47f0-b965-33a24293d259","Type":"ContainerStarted","Data":"44f2da6d4661e5514fc8c831451b16ebb11c14372ba3b3f22dfca3d0004ce214"} Nov 21 14:21:20 crc kubenswrapper[4897]: I1121 14:21:20.057811 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-7d97b76f9c-rfh2w" Nov 21 14:21:20 crc kubenswrapper[4897]: I1121 14:21:20.061460 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-78bb484b55-w5p5w" event={"ID":"fbbc7ea9-6d1b-485f-a0eb-c59ab28c77a3","Type":"ContainerStarted","Data":"b2627108d66096117cc51076e3448e1405f586a63aadafcec1c9b9ef60f9b06d"} Nov 21 14:21:20 crc kubenswrapper[4897]: I1121 14:21:20.061515 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-78bb484b55-w5p5w" event={"ID":"fbbc7ea9-6d1b-485f-a0eb-c59ab28c77a3","Type":"ContainerStarted","Data":"bd37cdbb807e3b816bd02e5be2728241f0d141f14036d8a329df4253f0a9e3ee"} Nov 21 14:21:20 crc kubenswrapper[4897]: I1121 14:21:20.061650 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-78bb484b55-w5p5w" Nov 21 14:21:20 crc kubenswrapper[4897]: I1121 14:21:20.063949 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-7d97b76f9c-rfh2w" Nov 21 14:21:20 crc kubenswrapper[4897]: I1121 14:21:20.081604 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-7d97b76f9c-rfh2w" podStartSLOduration=2.081586037 podStartE2EDuration="2.081586037s" podCreationTimestamp="2025-11-21 14:21:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:21:20.078167875 +0000 UTC m=+757.362761350" watchObservedRunningTime="2025-11-21 14:21:20.081586037 +0000 UTC m=+757.366179512" Nov 21 14:21:20 crc kubenswrapper[4897]: I1121 14:21:20.128338 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-78bb484b55-w5p5w" podStartSLOduration=2.128319467 podStartE2EDuration="2.128319467s" podCreationTimestamp="2025-11-21 14:21:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:21:20.124661989 +0000 UTC m=+757.409255464" watchObservedRunningTime="2025-11-21 14:21:20.128319467 +0000 UTC m=+757.412912932" Nov 21 14:21:20 crc kubenswrapper[4897]: I1121 14:21:20.228110 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-78bb484b55-w5p5w" Nov 21 14:21:24 crc kubenswrapper[4897]: I1121 14:21:24.616035 4897 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 21 14:21:27 crc kubenswrapper[4897]: I1121 14:21:27.120486 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/cluster-logging-operator-ff9846bd-q9vjt" event={"ID":"fcc70cd4-b166-456e-8a57-ea6515decc68","Type":"ContainerStarted","Data":"1ee2ec501b6f68b45b97f2040eca278012b0cc8edeb065083cc741b982e49b98"} Nov 21 14:21:27 crc kubenswrapper[4897]: I1121 14:21:27.124620 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators-redhat/loki-operator-controller-manager-757bf8868c-7gvjj" event={"ID":"053c0db4-be6c-4590-bf73-ac68334f67df","Type":"ContainerStarted","Data":"61bc41231eae77d3330989395a7127a1b1ac3970402adb303d95b6fefd24d4d4"} Nov 21 14:21:27 crc kubenswrapper[4897]: I1121 14:21:27.125098 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators-redhat/loki-operator-controller-manager-757bf8868c-7gvjj" Nov 21 14:21:27 crc kubenswrapper[4897]: I1121 14:21:27.126667 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators-redhat/loki-operator-controller-manager-757bf8868c-7gvjj" Nov 21 14:21:27 crc kubenswrapper[4897]: I1121 14:21:27.138993 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/cluster-logging-operator-ff9846bd-q9vjt" podStartSLOduration=3.859475642 podStartE2EDuration="13.138966792s" podCreationTimestamp="2025-11-21 14:21:14 +0000 UTC" firstStartedPulling="2025-11-21 14:21:16.81193728 +0000 UTC m=+754.096530745" lastFinishedPulling="2025-11-21 14:21:26.09142841 +0000 UTC m=+763.376021895" observedRunningTime="2025-11-21 14:21:27.135070248 +0000 UTC m=+764.419663743" watchObservedRunningTime="2025-11-21 14:21:27.138966792 +0000 UTC m=+764.423560267" Nov 21 14:21:27 crc kubenswrapper[4897]: I1121 14:21:27.156543 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators-redhat/loki-operator-controller-manager-757bf8868c-7gvjj" podStartSLOduration=2.765714794 podStartE2EDuration="17.156520853s" podCreationTimestamp="2025-11-21 14:21:10 +0000 UTC" firstStartedPulling="2025-11-21 14:21:11.729995357 +0000 UTC m=+749.014588842" lastFinishedPulling="2025-11-21 14:21:26.120801436 +0000 UTC m=+763.405394901" observedRunningTime="2025-11-21 14:21:27.152677379 +0000 UTC m=+764.437270854" watchObservedRunningTime="2025-11-21 14:21:27.156520853 +0000 UTC m=+764.441114328" Nov 21 14:21:31 crc kubenswrapper[4897]: I1121 14:21:31.000233 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["minio-dev/minio"] Nov 21 14:21:31 crc kubenswrapper[4897]: I1121 14:21:31.002383 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="minio-dev/minio" Nov 21 14:21:31 crc kubenswrapper[4897]: I1121 14:21:31.004586 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"minio-dev"/"kube-root-ca.crt" Nov 21 14:21:31 crc kubenswrapper[4897]: I1121 14:21:31.005643 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"minio-dev"/"openshift-service-ca.crt" Nov 21 14:21:31 crc kubenswrapper[4897]: I1121 14:21:31.005878 4897 reflector.go:368] Caches populated for *v1.Secret from object-"minio-dev"/"default-dockercfg-44srh" Nov 21 14:21:31 crc kubenswrapper[4897]: I1121 14:21:31.012624 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["minio-dev/minio"] Nov 21 14:21:31 crc kubenswrapper[4897]: I1121 14:21:31.082351 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-cb107f22-d199-40a9-a012-5cdcd3ba5cf6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-cb107f22-d199-40a9-a012-5cdcd3ba5cf6\") pod \"minio\" (UID: \"849188c8-a3da-4aff-a89d-90db776a3309\") " pod="minio-dev/minio" Nov 21 14:21:31 crc kubenswrapper[4897]: I1121 14:21:31.082478 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rwlrq\" (UniqueName: \"kubernetes.io/projected/849188c8-a3da-4aff-a89d-90db776a3309-kube-api-access-rwlrq\") pod \"minio\" (UID: \"849188c8-a3da-4aff-a89d-90db776a3309\") " pod="minio-dev/minio" Nov 21 14:21:31 crc kubenswrapper[4897]: I1121 14:21:31.183312 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rwlrq\" (UniqueName: \"kubernetes.io/projected/849188c8-a3da-4aff-a89d-90db776a3309-kube-api-access-rwlrq\") pod \"minio\" (UID: \"849188c8-a3da-4aff-a89d-90db776a3309\") " pod="minio-dev/minio" Nov 21 14:21:31 crc kubenswrapper[4897]: I1121 14:21:31.183488 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-cb107f22-d199-40a9-a012-5cdcd3ba5cf6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-cb107f22-d199-40a9-a012-5cdcd3ba5cf6\") pod \"minio\" (UID: \"849188c8-a3da-4aff-a89d-90db776a3309\") " pod="minio-dev/minio" Nov 21 14:21:31 crc kubenswrapper[4897]: I1121 14:21:31.186452 4897 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 21 14:21:31 crc kubenswrapper[4897]: I1121 14:21:31.186488 4897 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-cb107f22-d199-40a9-a012-5cdcd3ba5cf6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-cb107f22-d199-40a9-a012-5cdcd3ba5cf6\") pod \"minio\" (UID: \"849188c8-a3da-4aff-a89d-90db776a3309\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/cab0c9337509e6de44cd91a390e180cf8b5bf640988723b0788e4e8408d60c6c/globalmount\"" pod="minio-dev/minio" Nov 21 14:21:31 crc kubenswrapper[4897]: I1121 14:21:31.208132 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rwlrq\" (UniqueName: \"kubernetes.io/projected/849188c8-a3da-4aff-a89d-90db776a3309-kube-api-access-rwlrq\") pod \"minio\" (UID: \"849188c8-a3da-4aff-a89d-90db776a3309\") " pod="minio-dev/minio" Nov 21 14:21:31 crc kubenswrapper[4897]: I1121 14:21:31.227535 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-cb107f22-d199-40a9-a012-5cdcd3ba5cf6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-cb107f22-d199-40a9-a012-5cdcd3ba5cf6\") pod \"minio\" (UID: \"849188c8-a3da-4aff-a89d-90db776a3309\") " pod="minio-dev/minio" Nov 21 14:21:31 crc kubenswrapper[4897]: I1121 14:21:31.320539 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="minio-dev/minio" Nov 21 14:21:31 crc kubenswrapper[4897]: I1121 14:21:31.777486 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["minio-dev/minio"] Nov 21 14:21:31 crc kubenswrapper[4897]: W1121 14:21:31.791722 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod849188c8_a3da_4aff_a89d_90db776a3309.slice/crio-75fa9f708281f350ebe0ffb52e2ac567e997c11e83ee7de543a19a61f9e7a912 WatchSource:0}: Error finding container 75fa9f708281f350ebe0ffb52e2ac567e997c11e83ee7de543a19a61f9e7a912: Status 404 returned error can't find the container with id 75fa9f708281f350ebe0ffb52e2ac567e997c11e83ee7de543a19a61f9e7a912 Nov 21 14:21:32 crc kubenswrapper[4897]: I1121 14:21:32.158288 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="minio-dev/minio" event={"ID":"849188c8-a3da-4aff-a89d-90db776a3309","Type":"ContainerStarted","Data":"75fa9f708281f350ebe0ffb52e2ac567e997c11e83ee7de543a19a61f9e7a912"} Nov 21 14:21:34 crc kubenswrapper[4897]: I1121 14:21:34.372014 4897 patch_prober.go:28] interesting pod/machine-config-daemon-krv5b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 21 14:21:34 crc kubenswrapper[4897]: I1121 14:21:34.372497 4897 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 21 14:21:35 crc kubenswrapper[4897]: I1121 14:21:35.180155 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="minio-dev/minio" event={"ID":"849188c8-a3da-4aff-a89d-90db776a3309","Type":"ContainerStarted","Data":"c08a8b1c4e835bba9e75acaced5b99ddbe3c259bec38f64c3d6104d5f00824dd"} Nov 21 14:21:35 crc kubenswrapper[4897]: I1121 14:21:35.199741 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="minio-dev/minio" podStartSLOduration=4.275943148 podStartE2EDuration="7.199720859s" podCreationTimestamp="2025-11-21 14:21:28 +0000 UTC" firstStartedPulling="2025-11-21 14:21:31.793279502 +0000 UTC m=+769.077872977" lastFinishedPulling="2025-11-21 14:21:34.717057203 +0000 UTC m=+772.001650688" observedRunningTime="2025-11-21 14:21:35.193972805 +0000 UTC m=+772.478566310" watchObservedRunningTime="2025-11-21 14:21:35.199720859 +0000 UTC m=+772.484314344" Nov 21 14:21:38 crc kubenswrapper[4897]: I1121 14:21:38.520311 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-distributor-76cc67bf56-m64dx"] Nov 21 14:21:38 crc kubenswrapper[4897]: I1121 14:21:38.521497 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-distributor-76cc67bf56-m64dx" Nov 21 14:21:38 crc kubenswrapper[4897]: I1121 14:21:38.524169 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-distributor-grpc" Nov 21 14:21:38 crc kubenswrapper[4897]: I1121 14:21:38.524171 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"logging-loki-ca-bundle" Nov 21 14:21:38 crc kubenswrapper[4897]: I1121 14:21:38.524199 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-dockercfg-mlp22" Nov 21 14:21:38 crc kubenswrapper[4897]: I1121 14:21:38.524208 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"logging-loki-config" Nov 21 14:21:38 crc kubenswrapper[4897]: I1121 14:21:38.525426 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-distributor-http" Nov 21 14:21:38 crc kubenswrapper[4897]: I1121 14:21:38.542313 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-distributor-76cc67bf56-m64dx"] Nov 21 14:21:38 crc kubenswrapper[4897]: I1121 14:21:38.614875 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d3f767a5-605a-4b21-a15a-68cde314d5e6-logging-loki-ca-bundle\") pod \"logging-loki-distributor-76cc67bf56-m64dx\" (UID: \"d3f767a5-605a-4b21-a15a-68cde314d5e6\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-m64dx" Nov 21 14:21:38 crc kubenswrapper[4897]: I1121 14:21:38.614944 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3f767a5-605a-4b21-a15a-68cde314d5e6-config\") pod \"logging-loki-distributor-76cc67bf56-m64dx\" (UID: \"d3f767a5-605a-4b21-a15a-68cde314d5e6\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-m64dx" Nov 21 14:21:38 crc kubenswrapper[4897]: I1121 14:21:38.614982 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dlbd6\" (UniqueName: \"kubernetes.io/projected/d3f767a5-605a-4b21-a15a-68cde314d5e6-kube-api-access-dlbd6\") pod \"logging-loki-distributor-76cc67bf56-m64dx\" (UID: \"d3f767a5-605a-4b21-a15a-68cde314d5e6\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-m64dx" Nov 21 14:21:38 crc kubenswrapper[4897]: I1121 14:21:38.615039 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-distributor-http\" (UniqueName: \"kubernetes.io/secret/d3f767a5-605a-4b21-a15a-68cde314d5e6-logging-loki-distributor-http\") pod \"logging-loki-distributor-76cc67bf56-m64dx\" (UID: \"d3f767a5-605a-4b21-a15a-68cde314d5e6\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-m64dx" Nov 21 14:21:38 crc kubenswrapper[4897]: I1121 14:21:38.615073 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-distributor-grpc\" (UniqueName: \"kubernetes.io/secret/d3f767a5-605a-4b21-a15a-68cde314d5e6-logging-loki-distributor-grpc\") pod \"logging-loki-distributor-76cc67bf56-m64dx\" (UID: \"d3f767a5-605a-4b21-a15a-68cde314d5e6\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-m64dx" Nov 21 14:21:38 crc kubenswrapper[4897]: I1121 14:21:38.648029 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-querier-5895d59bb8-8lqvs"] Nov 21 14:21:38 crc kubenswrapper[4897]: I1121 14:21:38.653639 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-querier-5895d59bb8-8lqvs" Nov 21 14:21:38 crc kubenswrapper[4897]: I1121 14:21:38.661087 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-querier-grpc" Nov 21 14:21:38 crc kubenswrapper[4897]: I1121 14:21:38.661346 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-s3" Nov 21 14:21:38 crc kubenswrapper[4897]: I1121 14:21:38.661499 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-querier-http" Nov 21 14:21:38 crc kubenswrapper[4897]: I1121 14:21:38.681805 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-querier-5895d59bb8-8lqvs"] Nov 21 14:21:38 crc kubenswrapper[4897]: I1121 14:21:38.719334 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-distributor-http\" (UniqueName: \"kubernetes.io/secret/d3f767a5-605a-4b21-a15a-68cde314d5e6-logging-loki-distributor-http\") pod \"logging-loki-distributor-76cc67bf56-m64dx\" (UID: \"d3f767a5-605a-4b21-a15a-68cde314d5e6\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-m64dx" Nov 21 14:21:38 crc kubenswrapper[4897]: I1121 14:21:38.719416 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-distributor-grpc\" (UniqueName: \"kubernetes.io/secret/d3f767a5-605a-4b21-a15a-68cde314d5e6-logging-loki-distributor-grpc\") pod \"logging-loki-distributor-76cc67bf56-m64dx\" (UID: \"d3f767a5-605a-4b21-a15a-68cde314d5e6\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-m64dx" Nov 21 14:21:38 crc kubenswrapper[4897]: I1121 14:21:38.719470 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bqlw4\" (UniqueName: \"kubernetes.io/projected/53c2fb7b-24cd-4da0-9016-e35e3223ab06-kube-api-access-bqlw4\") pod \"logging-loki-querier-5895d59bb8-8lqvs\" (UID: \"53c2fb7b-24cd-4da0-9016-e35e3223ab06\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-8lqvs" Nov 21 14:21:38 crc kubenswrapper[4897]: I1121 14:21:38.719533 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-querier-http\" (UniqueName: \"kubernetes.io/secret/53c2fb7b-24cd-4da0-9016-e35e3223ab06-logging-loki-querier-http\") pod \"logging-loki-querier-5895d59bb8-8lqvs\" (UID: \"53c2fb7b-24cd-4da0-9016-e35e3223ab06\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-8lqvs" Nov 21 14:21:38 crc kubenswrapper[4897]: I1121 14:21:38.719572 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/53c2fb7b-24cd-4da0-9016-e35e3223ab06-config\") pod \"logging-loki-querier-5895d59bb8-8lqvs\" (UID: \"53c2fb7b-24cd-4da0-9016-e35e3223ab06\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-8lqvs" Nov 21 14:21:38 crc kubenswrapper[4897]: I1121 14:21:38.719637 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/53c2fb7b-24cd-4da0-9016-e35e3223ab06-logging-loki-s3\") pod \"logging-loki-querier-5895d59bb8-8lqvs\" (UID: \"53c2fb7b-24cd-4da0-9016-e35e3223ab06\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-8lqvs" Nov 21 14:21:38 crc kubenswrapper[4897]: I1121 14:21:38.719709 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-querier-grpc\" (UniqueName: \"kubernetes.io/secret/53c2fb7b-24cd-4da0-9016-e35e3223ab06-logging-loki-querier-grpc\") pod \"logging-loki-querier-5895d59bb8-8lqvs\" (UID: \"53c2fb7b-24cd-4da0-9016-e35e3223ab06\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-8lqvs" Nov 21 14:21:38 crc kubenswrapper[4897]: I1121 14:21:38.719738 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d3f767a5-605a-4b21-a15a-68cde314d5e6-logging-loki-ca-bundle\") pod \"logging-loki-distributor-76cc67bf56-m64dx\" (UID: \"d3f767a5-605a-4b21-a15a-68cde314d5e6\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-m64dx" Nov 21 14:21:38 crc kubenswrapper[4897]: I1121 14:21:38.719804 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3f767a5-605a-4b21-a15a-68cde314d5e6-config\") pod \"logging-loki-distributor-76cc67bf56-m64dx\" (UID: \"d3f767a5-605a-4b21-a15a-68cde314d5e6\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-m64dx" Nov 21 14:21:38 crc kubenswrapper[4897]: I1121 14:21:38.719834 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/53c2fb7b-24cd-4da0-9016-e35e3223ab06-logging-loki-ca-bundle\") pod \"logging-loki-querier-5895d59bb8-8lqvs\" (UID: \"53c2fb7b-24cd-4da0-9016-e35e3223ab06\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-8lqvs" Nov 21 14:21:38 crc kubenswrapper[4897]: I1121 14:21:38.719866 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dlbd6\" (UniqueName: \"kubernetes.io/projected/d3f767a5-605a-4b21-a15a-68cde314d5e6-kube-api-access-dlbd6\") pod \"logging-loki-distributor-76cc67bf56-m64dx\" (UID: \"d3f767a5-605a-4b21-a15a-68cde314d5e6\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-m64dx" Nov 21 14:21:38 crc kubenswrapper[4897]: I1121 14:21:38.720227 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-query-frontend-84558f7c9f-vltt5"] Nov 21 14:21:38 crc kubenswrapper[4897]: I1121 14:21:38.721225 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-vltt5" Nov 21 14:21:38 crc kubenswrapper[4897]: I1121 14:21:38.724466 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d3f767a5-605a-4b21-a15a-68cde314d5e6-logging-loki-ca-bundle\") pod \"logging-loki-distributor-76cc67bf56-m64dx\" (UID: \"d3f767a5-605a-4b21-a15a-68cde314d5e6\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-m64dx" Nov 21 14:21:38 crc kubenswrapper[4897]: I1121 14:21:38.724842 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-query-frontend-grpc" Nov 21 14:21:38 crc kubenswrapper[4897]: I1121 14:21:38.725089 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-query-frontend-http" Nov 21 14:21:38 crc kubenswrapper[4897]: I1121 14:21:38.725378 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3f767a5-605a-4b21-a15a-68cde314d5e6-config\") pod \"logging-loki-distributor-76cc67bf56-m64dx\" (UID: \"d3f767a5-605a-4b21-a15a-68cde314d5e6\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-m64dx" Nov 21 14:21:38 crc kubenswrapper[4897]: I1121 14:21:38.736651 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-distributor-grpc\" (UniqueName: \"kubernetes.io/secret/d3f767a5-605a-4b21-a15a-68cde314d5e6-logging-loki-distributor-grpc\") pod \"logging-loki-distributor-76cc67bf56-m64dx\" (UID: \"d3f767a5-605a-4b21-a15a-68cde314d5e6\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-m64dx" Nov 21 14:21:38 crc kubenswrapper[4897]: I1121 14:21:38.738391 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-query-frontend-84558f7c9f-vltt5"] Nov 21 14:21:38 crc kubenswrapper[4897]: I1121 14:21:38.748394 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dlbd6\" (UniqueName: \"kubernetes.io/projected/d3f767a5-605a-4b21-a15a-68cde314d5e6-kube-api-access-dlbd6\") pod \"logging-loki-distributor-76cc67bf56-m64dx\" (UID: \"d3f767a5-605a-4b21-a15a-68cde314d5e6\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-m64dx" Nov 21 14:21:38 crc kubenswrapper[4897]: I1121 14:21:38.753572 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-distributor-http\" (UniqueName: \"kubernetes.io/secret/d3f767a5-605a-4b21-a15a-68cde314d5e6-logging-loki-distributor-http\") pod \"logging-loki-distributor-76cc67bf56-m64dx\" (UID: \"d3f767a5-605a-4b21-a15a-68cde314d5e6\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-m64dx" Nov 21 14:21:38 crc kubenswrapper[4897]: I1121 14:21:38.821454 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bqlw4\" (UniqueName: \"kubernetes.io/projected/53c2fb7b-24cd-4da0-9016-e35e3223ab06-kube-api-access-bqlw4\") pod \"logging-loki-querier-5895d59bb8-8lqvs\" (UID: \"53c2fb7b-24cd-4da0-9016-e35e3223ab06\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-8lqvs" Nov 21 14:21:38 crc kubenswrapper[4897]: I1121 14:21:38.821566 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-querier-http\" (UniqueName: \"kubernetes.io/secret/53c2fb7b-24cd-4da0-9016-e35e3223ab06-logging-loki-querier-http\") pod \"logging-loki-querier-5895d59bb8-8lqvs\" (UID: \"53c2fb7b-24cd-4da0-9016-e35e3223ab06\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-8lqvs" Nov 21 14:21:38 crc kubenswrapper[4897]: I1121 14:21:38.821605 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/53c2fb7b-24cd-4da0-9016-e35e3223ab06-config\") pod \"logging-loki-querier-5895d59bb8-8lqvs\" (UID: \"53c2fb7b-24cd-4da0-9016-e35e3223ab06\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-8lqvs" Nov 21 14:21:38 crc kubenswrapper[4897]: I1121 14:21:38.821655 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/53c2fb7b-24cd-4da0-9016-e35e3223ab06-logging-loki-s3\") pod \"logging-loki-querier-5895d59bb8-8lqvs\" (UID: \"53c2fb7b-24cd-4da0-9016-e35e3223ab06\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-8lqvs" Nov 21 14:21:38 crc kubenswrapper[4897]: I1121 14:21:38.821682 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-querier-grpc\" (UniqueName: \"kubernetes.io/secret/53c2fb7b-24cd-4da0-9016-e35e3223ab06-logging-loki-querier-grpc\") pod \"logging-loki-querier-5895d59bb8-8lqvs\" (UID: \"53c2fb7b-24cd-4da0-9016-e35e3223ab06\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-8lqvs" Nov 21 14:21:38 crc kubenswrapper[4897]: I1121 14:21:38.821720 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/53c2fb7b-24cd-4da0-9016-e35e3223ab06-logging-loki-ca-bundle\") pod \"logging-loki-querier-5895d59bb8-8lqvs\" (UID: \"53c2fb7b-24cd-4da0-9016-e35e3223ab06\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-8lqvs" Nov 21 14:21:38 crc kubenswrapper[4897]: I1121 14:21:38.822682 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/53c2fb7b-24cd-4da0-9016-e35e3223ab06-logging-loki-ca-bundle\") pod \"logging-loki-querier-5895d59bb8-8lqvs\" (UID: \"53c2fb7b-24cd-4da0-9016-e35e3223ab06\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-8lqvs" Nov 21 14:21:38 crc kubenswrapper[4897]: I1121 14:21:38.825616 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/53c2fb7b-24cd-4da0-9016-e35e3223ab06-config\") pod \"logging-loki-querier-5895d59bb8-8lqvs\" (UID: \"53c2fb7b-24cd-4da0-9016-e35e3223ab06\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-8lqvs" Nov 21 14:21:38 crc kubenswrapper[4897]: I1121 14:21:38.826742 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-querier-grpc\" (UniqueName: \"kubernetes.io/secret/53c2fb7b-24cd-4da0-9016-e35e3223ab06-logging-loki-querier-grpc\") pod \"logging-loki-querier-5895d59bb8-8lqvs\" (UID: \"53c2fb7b-24cd-4da0-9016-e35e3223ab06\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-8lqvs" Nov 21 14:21:38 crc kubenswrapper[4897]: I1121 14:21:38.836945 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-gateway-78b4f7fc55-dbkmd"] Nov 21 14:21:38 crc kubenswrapper[4897]: I1121 14:21:38.837962 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-gateway-78b4f7fc55-dbkmd" Nov 21 14:21:38 crc kubenswrapper[4897]: I1121 14:21:38.844041 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-distributor-76cc67bf56-m64dx" Nov 21 14:21:38 crc kubenswrapper[4897]: I1121 14:21:38.844047 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-gateway-client-http" Nov 21 14:21:38 crc kubenswrapper[4897]: I1121 14:21:38.844111 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"logging-loki-gateway" Nov 21 14:21:38 crc kubenswrapper[4897]: I1121 14:21:38.844184 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"logging-loki-gateway-ca-bundle" Nov 21 14:21:38 crc kubenswrapper[4897]: I1121 14:21:38.844209 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-gateway" Nov 21 14:21:38 crc kubenswrapper[4897]: I1121 14:21:38.844968 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-gateway-78b4f7fc55-txlqp"] Nov 21 14:21:38 crc kubenswrapper[4897]: I1121 14:21:38.844254 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-gateway-dockercfg-sz4hq" Nov 21 14:21:38 crc kubenswrapper[4897]: I1121 14:21:38.844283 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-gateway-http" Nov 21 14:21:38 crc kubenswrapper[4897]: I1121 14:21:38.845952 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-gateway-78b4f7fc55-txlqp" Nov 21 14:21:38 crc kubenswrapper[4897]: I1121 14:21:38.852423 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-querier-http\" (UniqueName: \"kubernetes.io/secret/53c2fb7b-24cd-4da0-9016-e35e3223ab06-logging-loki-querier-http\") pod \"logging-loki-querier-5895d59bb8-8lqvs\" (UID: \"53c2fb7b-24cd-4da0-9016-e35e3223ab06\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-8lqvs" Nov 21 14:21:38 crc kubenswrapper[4897]: I1121 14:21:38.853480 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/53c2fb7b-24cd-4da0-9016-e35e3223ab06-logging-loki-s3\") pod \"logging-loki-querier-5895d59bb8-8lqvs\" (UID: \"53c2fb7b-24cd-4da0-9016-e35e3223ab06\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-8lqvs" Nov 21 14:21:38 crc kubenswrapper[4897]: I1121 14:21:38.875691 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-gateway-78b4f7fc55-txlqp"] Nov 21 14:21:38 crc kubenswrapper[4897]: I1121 14:21:38.886788 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-gateway-78b4f7fc55-dbkmd"] Nov 21 14:21:38 crc kubenswrapper[4897]: I1121 14:21:38.892658 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bqlw4\" (UniqueName: \"kubernetes.io/projected/53c2fb7b-24cd-4da0-9016-e35e3223ab06-kube-api-access-bqlw4\") pod \"logging-loki-querier-5895d59bb8-8lqvs\" (UID: \"53c2fb7b-24cd-4da0-9016-e35e3223ab06\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-8lqvs" Nov 21 14:21:38 crc kubenswrapper[4897]: I1121 14:21:38.923515 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-query-frontend-grpc\" (UniqueName: \"kubernetes.io/secret/a77cc078-3448-4f4b-86f4-90132ec0fd88-logging-loki-query-frontend-grpc\") pod \"logging-loki-query-frontend-84558f7c9f-vltt5\" (UID: \"a77cc078-3448-4f4b-86f4-90132ec0fd88\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-vltt5" Nov 21 14:21:38 crc kubenswrapper[4897]: I1121 14:21:38.923558 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a77cc078-3448-4f4b-86f4-90132ec0fd88-config\") pod \"logging-loki-query-frontend-84558f7c9f-vltt5\" (UID: \"a77cc078-3448-4f4b-86f4-90132ec0fd88\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-vltt5" Nov 21 14:21:38 crc kubenswrapper[4897]: I1121 14:21:38.923624 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j5m2c\" (UniqueName: \"kubernetes.io/projected/a77cc078-3448-4f4b-86f4-90132ec0fd88-kube-api-access-j5m2c\") pod \"logging-loki-query-frontend-84558f7c9f-vltt5\" (UID: \"a77cc078-3448-4f4b-86f4-90132ec0fd88\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-vltt5" Nov 21 14:21:38 crc kubenswrapper[4897]: I1121 14:21:38.923668 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-query-frontend-http\" (UniqueName: \"kubernetes.io/secret/a77cc078-3448-4f4b-86f4-90132ec0fd88-logging-loki-query-frontend-http\") pod \"logging-loki-query-frontend-84558f7c9f-vltt5\" (UID: \"a77cc078-3448-4f4b-86f4-90132ec0fd88\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-vltt5" Nov 21 14:21:38 crc kubenswrapper[4897]: I1121 14:21:38.923694 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a77cc078-3448-4f4b-86f4-90132ec0fd88-logging-loki-ca-bundle\") pod \"logging-loki-query-frontend-84558f7c9f-vltt5\" (UID: \"a77cc078-3448-4f4b-86f4-90132ec0fd88\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-vltt5" Nov 21 14:21:38 crc kubenswrapper[4897]: I1121 14:21:38.982901 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-querier-5895d59bb8-8lqvs" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.025579 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a77cc078-3448-4f4b-86f4-90132ec0fd88-logging-loki-ca-bundle\") pod \"logging-loki-query-frontend-84558f7c9f-vltt5\" (UID: \"a77cc078-3448-4f4b-86f4-90132ec0fd88\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-vltt5" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.025650 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/268847f7-dd43-4094-a01c-e5d795722e82-logging-loki-ca-bundle\") pod \"logging-loki-gateway-78b4f7fc55-txlqp\" (UID: \"268847f7-dd43-4094-a01c-e5d795722e82\") " pod="openshift-logging/logging-loki-gateway-78b4f7fc55-txlqp" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.025698 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/268847f7-dd43-4094-a01c-e5d795722e82-lokistack-gateway\") pod \"logging-loki-gateway-78b4f7fc55-txlqp\" (UID: \"268847f7-dd43-4094-a01c-e5d795722e82\") " pod="openshift-logging/logging-loki-gateway-78b4f7fc55-txlqp" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.025732 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/268847f7-dd43-4094-a01c-e5d795722e82-tls-secret\") pod \"logging-loki-gateway-78b4f7fc55-txlqp\" (UID: \"268847f7-dd43-4094-a01c-e5d795722e82\") " pod="openshift-logging/logging-loki-gateway-78b4f7fc55-txlqp" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.025765 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/6dda2ab1-6304-4116-90f3-b9219aff2664-rbac\") pod \"logging-loki-gateway-78b4f7fc55-dbkmd\" (UID: \"6dda2ab1-6304-4116-90f3-b9219aff2664\") " pod="openshift-logging/logging-loki-gateway-78b4f7fc55-dbkmd" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.025790 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6dda2ab1-6304-4116-90f3-b9219aff2664-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-78b4f7fc55-dbkmd\" (UID: \"6dda2ab1-6304-4116-90f3-b9219aff2664\") " pod="openshift-logging/logging-loki-gateway-78b4f7fc55-dbkmd" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.025821 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-query-frontend-grpc\" (UniqueName: \"kubernetes.io/secret/a77cc078-3448-4f4b-86f4-90132ec0fd88-logging-loki-query-frontend-grpc\") pod \"logging-loki-query-frontend-84558f7c9f-vltt5\" (UID: \"a77cc078-3448-4f4b-86f4-90132ec0fd88\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-vltt5" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.025839 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wkg8s\" (UniqueName: \"kubernetes.io/projected/6dda2ab1-6304-4116-90f3-b9219aff2664-kube-api-access-wkg8s\") pod \"logging-loki-gateway-78b4f7fc55-dbkmd\" (UID: \"6dda2ab1-6304-4116-90f3-b9219aff2664\") " pod="openshift-logging/logging-loki-gateway-78b4f7fc55-dbkmd" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.025863 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a77cc078-3448-4f4b-86f4-90132ec0fd88-config\") pod \"logging-loki-query-frontend-84558f7c9f-vltt5\" (UID: \"a77cc078-3448-4f4b-86f4-90132ec0fd88\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-vltt5" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.025879 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/6dda2ab1-6304-4116-90f3-b9219aff2664-tenants\") pod \"logging-loki-gateway-78b4f7fc55-dbkmd\" (UID: \"6dda2ab1-6304-4116-90f3-b9219aff2664\") " pod="openshift-logging/logging-loki-gateway-78b4f7fc55-dbkmd" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.025900 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/268847f7-dd43-4094-a01c-e5d795722e82-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-78b4f7fc55-txlqp\" (UID: \"268847f7-dd43-4094-a01c-e5d795722e82\") " pod="openshift-logging/logging-loki-gateway-78b4f7fc55-txlqp" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.025925 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/268847f7-dd43-4094-a01c-e5d795722e82-rbac\") pod \"logging-loki-gateway-78b4f7fc55-txlqp\" (UID: \"268847f7-dd43-4094-a01c-e5d795722e82\") " pod="openshift-logging/logging-loki-gateway-78b4f7fc55-txlqp" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.025942 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/268847f7-dd43-4094-a01c-e5d795722e82-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-78b4f7fc55-txlqp\" (UID: \"268847f7-dd43-4094-a01c-e5d795722e82\") " pod="openshift-logging/logging-loki-gateway-78b4f7fc55-txlqp" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.025979 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j5m2c\" (UniqueName: \"kubernetes.io/projected/a77cc078-3448-4f4b-86f4-90132ec0fd88-kube-api-access-j5m2c\") pod \"logging-loki-query-frontend-84558f7c9f-vltt5\" (UID: \"a77cc078-3448-4f4b-86f4-90132ec0fd88\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-vltt5" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.026001 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/6dda2ab1-6304-4116-90f3-b9219aff2664-lokistack-gateway\") pod \"logging-loki-gateway-78b4f7fc55-dbkmd\" (UID: \"6dda2ab1-6304-4116-90f3-b9219aff2664\") " pod="openshift-logging/logging-loki-gateway-78b4f7fc55-dbkmd" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.026018 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/268847f7-dd43-4094-a01c-e5d795722e82-tenants\") pod \"logging-loki-gateway-78b4f7fc55-txlqp\" (UID: \"268847f7-dd43-4094-a01c-e5d795722e82\") " pod="openshift-logging/logging-loki-gateway-78b4f7fc55-txlqp" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.026047 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/6dda2ab1-6304-4116-90f3-b9219aff2664-tls-secret\") pod \"logging-loki-gateway-78b4f7fc55-dbkmd\" (UID: \"6dda2ab1-6304-4116-90f3-b9219aff2664\") " pod="openshift-logging/logging-loki-gateway-78b4f7fc55-dbkmd" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.026069 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6dda2ab1-6304-4116-90f3-b9219aff2664-logging-loki-ca-bundle\") pod \"logging-loki-gateway-78b4f7fc55-dbkmd\" (UID: \"6dda2ab1-6304-4116-90f3-b9219aff2664\") " pod="openshift-logging/logging-loki-gateway-78b4f7fc55-dbkmd" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.026087 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/6dda2ab1-6304-4116-90f3-b9219aff2664-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-78b4f7fc55-dbkmd\" (UID: \"6dda2ab1-6304-4116-90f3-b9219aff2664\") " pod="openshift-logging/logging-loki-gateway-78b4f7fc55-dbkmd" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.026104 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xqd52\" (UniqueName: \"kubernetes.io/projected/268847f7-dd43-4094-a01c-e5d795722e82-kube-api-access-xqd52\") pod \"logging-loki-gateway-78b4f7fc55-txlqp\" (UID: \"268847f7-dd43-4094-a01c-e5d795722e82\") " pod="openshift-logging/logging-loki-gateway-78b4f7fc55-txlqp" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.026123 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-query-frontend-http\" (UniqueName: \"kubernetes.io/secret/a77cc078-3448-4f4b-86f4-90132ec0fd88-logging-loki-query-frontend-http\") pod \"logging-loki-query-frontend-84558f7c9f-vltt5\" (UID: \"a77cc078-3448-4f4b-86f4-90132ec0fd88\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-vltt5" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.028489 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a77cc078-3448-4f4b-86f4-90132ec0fd88-config\") pod \"logging-loki-query-frontend-84558f7c9f-vltt5\" (UID: \"a77cc078-3448-4f4b-86f4-90132ec0fd88\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-vltt5" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.029273 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a77cc078-3448-4f4b-86f4-90132ec0fd88-logging-loki-ca-bundle\") pod \"logging-loki-query-frontend-84558f7c9f-vltt5\" (UID: \"a77cc078-3448-4f4b-86f4-90132ec0fd88\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-vltt5" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.036906 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-query-frontend-http\" (UniqueName: \"kubernetes.io/secret/a77cc078-3448-4f4b-86f4-90132ec0fd88-logging-loki-query-frontend-http\") pod \"logging-loki-query-frontend-84558f7c9f-vltt5\" (UID: \"a77cc078-3448-4f4b-86f4-90132ec0fd88\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-vltt5" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.037234 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-query-frontend-grpc\" (UniqueName: \"kubernetes.io/secret/a77cc078-3448-4f4b-86f4-90132ec0fd88-logging-loki-query-frontend-grpc\") pod \"logging-loki-query-frontend-84558f7c9f-vltt5\" (UID: \"a77cc078-3448-4f4b-86f4-90132ec0fd88\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-vltt5" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.052937 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j5m2c\" (UniqueName: \"kubernetes.io/projected/a77cc078-3448-4f4b-86f4-90132ec0fd88-kube-api-access-j5m2c\") pod \"logging-loki-query-frontend-84558f7c9f-vltt5\" (UID: \"a77cc078-3448-4f4b-86f4-90132ec0fd88\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-vltt5" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.105070 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-vltt5" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.134666 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/268847f7-dd43-4094-a01c-e5d795722e82-tls-secret\") pod \"logging-loki-gateway-78b4f7fc55-txlqp\" (UID: \"268847f7-dd43-4094-a01c-e5d795722e82\") " pod="openshift-logging/logging-loki-gateway-78b4f7fc55-txlqp" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.134701 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/6dda2ab1-6304-4116-90f3-b9219aff2664-rbac\") pod \"logging-loki-gateway-78b4f7fc55-dbkmd\" (UID: \"6dda2ab1-6304-4116-90f3-b9219aff2664\") " pod="openshift-logging/logging-loki-gateway-78b4f7fc55-dbkmd" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.134726 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6dda2ab1-6304-4116-90f3-b9219aff2664-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-78b4f7fc55-dbkmd\" (UID: \"6dda2ab1-6304-4116-90f3-b9219aff2664\") " pod="openshift-logging/logging-loki-gateway-78b4f7fc55-dbkmd" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.134756 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wkg8s\" (UniqueName: \"kubernetes.io/projected/6dda2ab1-6304-4116-90f3-b9219aff2664-kube-api-access-wkg8s\") pod \"logging-loki-gateway-78b4f7fc55-dbkmd\" (UID: \"6dda2ab1-6304-4116-90f3-b9219aff2664\") " pod="openshift-logging/logging-loki-gateway-78b4f7fc55-dbkmd" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.134775 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/6dda2ab1-6304-4116-90f3-b9219aff2664-tenants\") pod \"logging-loki-gateway-78b4f7fc55-dbkmd\" (UID: \"6dda2ab1-6304-4116-90f3-b9219aff2664\") " pod="openshift-logging/logging-loki-gateway-78b4f7fc55-dbkmd" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.134790 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/268847f7-dd43-4094-a01c-e5d795722e82-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-78b4f7fc55-txlqp\" (UID: \"268847f7-dd43-4094-a01c-e5d795722e82\") " pod="openshift-logging/logging-loki-gateway-78b4f7fc55-txlqp" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.134815 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/268847f7-dd43-4094-a01c-e5d795722e82-rbac\") pod \"logging-loki-gateway-78b4f7fc55-txlqp\" (UID: \"268847f7-dd43-4094-a01c-e5d795722e82\") " pod="openshift-logging/logging-loki-gateway-78b4f7fc55-txlqp" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.134833 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/268847f7-dd43-4094-a01c-e5d795722e82-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-78b4f7fc55-txlqp\" (UID: \"268847f7-dd43-4094-a01c-e5d795722e82\") " pod="openshift-logging/logging-loki-gateway-78b4f7fc55-txlqp" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.134854 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/6dda2ab1-6304-4116-90f3-b9219aff2664-lokistack-gateway\") pod \"logging-loki-gateway-78b4f7fc55-dbkmd\" (UID: \"6dda2ab1-6304-4116-90f3-b9219aff2664\") " pod="openshift-logging/logging-loki-gateway-78b4f7fc55-dbkmd" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.134869 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/268847f7-dd43-4094-a01c-e5d795722e82-tenants\") pod \"logging-loki-gateway-78b4f7fc55-txlqp\" (UID: \"268847f7-dd43-4094-a01c-e5d795722e82\") " pod="openshift-logging/logging-loki-gateway-78b4f7fc55-txlqp" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.134887 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/6dda2ab1-6304-4116-90f3-b9219aff2664-tls-secret\") pod \"logging-loki-gateway-78b4f7fc55-dbkmd\" (UID: \"6dda2ab1-6304-4116-90f3-b9219aff2664\") " pod="openshift-logging/logging-loki-gateway-78b4f7fc55-dbkmd" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.134908 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6dda2ab1-6304-4116-90f3-b9219aff2664-logging-loki-ca-bundle\") pod \"logging-loki-gateway-78b4f7fc55-dbkmd\" (UID: \"6dda2ab1-6304-4116-90f3-b9219aff2664\") " pod="openshift-logging/logging-loki-gateway-78b4f7fc55-dbkmd" Nov 21 14:21:39 crc kubenswrapper[4897]: E1121 14:21:39.135363 4897 secret.go:188] Couldn't get secret openshift-logging/logging-loki-gateway-http: secret "logging-loki-gateway-http" not found Nov 21 14:21:39 crc kubenswrapper[4897]: E1121 14:21:39.135413 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/268847f7-dd43-4094-a01c-e5d795722e82-tls-secret podName:268847f7-dd43-4094-a01c-e5d795722e82 nodeName:}" failed. No retries permitted until 2025-11-21 14:21:39.635395788 +0000 UTC m=+776.919989253 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-secret" (UniqueName: "kubernetes.io/secret/268847f7-dd43-4094-a01c-e5d795722e82-tls-secret") pod "logging-loki-gateway-78b4f7fc55-txlqp" (UID: "268847f7-dd43-4094-a01c-e5d795722e82") : secret "logging-loki-gateway-http" not found Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.136394 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/6dda2ab1-6304-4116-90f3-b9219aff2664-rbac\") pod \"logging-loki-gateway-78b4f7fc55-dbkmd\" (UID: \"6dda2ab1-6304-4116-90f3-b9219aff2664\") " pod="openshift-logging/logging-loki-gateway-78b4f7fc55-dbkmd" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.136914 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6dda2ab1-6304-4116-90f3-b9219aff2664-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-78b4f7fc55-dbkmd\" (UID: \"6dda2ab1-6304-4116-90f3-b9219aff2664\") " pod="openshift-logging/logging-loki-gateway-78b4f7fc55-dbkmd" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.137335 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/6dda2ab1-6304-4116-90f3-b9219aff2664-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-78b4f7fc55-dbkmd\" (UID: \"6dda2ab1-6304-4116-90f3-b9219aff2664\") " pod="openshift-logging/logging-loki-gateway-78b4f7fc55-dbkmd" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.137372 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xqd52\" (UniqueName: \"kubernetes.io/projected/268847f7-dd43-4094-a01c-e5d795722e82-kube-api-access-xqd52\") pod \"logging-loki-gateway-78b4f7fc55-txlqp\" (UID: \"268847f7-dd43-4094-a01c-e5d795722e82\") " pod="openshift-logging/logging-loki-gateway-78b4f7fc55-txlqp" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.137424 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/268847f7-dd43-4094-a01c-e5d795722e82-logging-loki-ca-bundle\") pod \"logging-loki-gateway-78b4f7fc55-txlqp\" (UID: \"268847f7-dd43-4094-a01c-e5d795722e82\") " pod="openshift-logging/logging-loki-gateway-78b4f7fc55-txlqp" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.141882 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/6dda2ab1-6304-4116-90f3-b9219aff2664-lokistack-gateway\") pod \"logging-loki-gateway-78b4f7fc55-dbkmd\" (UID: \"6dda2ab1-6304-4116-90f3-b9219aff2664\") " pod="openshift-logging/logging-loki-gateway-78b4f7fc55-dbkmd" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.143388 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/268847f7-dd43-4094-a01c-e5d795722e82-rbac\") pod \"logging-loki-gateway-78b4f7fc55-txlqp\" (UID: \"268847f7-dd43-4094-a01c-e5d795722e82\") " pod="openshift-logging/logging-loki-gateway-78b4f7fc55-txlqp" Nov 21 14:21:39 crc kubenswrapper[4897]: E1121 14:21:39.144211 4897 secret.go:188] Couldn't get secret openshift-logging/logging-loki-gateway-http: secret "logging-loki-gateway-http" not found Nov 21 14:21:39 crc kubenswrapper[4897]: E1121 14:21:39.144282 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6dda2ab1-6304-4116-90f3-b9219aff2664-tls-secret podName:6dda2ab1-6304-4116-90f3-b9219aff2664 nodeName:}" failed. No retries permitted until 2025-11-21 14:21:39.644263005 +0000 UTC m=+776.928856480 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-secret" (UniqueName: "kubernetes.io/secret/6dda2ab1-6304-4116-90f3-b9219aff2664-tls-secret") pod "logging-loki-gateway-78b4f7fc55-dbkmd" (UID: "6dda2ab1-6304-4116-90f3-b9219aff2664") : secret "logging-loki-gateway-http" not found Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.144843 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/268847f7-dd43-4094-a01c-e5d795722e82-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-78b4f7fc55-txlqp\" (UID: \"268847f7-dd43-4094-a01c-e5d795722e82\") " pod="openshift-logging/logging-loki-gateway-78b4f7fc55-txlqp" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.146369 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/6dda2ab1-6304-4116-90f3-b9219aff2664-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-78b4f7fc55-dbkmd\" (UID: \"6dda2ab1-6304-4116-90f3-b9219aff2664\") " pod="openshift-logging/logging-loki-gateway-78b4f7fc55-dbkmd" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.146906 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6dda2ab1-6304-4116-90f3-b9219aff2664-logging-loki-ca-bundle\") pod \"logging-loki-gateway-78b4f7fc55-dbkmd\" (UID: \"6dda2ab1-6304-4116-90f3-b9219aff2664\") " pod="openshift-logging/logging-loki-gateway-78b4f7fc55-dbkmd" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.147948 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/268847f7-dd43-4094-a01c-e5d795722e82-logging-loki-ca-bundle\") pod \"logging-loki-gateway-78b4f7fc55-txlqp\" (UID: \"268847f7-dd43-4094-a01c-e5d795722e82\") " pod="openshift-logging/logging-loki-gateway-78b4f7fc55-txlqp" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.149277 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/268847f7-dd43-4094-a01c-e5d795722e82-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-78b4f7fc55-txlqp\" (UID: \"268847f7-dd43-4094-a01c-e5d795722e82\") " pod="openshift-logging/logging-loki-gateway-78b4f7fc55-txlqp" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.153293 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/6dda2ab1-6304-4116-90f3-b9219aff2664-tenants\") pod \"logging-loki-gateway-78b4f7fc55-dbkmd\" (UID: \"6dda2ab1-6304-4116-90f3-b9219aff2664\") " pod="openshift-logging/logging-loki-gateway-78b4f7fc55-dbkmd" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.154061 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/268847f7-dd43-4094-a01c-e5d795722e82-lokistack-gateway\") pod \"logging-loki-gateway-78b4f7fc55-txlqp\" (UID: \"268847f7-dd43-4094-a01c-e5d795722e82\") " pod="openshift-logging/logging-loki-gateway-78b4f7fc55-txlqp" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.156942 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/268847f7-dd43-4094-a01c-e5d795722e82-lokistack-gateway\") pod \"logging-loki-gateway-78b4f7fc55-txlqp\" (UID: \"268847f7-dd43-4094-a01c-e5d795722e82\") " pod="openshift-logging/logging-loki-gateway-78b4f7fc55-txlqp" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.158160 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/268847f7-dd43-4094-a01c-e5d795722e82-tenants\") pod \"logging-loki-gateway-78b4f7fc55-txlqp\" (UID: \"268847f7-dd43-4094-a01c-e5d795722e82\") " pod="openshift-logging/logging-loki-gateway-78b4f7fc55-txlqp" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.164008 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wkg8s\" (UniqueName: \"kubernetes.io/projected/6dda2ab1-6304-4116-90f3-b9219aff2664-kube-api-access-wkg8s\") pod \"logging-loki-gateway-78b4f7fc55-dbkmd\" (UID: \"6dda2ab1-6304-4116-90f3-b9219aff2664\") " pod="openshift-logging/logging-loki-gateway-78b4f7fc55-dbkmd" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.165375 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xqd52\" (UniqueName: \"kubernetes.io/projected/268847f7-dd43-4094-a01c-e5d795722e82-kube-api-access-xqd52\") pod \"logging-loki-gateway-78b4f7fc55-txlqp\" (UID: \"268847f7-dd43-4094-a01c-e5d795722e82\") " pod="openshift-logging/logging-loki-gateway-78b4f7fc55-txlqp" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.324750 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-distributor-76cc67bf56-m64dx"] Nov 21 14:21:39 crc kubenswrapper[4897]: W1121 14:21:39.331097 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd3f767a5_605a_4b21_a15a_68cde314d5e6.slice/crio-30728ff8226f87524cc4b2617d6fec8b807de6dd84c961fc6f8380af23826854 WatchSource:0}: Error finding container 30728ff8226f87524cc4b2617d6fec8b807de6dd84c961fc6f8380af23826854: Status 404 returned error can't find the container with id 30728ff8226f87524cc4b2617d6fec8b807de6dd84c961fc6f8380af23826854 Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.496478 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-querier-5895d59bb8-8lqvs"] Nov 21 14:21:39 crc kubenswrapper[4897]: W1121 14:21:39.504727 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod53c2fb7b_24cd_4da0_9016_e35e3223ab06.slice/crio-d0e891bbde2b1d678fca4c5dc5b7ccc6a3c4715333eb32997c6a3ae8d2c8685c WatchSource:0}: Error finding container d0e891bbde2b1d678fca4c5dc5b7ccc6a3c4715333eb32997c6a3ae8d2c8685c: Status 404 returned error can't find the container with id d0e891bbde2b1d678fca4c5dc5b7ccc6a3c4715333eb32997c6a3ae8d2c8685c Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.610100 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-query-frontend-84558f7c9f-vltt5"] Nov 21 14:21:39 crc kubenswrapper[4897]: W1121 14:21:39.614617 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda77cc078_3448_4f4b_86f4_90132ec0fd88.slice/crio-8f142860e9ddef0af2fc98c7308e6bd76ab694b474d8f47eb85a8c9f50e8b143 WatchSource:0}: Error finding container 8f142860e9ddef0af2fc98c7308e6bd76ab694b474d8f47eb85a8c9f50e8b143: Status 404 returned error can't find the container with id 8f142860e9ddef0af2fc98c7308e6bd76ab694b474d8f47eb85a8c9f50e8b143 Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.640288 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-ingester-0"] Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.641088 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-ingester-0" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.643589 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-ingester-http" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.643771 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-ingester-grpc" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.655271 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-ingester-0"] Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.665130 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/268847f7-dd43-4094-a01c-e5d795722e82-tls-secret\") pod \"logging-loki-gateway-78b4f7fc55-txlqp\" (UID: \"268847f7-dd43-4094-a01c-e5d795722e82\") " pod="openshift-logging/logging-loki-gateway-78b4f7fc55-txlqp" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.665224 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-58f9c9d2-2f87-425a-a32e-4717701acf0b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-58f9c9d2-2f87-425a-a32e-4717701acf0b\") pod \"logging-loki-ingester-0\" (UID: \"69b9e549-607c-4707-9ff5-2dddbcc54b18\") " pod="openshift-logging/logging-loki-ingester-0" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.665247 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/6dda2ab1-6304-4116-90f3-b9219aff2664-tls-secret\") pod \"logging-loki-gateway-78b4f7fc55-dbkmd\" (UID: \"6dda2ab1-6304-4116-90f3-b9219aff2664\") " pod="openshift-logging/logging-loki-gateway-78b4f7fc55-dbkmd" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.668560 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/6dda2ab1-6304-4116-90f3-b9219aff2664-tls-secret\") pod \"logging-loki-gateway-78b4f7fc55-dbkmd\" (UID: \"6dda2ab1-6304-4116-90f3-b9219aff2664\") " pod="openshift-logging/logging-loki-gateway-78b4f7fc55-dbkmd" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.671759 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/268847f7-dd43-4094-a01c-e5d795722e82-tls-secret\") pod \"logging-loki-gateway-78b4f7fc55-txlqp\" (UID: \"268847f7-dd43-4094-a01c-e5d795722e82\") " pod="openshift-logging/logging-loki-gateway-78b4f7fc55-txlqp" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.722708 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-compactor-0"] Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.724193 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-compactor-0" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.726488 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-compactor-http" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.726778 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-compactor-grpc" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.737043 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-compactor-0"] Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.767149 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-58f9c9d2-2f87-425a-a32e-4717701acf0b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-58f9c9d2-2f87-425a-a32e-4717701acf0b\") pod \"logging-loki-ingester-0\" (UID: \"69b9e549-607c-4707-9ff5-2dddbcc54b18\") " pod="openshift-logging/logging-loki-ingester-0" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.767411 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/69b9e549-607c-4707-9ff5-2dddbcc54b18-config\") pod \"logging-loki-ingester-0\" (UID: \"69b9e549-607c-4707-9ff5-2dddbcc54b18\") " pod="openshift-logging/logging-loki-ingester-0" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.767542 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ingester-grpc\" (UniqueName: \"kubernetes.io/secret/69b9e549-607c-4707-9ff5-2dddbcc54b18-logging-loki-ingester-grpc\") pod \"logging-loki-ingester-0\" (UID: \"69b9e549-607c-4707-9ff5-2dddbcc54b18\") " pod="openshift-logging/logging-loki-ingester-0" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.767653 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/69b9e549-607c-4707-9ff5-2dddbcc54b18-logging-loki-ca-bundle\") pod \"logging-loki-ingester-0\" (UID: \"69b9e549-607c-4707-9ff5-2dddbcc54b18\") " pod="openshift-logging/logging-loki-ingester-0" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.767775 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-65swl\" (UniqueName: \"kubernetes.io/projected/69b9e549-607c-4707-9ff5-2dddbcc54b18-kube-api-access-65swl\") pod \"logging-loki-ingester-0\" (UID: \"69b9e549-607c-4707-9ff5-2dddbcc54b18\") " pod="openshift-logging/logging-loki-ingester-0" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.767874 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ingester-http\" (UniqueName: \"kubernetes.io/secret/69b9e549-607c-4707-9ff5-2dddbcc54b18-logging-loki-ingester-http\") pod \"logging-loki-ingester-0\" (UID: \"69b9e549-607c-4707-9ff5-2dddbcc54b18\") " pod="openshift-logging/logging-loki-ingester-0" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.767997 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-579e4bea-0fda-4109-bd9d-08fcde44d914\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-579e4bea-0fda-4109-bd9d-08fcde44d914\") pod \"logging-loki-ingester-0\" (UID: \"69b9e549-607c-4707-9ff5-2dddbcc54b18\") " pod="openshift-logging/logging-loki-ingester-0" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.768125 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/69b9e549-607c-4707-9ff5-2dddbcc54b18-logging-loki-s3\") pod \"logging-loki-ingester-0\" (UID: \"69b9e549-607c-4707-9ff5-2dddbcc54b18\") " pod="openshift-logging/logging-loki-ingester-0" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.770689 4897 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.770740 4897 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-58f9c9d2-2f87-425a-a32e-4717701acf0b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-58f9c9d2-2f87-425a-a32e-4717701acf0b\") pod \"logging-loki-ingester-0\" (UID: \"69b9e549-607c-4707-9ff5-2dddbcc54b18\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/edd1319584f5d7c1a6d492fee84b23dc551b2380fd1395fc1c452cb15602de63/globalmount\"" pod="openshift-logging/logging-loki-ingester-0" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.796936 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-index-gateway-0"] Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.798048 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-index-gateway-0" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.800344 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-58f9c9d2-2f87-425a-a32e-4717701acf0b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-58f9c9d2-2f87-425a-a32e-4717701acf0b\") pod \"logging-loki-ingester-0\" (UID: \"69b9e549-607c-4707-9ff5-2dddbcc54b18\") " pod="openshift-logging/logging-loki-ingester-0" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.802323 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-index-gateway-grpc" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.802865 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-index-gateway-http" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.808990 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-index-gateway-0"] Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.812523 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-gateway-78b4f7fc55-dbkmd" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.826121 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-gateway-78b4f7fc55-txlqp" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.870249 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-65swl\" (UniqueName: \"kubernetes.io/projected/69b9e549-607c-4707-9ff5-2dddbcc54b18-kube-api-access-65swl\") pod \"logging-loki-ingester-0\" (UID: \"69b9e549-607c-4707-9ff5-2dddbcc54b18\") " pod="openshift-logging/logging-loki-ingester-0" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.870303 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ingester-http\" (UniqueName: \"kubernetes.io/secret/69b9e549-607c-4707-9ff5-2dddbcc54b18-logging-loki-ingester-http\") pod \"logging-loki-ingester-0\" (UID: \"69b9e549-607c-4707-9ff5-2dddbcc54b18\") " pod="openshift-logging/logging-loki-ingester-0" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.870336 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9241fda3-de39-4efb-8825-a5d706a4dcc9-logging-loki-ca-bundle\") pod \"logging-loki-compactor-0\" (UID: \"9241fda3-de39-4efb-8825-a5d706a4dcc9\") " pod="openshift-logging/logging-loki-compactor-0" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.870372 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-579e4bea-0fda-4109-bd9d-08fcde44d914\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-579e4bea-0fda-4109-bd9d-08fcde44d914\") pod \"logging-loki-ingester-0\" (UID: \"69b9e549-607c-4707-9ff5-2dddbcc54b18\") " pod="openshift-logging/logging-loki-ingester-0" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.870396 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cwz72\" (UniqueName: \"kubernetes.io/projected/9241fda3-de39-4efb-8825-a5d706a4dcc9-kube-api-access-cwz72\") pod \"logging-loki-compactor-0\" (UID: \"9241fda3-de39-4efb-8825-a5d706a4dcc9\") " pod="openshift-logging/logging-loki-compactor-0" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.870600 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/69b9e549-607c-4707-9ff5-2dddbcc54b18-logging-loki-s3\") pod \"logging-loki-ingester-0\" (UID: \"69b9e549-607c-4707-9ff5-2dddbcc54b18\") " pod="openshift-logging/logging-loki-ingester-0" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.870656 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/9241fda3-de39-4efb-8825-a5d706a4dcc9-logging-loki-s3\") pod \"logging-loki-compactor-0\" (UID: \"9241fda3-de39-4efb-8825-a5d706a4dcc9\") " pod="openshift-logging/logging-loki-compactor-0" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.870690 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-compactor-grpc\" (UniqueName: \"kubernetes.io/secret/9241fda3-de39-4efb-8825-a5d706a4dcc9-logging-loki-compactor-grpc\") pod \"logging-loki-compactor-0\" (UID: \"9241fda3-de39-4efb-8825-a5d706a4dcc9\") " pod="openshift-logging/logging-loki-compactor-0" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.870718 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9241fda3-de39-4efb-8825-a5d706a4dcc9-config\") pod \"logging-loki-compactor-0\" (UID: \"9241fda3-de39-4efb-8825-a5d706a4dcc9\") " pod="openshift-logging/logging-loki-compactor-0" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.870973 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/69b9e549-607c-4707-9ff5-2dddbcc54b18-config\") pod \"logging-loki-ingester-0\" (UID: \"69b9e549-607c-4707-9ff5-2dddbcc54b18\") " pod="openshift-logging/logging-loki-ingester-0" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.871054 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-756f1856-2b34-415c-9c29-0a61b4c29cfc\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-756f1856-2b34-415c-9c29-0a61b4c29cfc\") pod \"logging-loki-compactor-0\" (UID: \"9241fda3-de39-4efb-8825-a5d706a4dcc9\") " pod="openshift-logging/logging-loki-compactor-0" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.871094 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ingester-grpc\" (UniqueName: \"kubernetes.io/secret/69b9e549-607c-4707-9ff5-2dddbcc54b18-logging-loki-ingester-grpc\") pod \"logging-loki-ingester-0\" (UID: \"69b9e549-607c-4707-9ff5-2dddbcc54b18\") " pod="openshift-logging/logging-loki-ingester-0" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.871166 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-compactor-http\" (UniqueName: \"kubernetes.io/secret/9241fda3-de39-4efb-8825-a5d706a4dcc9-logging-loki-compactor-http\") pod \"logging-loki-compactor-0\" (UID: \"9241fda3-de39-4efb-8825-a5d706a4dcc9\") " pod="openshift-logging/logging-loki-compactor-0" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.871246 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/69b9e549-607c-4707-9ff5-2dddbcc54b18-logging-loki-ca-bundle\") pod \"logging-loki-ingester-0\" (UID: \"69b9e549-607c-4707-9ff5-2dddbcc54b18\") " pod="openshift-logging/logging-loki-ingester-0" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.872877 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/69b9e549-607c-4707-9ff5-2dddbcc54b18-config\") pod \"logging-loki-ingester-0\" (UID: \"69b9e549-607c-4707-9ff5-2dddbcc54b18\") " pod="openshift-logging/logging-loki-ingester-0" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.873558 4897 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.873620 4897 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-579e4bea-0fda-4109-bd9d-08fcde44d914\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-579e4bea-0fda-4109-bd9d-08fcde44d914\") pod \"logging-loki-ingester-0\" (UID: \"69b9e549-607c-4707-9ff5-2dddbcc54b18\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/364be16a9da25cb4082b7b8c87d3dcb0cf8d373f8849c58aab4db240021e8350/globalmount\"" pod="openshift-logging/logging-loki-ingester-0" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.873945 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/69b9e549-607c-4707-9ff5-2dddbcc54b18-logging-loki-ca-bundle\") pod \"logging-loki-ingester-0\" (UID: \"69b9e549-607c-4707-9ff5-2dddbcc54b18\") " pod="openshift-logging/logging-loki-ingester-0" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.875291 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/69b9e549-607c-4707-9ff5-2dddbcc54b18-logging-loki-s3\") pod \"logging-loki-ingester-0\" (UID: \"69b9e549-607c-4707-9ff5-2dddbcc54b18\") " pod="openshift-logging/logging-loki-ingester-0" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.877847 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ingester-http\" (UniqueName: \"kubernetes.io/secret/69b9e549-607c-4707-9ff5-2dddbcc54b18-logging-loki-ingester-http\") pod \"logging-loki-ingester-0\" (UID: \"69b9e549-607c-4707-9ff5-2dddbcc54b18\") " pod="openshift-logging/logging-loki-ingester-0" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.891389 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-65swl\" (UniqueName: \"kubernetes.io/projected/69b9e549-607c-4707-9ff5-2dddbcc54b18-kube-api-access-65swl\") pod \"logging-loki-ingester-0\" (UID: \"69b9e549-607c-4707-9ff5-2dddbcc54b18\") " pod="openshift-logging/logging-loki-ingester-0" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.891719 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ingester-grpc\" (UniqueName: \"kubernetes.io/secret/69b9e549-607c-4707-9ff5-2dddbcc54b18-logging-loki-ingester-grpc\") pod \"logging-loki-ingester-0\" (UID: \"69b9e549-607c-4707-9ff5-2dddbcc54b18\") " pod="openshift-logging/logging-loki-ingester-0" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.900250 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-579e4bea-0fda-4109-bd9d-08fcde44d914\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-579e4bea-0fda-4109-bd9d-08fcde44d914\") pod \"logging-loki-ingester-0\" (UID: \"69b9e549-607c-4707-9ff5-2dddbcc54b18\") " pod="openshift-logging/logging-loki-ingester-0" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.959069 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-ingester-0" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.972746 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/9241fda3-de39-4efb-8825-a5d706a4dcc9-logging-loki-s3\") pod \"logging-loki-compactor-0\" (UID: \"9241fda3-de39-4efb-8825-a5d706a4dcc9\") " pod="openshift-logging/logging-loki-compactor-0" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.972831 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-index-gateway-http\" (UniqueName: \"kubernetes.io/secret/becf18d3-fc70-4b68-b16e-f03bbf0beb84-logging-loki-index-gateway-http\") pod \"logging-loki-index-gateway-0\" (UID: \"becf18d3-fc70-4b68-b16e-f03bbf0beb84\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.972894 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-756f1856-2b34-415c-9c29-0a61b4c29cfc\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-756f1856-2b34-415c-9c29-0a61b4c29cfc\") pod \"logging-loki-compactor-0\" (UID: \"9241fda3-de39-4efb-8825-a5d706a4dcc9\") " pod="openshift-logging/logging-loki-compactor-0" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.972938 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-c988cb84-c0c1-4fa9-be32-71560c8f12dc\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c988cb84-c0c1-4fa9-be32-71560c8f12dc\") pod \"logging-loki-index-gateway-0\" (UID: \"becf18d3-fc70-4b68-b16e-f03bbf0beb84\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.972966 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9241fda3-de39-4efb-8825-a5d706a4dcc9-logging-loki-ca-bundle\") pod \"logging-loki-compactor-0\" (UID: \"9241fda3-de39-4efb-8825-a5d706a4dcc9\") " pod="openshift-logging/logging-loki-compactor-0" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.972995 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cwz72\" (UniqueName: \"kubernetes.io/projected/9241fda3-de39-4efb-8825-a5d706a4dcc9-kube-api-access-cwz72\") pod \"logging-loki-compactor-0\" (UID: \"9241fda3-de39-4efb-8825-a5d706a4dcc9\") " pod="openshift-logging/logging-loki-compactor-0" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.973024 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/becf18d3-fc70-4b68-b16e-f03bbf0beb84-config\") pod \"logging-loki-index-gateway-0\" (UID: \"becf18d3-fc70-4b68-b16e-f03bbf0beb84\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.973058 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-compactor-grpc\" (UniqueName: \"kubernetes.io/secret/9241fda3-de39-4efb-8825-a5d706a4dcc9-logging-loki-compactor-grpc\") pod \"logging-loki-compactor-0\" (UID: \"9241fda3-de39-4efb-8825-a5d706a4dcc9\") " pod="openshift-logging/logging-loki-compactor-0" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.973079 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9241fda3-de39-4efb-8825-a5d706a4dcc9-config\") pod \"logging-loki-compactor-0\" (UID: \"9241fda3-de39-4efb-8825-a5d706a4dcc9\") " pod="openshift-logging/logging-loki-compactor-0" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.973107 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/becf18d3-fc70-4b68-b16e-f03bbf0beb84-logging-loki-s3\") pod \"logging-loki-index-gateway-0\" (UID: \"becf18d3-fc70-4b68-b16e-f03bbf0beb84\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.973157 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-index-gateway-grpc\" (UniqueName: \"kubernetes.io/secret/becf18d3-fc70-4b68-b16e-f03bbf0beb84-logging-loki-index-gateway-grpc\") pod \"logging-loki-index-gateway-0\" (UID: \"becf18d3-fc70-4b68-b16e-f03bbf0beb84\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.973194 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-compactor-http\" (UniqueName: \"kubernetes.io/secret/9241fda3-de39-4efb-8825-a5d706a4dcc9-logging-loki-compactor-http\") pod \"logging-loki-compactor-0\" (UID: \"9241fda3-de39-4efb-8825-a5d706a4dcc9\") " pod="openshift-logging/logging-loki-compactor-0" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.973227 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjhrl\" (UniqueName: \"kubernetes.io/projected/becf18d3-fc70-4b68-b16e-f03bbf0beb84-kube-api-access-hjhrl\") pod \"logging-loki-index-gateway-0\" (UID: \"becf18d3-fc70-4b68-b16e-f03bbf0beb84\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.973315 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/becf18d3-fc70-4b68-b16e-f03bbf0beb84-logging-loki-ca-bundle\") pod \"logging-loki-index-gateway-0\" (UID: \"becf18d3-fc70-4b68-b16e-f03bbf0beb84\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.974916 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9241fda3-de39-4efb-8825-a5d706a4dcc9-config\") pod \"logging-loki-compactor-0\" (UID: \"9241fda3-de39-4efb-8825-a5d706a4dcc9\") " pod="openshift-logging/logging-loki-compactor-0" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.975541 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9241fda3-de39-4efb-8825-a5d706a4dcc9-logging-loki-ca-bundle\") pod \"logging-loki-compactor-0\" (UID: \"9241fda3-de39-4efb-8825-a5d706a4dcc9\") " pod="openshift-logging/logging-loki-compactor-0" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.979369 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-compactor-grpc\" (UniqueName: \"kubernetes.io/secret/9241fda3-de39-4efb-8825-a5d706a4dcc9-logging-loki-compactor-grpc\") pod \"logging-loki-compactor-0\" (UID: \"9241fda3-de39-4efb-8825-a5d706a4dcc9\") " pod="openshift-logging/logging-loki-compactor-0" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.980668 4897 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.980715 4897 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-756f1856-2b34-415c-9c29-0a61b4c29cfc\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-756f1856-2b34-415c-9c29-0a61b4c29cfc\") pod \"logging-loki-compactor-0\" (UID: \"9241fda3-de39-4efb-8825-a5d706a4dcc9\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/25735e522a5217e27464f78a0b94a9c91adfda307da9e7a40c13e2fa8b3b75d9/globalmount\"" pod="openshift-logging/logging-loki-compactor-0" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.980723 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/9241fda3-de39-4efb-8825-a5d706a4dcc9-logging-loki-s3\") pod \"logging-loki-compactor-0\" (UID: \"9241fda3-de39-4efb-8825-a5d706a4dcc9\") " pod="openshift-logging/logging-loki-compactor-0" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.984044 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-compactor-http\" (UniqueName: \"kubernetes.io/secret/9241fda3-de39-4efb-8825-a5d706a4dcc9-logging-loki-compactor-http\") pod \"logging-loki-compactor-0\" (UID: \"9241fda3-de39-4efb-8825-a5d706a4dcc9\") " pod="openshift-logging/logging-loki-compactor-0" Nov 21 14:21:39 crc kubenswrapper[4897]: I1121 14:21:39.993036 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cwz72\" (UniqueName: \"kubernetes.io/projected/9241fda3-de39-4efb-8825-a5d706a4dcc9-kube-api-access-cwz72\") pod \"logging-loki-compactor-0\" (UID: \"9241fda3-de39-4efb-8825-a5d706a4dcc9\") " pod="openshift-logging/logging-loki-compactor-0" Nov 21 14:21:40 crc kubenswrapper[4897]: I1121 14:21:40.007008 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-756f1856-2b34-415c-9c29-0a61b4c29cfc\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-756f1856-2b34-415c-9c29-0a61b4c29cfc\") pod \"logging-loki-compactor-0\" (UID: \"9241fda3-de39-4efb-8825-a5d706a4dcc9\") " pod="openshift-logging/logging-loki-compactor-0" Nov 21 14:21:40 crc kubenswrapper[4897]: I1121 14:21:40.074460 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/becf18d3-fc70-4b68-b16e-f03bbf0beb84-config\") pod \"logging-loki-index-gateway-0\" (UID: \"becf18d3-fc70-4b68-b16e-f03bbf0beb84\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 21 14:21:40 crc kubenswrapper[4897]: I1121 14:21:40.074569 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/becf18d3-fc70-4b68-b16e-f03bbf0beb84-logging-loki-s3\") pod \"logging-loki-index-gateway-0\" (UID: \"becf18d3-fc70-4b68-b16e-f03bbf0beb84\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 21 14:21:40 crc kubenswrapper[4897]: I1121 14:21:40.074612 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-index-gateway-grpc\" (UniqueName: \"kubernetes.io/secret/becf18d3-fc70-4b68-b16e-f03bbf0beb84-logging-loki-index-gateway-grpc\") pod \"logging-loki-index-gateway-0\" (UID: \"becf18d3-fc70-4b68-b16e-f03bbf0beb84\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 21 14:21:40 crc kubenswrapper[4897]: I1121 14:21:40.074654 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hjhrl\" (UniqueName: \"kubernetes.io/projected/becf18d3-fc70-4b68-b16e-f03bbf0beb84-kube-api-access-hjhrl\") pod \"logging-loki-index-gateway-0\" (UID: \"becf18d3-fc70-4b68-b16e-f03bbf0beb84\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 21 14:21:40 crc kubenswrapper[4897]: I1121 14:21:40.074700 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/becf18d3-fc70-4b68-b16e-f03bbf0beb84-logging-loki-ca-bundle\") pod \"logging-loki-index-gateway-0\" (UID: \"becf18d3-fc70-4b68-b16e-f03bbf0beb84\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 21 14:21:40 crc kubenswrapper[4897]: I1121 14:21:40.074756 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-index-gateway-http\" (UniqueName: \"kubernetes.io/secret/becf18d3-fc70-4b68-b16e-f03bbf0beb84-logging-loki-index-gateway-http\") pod \"logging-loki-index-gateway-0\" (UID: \"becf18d3-fc70-4b68-b16e-f03bbf0beb84\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 21 14:21:40 crc kubenswrapper[4897]: I1121 14:21:40.074805 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-c988cb84-c0c1-4fa9-be32-71560c8f12dc\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c988cb84-c0c1-4fa9-be32-71560c8f12dc\") pod \"logging-loki-index-gateway-0\" (UID: \"becf18d3-fc70-4b68-b16e-f03bbf0beb84\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 21 14:21:40 crc kubenswrapper[4897]: I1121 14:21:40.075591 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/becf18d3-fc70-4b68-b16e-f03bbf0beb84-config\") pod \"logging-loki-index-gateway-0\" (UID: \"becf18d3-fc70-4b68-b16e-f03bbf0beb84\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 21 14:21:40 crc kubenswrapper[4897]: I1121 14:21:40.077116 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/becf18d3-fc70-4b68-b16e-f03bbf0beb84-logging-loki-ca-bundle\") pod \"logging-loki-index-gateway-0\" (UID: \"becf18d3-fc70-4b68-b16e-f03bbf0beb84\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 21 14:21:40 crc kubenswrapper[4897]: I1121 14:21:40.078472 4897 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 21 14:21:40 crc kubenswrapper[4897]: I1121 14:21:40.078534 4897 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-c988cb84-c0c1-4fa9-be32-71560c8f12dc\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c988cb84-c0c1-4fa9-be32-71560c8f12dc\") pod \"logging-loki-index-gateway-0\" (UID: \"becf18d3-fc70-4b68-b16e-f03bbf0beb84\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/4b8c61d76df4718c57b92f917b36622a4c32ff6b38e7b35ac6e36f1e317f8a16/globalmount\"" pod="openshift-logging/logging-loki-index-gateway-0" Nov 21 14:21:40 crc kubenswrapper[4897]: I1121 14:21:40.078479 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/becf18d3-fc70-4b68-b16e-f03bbf0beb84-logging-loki-s3\") pod \"logging-loki-index-gateway-0\" (UID: \"becf18d3-fc70-4b68-b16e-f03bbf0beb84\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 21 14:21:40 crc kubenswrapper[4897]: I1121 14:21:40.078689 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-index-gateway-grpc\" (UniqueName: \"kubernetes.io/secret/becf18d3-fc70-4b68-b16e-f03bbf0beb84-logging-loki-index-gateway-grpc\") pod \"logging-loki-index-gateway-0\" (UID: \"becf18d3-fc70-4b68-b16e-f03bbf0beb84\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 21 14:21:40 crc kubenswrapper[4897]: I1121 14:21:40.081456 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-index-gateway-http\" (UniqueName: \"kubernetes.io/secret/becf18d3-fc70-4b68-b16e-f03bbf0beb84-logging-loki-index-gateway-http\") pod \"logging-loki-index-gateway-0\" (UID: \"becf18d3-fc70-4b68-b16e-f03bbf0beb84\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 21 14:21:40 crc kubenswrapper[4897]: I1121 14:21:40.096215 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjhrl\" (UniqueName: \"kubernetes.io/projected/becf18d3-fc70-4b68-b16e-f03bbf0beb84-kube-api-access-hjhrl\") pod \"logging-loki-index-gateway-0\" (UID: \"becf18d3-fc70-4b68-b16e-f03bbf0beb84\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 21 14:21:40 crc kubenswrapper[4897]: I1121 14:21:40.111036 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-compactor-0" Nov 21 14:21:40 crc kubenswrapper[4897]: I1121 14:21:40.112007 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-c988cb84-c0c1-4fa9-be32-71560c8f12dc\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c988cb84-c0c1-4fa9-be32-71560c8f12dc\") pod \"logging-loki-index-gateway-0\" (UID: \"becf18d3-fc70-4b68-b16e-f03bbf0beb84\") " pod="openshift-logging/logging-loki-index-gateway-0" Nov 21 14:21:40 crc kubenswrapper[4897]: I1121 14:21:40.132291 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-index-gateway-0" Nov 21 14:21:40 crc kubenswrapper[4897]: I1121 14:21:40.222920 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-distributor-76cc67bf56-m64dx" event={"ID":"d3f767a5-605a-4b21-a15a-68cde314d5e6","Type":"ContainerStarted","Data":"30728ff8226f87524cc4b2617d6fec8b807de6dd84c961fc6f8380af23826854"} Nov 21 14:21:40 crc kubenswrapper[4897]: I1121 14:21:40.224811 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-vltt5" event={"ID":"a77cc078-3448-4f4b-86f4-90132ec0fd88","Type":"ContainerStarted","Data":"8f142860e9ddef0af2fc98c7308e6bd76ab694b474d8f47eb85a8c9f50e8b143"} Nov 21 14:21:40 crc kubenswrapper[4897]: I1121 14:21:40.227170 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-querier-5895d59bb8-8lqvs" event={"ID":"53c2fb7b-24cd-4da0-9016-e35e3223ab06","Type":"ContainerStarted","Data":"d0e891bbde2b1d678fca4c5dc5b7ccc6a3c4715333eb32997c6a3ae8d2c8685c"} Nov 21 14:21:40 crc kubenswrapper[4897]: I1121 14:21:40.281326 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-gateway-78b4f7fc55-dbkmd"] Nov 21 14:21:40 crc kubenswrapper[4897]: I1121 14:21:40.332895 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-gateway-78b4f7fc55-txlqp"] Nov 21 14:21:40 crc kubenswrapper[4897]: W1121 14:21:40.339330 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod268847f7_dd43_4094_a01c_e5d795722e82.slice/crio-479f8eed96e43c02b233f7e8f6d5b4ad32fe8e037f6c2ce159e3371cdd2cffda WatchSource:0}: Error finding container 479f8eed96e43c02b233f7e8f6d5b4ad32fe8e037f6c2ce159e3371cdd2cffda: Status 404 returned error can't find the container with id 479f8eed96e43c02b233f7e8f6d5b4ad32fe8e037f6c2ce159e3371cdd2cffda Nov 21 14:21:40 crc kubenswrapper[4897]: I1121 14:21:40.428936 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-ingester-0"] Nov 21 14:21:40 crc kubenswrapper[4897]: W1121 14:21:40.433213 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod69b9e549_607c_4707_9ff5_2dddbcc54b18.slice/crio-cc1a715b8753b8b0d4d484da82e0d99d9eca0666cfff201d89aa36e604e80022 WatchSource:0}: Error finding container cc1a715b8753b8b0d4d484da82e0d99d9eca0666cfff201d89aa36e604e80022: Status 404 returned error can't find the container with id cc1a715b8753b8b0d4d484da82e0d99d9eca0666cfff201d89aa36e604e80022 Nov 21 14:21:40 crc kubenswrapper[4897]: I1121 14:21:40.555363 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-compactor-0"] Nov 21 14:21:40 crc kubenswrapper[4897]: W1121 14:21:40.566853 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9241fda3_de39_4efb_8825_a5d706a4dcc9.slice/crio-184f859c07e66be6fc63a82ed1a5cce4efa61d0fdf55fe1fb7b0aceb122652a5 WatchSource:0}: Error finding container 184f859c07e66be6fc63a82ed1a5cce4efa61d0fdf55fe1fb7b0aceb122652a5: Status 404 returned error can't find the container with id 184f859c07e66be6fc63a82ed1a5cce4efa61d0fdf55fe1fb7b0aceb122652a5 Nov 21 14:21:40 crc kubenswrapper[4897]: I1121 14:21:40.681963 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-index-gateway-0"] Nov 21 14:21:41 crc kubenswrapper[4897]: I1121 14:21:41.235214 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-78b4f7fc55-txlqp" event={"ID":"268847f7-dd43-4094-a01c-e5d795722e82","Type":"ContainerStarted","Data":"479f8eed96e43c02b233f7e8f6d5b4ad32fe8e037f6c2ce159e3371cdd2cffda"} Nov 21 14:21:41 crc kubenswrapper[4897]: I1121 14:21:41.236288 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-index-gateway-0" event={"ID":"becf18d3-fc70-4b68-b16e-f03bbf0beb84","Type":"ContainerStarted","Data":"6c7343490af54878e3ccbcbb68abea1518439e5515b561f5222253e19c3a7820"} Nov 21 14:21:41 crc kubenswrapper[4897]: I1121 14:21:41.237649 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-ingester-0" event={"ID":"69b9e549-607c-4707-9ff5-2dddbcc54b18","Type":"ContainerStarted","Data":"cc1a715b8753b8b0d4d484da82e0d99d9eca0666cfff201d89aa36e604e80022"} Nov 21 14:21:41 crc kubenswrapper[4897]: I1121 14:21:41.238707 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-compactor-0" event={"ID":"9241fda3-de39-4efb-8825-a5d706a4dcc9","Type":"ContainerStarted","Data":"184f859c07e66be6fc63a82ed1a5cce4efa61d0fdf55fe1fb7b0aceb122652a5"} Nov 21 14:21:41 crc kubenswrapper[4897]: I1121 14:21:41.239771 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-78b4f7fc55-dbkmd" event={"ID":"6dda2ab1-6304-4116-90f3-b9219aff2664","Type":"ContainerStarted","Data":"e8f13500555b5e86e7504ae5d552c90e0e3dd73036d26a98c2623c5c4a594ed9"} Nov 21 14:21:43 crc kubenswrapper[4897]: I1121 14:21:43.284998 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-querier-5895d59bb8-8lqvs" event={"ID":"53c2fb7b-24cd-4da0-9016-e35e3223ab06","Type":"ContainerStarted","Data":"5612de5603c91a11f5a598ec95abae6b99fcd08184477782487b8943042e597a"} Nov 21 14:21:43 crc kubenswrapper[4897]: I1121 14:21:43.285531 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-querier-5895d59bb8-8lqvs" Nov 21 14:21:43 crc kubenswrapper[4897]: I1121 14:21:43.303644 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-querier-5895d59bb8-8lqvs" podStartSLOduration=1.802902191 podStartE2EDuration="5.30362658s" podCreationTimestamp="2025-11-21 14:21:38 +0000 UTC" firstStartedPulling="2025-11-21 14:21:39.512014957 +0000 UTC m=+776.796608432" lastFinishedPulling="2025-11-21 14:21:43.012739346 +0000 UTC m=+780.297332821" observedRunningTime="2025-11-21 14:21:43.302643264 +0000 UTC m=+780.587236739" watchObservedRunningTime="2025-11-21 14:21:43.30362658 +0000 UTC m=+780.588220055" Nov 21 14:21:44 crc kubenswrapper[4897]: I1121 14:21:44.294427 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-78b4f7fc55-dbkmd" event={"ID":"6dda2ab1-6304-4116-90f3-b9219aff2664","Type":"ContainerStarted","Data":"90f3f3400338672323b3c27b6d0a22cf5392b12cbf7a49bbe50bc05ca5e23ed4"} Nov 21 14:21:44 crc kubenswrapper[4897]: I1121 14:21:44.296328 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-distributor-76cc67bf56-m64dx" event={"ID":"d3f767a5-605a-4b21-a15a-68cde314d5e6","Type":"ContainerStarted","Data":"348ff68fc5dcdb278267662ff3d2dc6af578835159c153576d40353539ea8504"} Nov 21 14:21:44 crc kubenswrapper[4897]: I1121 14:21:44.296545 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-distributor-76cc67bf56-m64dx" Nov 21 14:21:44 crc kubenswrapper[4897]: I1121 14:21:44.297936 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-78b4f7fc55-txlqp" event={"ID":"268847f7-dd43-4094-a01c-e5d795722e82","Type":"ContainerStarted","Data":"4892c85303b09afa892229afee62cf278c65f828492d5f40c16641eadb5d7271"} Nov 21 14:21:44 crc kubenswrapper[4897]: I1121 14:21:44.299737 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-vltt5" event={"ID":"a77cc078-3448-4f4b-86f4-90132ec0fd88","Type":"ContainerStarted","Data":"82a8540d4e27df6215b3cd0e5eaae72d1709fd78438cf9927fa61ca8564f28ac"} Nov 21 14:21:44 crc kubenswrapper[4897]: I1121 14:21:44.299928 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-vltt5" Nov 21 14:21:44 crc kubenswrapper[4897]: I1121 14:21:44.301297 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-index-gateway-0" event={"ID":"becf18d3-fc70-4b68-b16e-f03bbf0beb84","Type":"ContainerStarted","Data":"8c65475eceeb67e5a9c6a49c1aa8e85a8328763afc0767c97a17c46581eb8c9c"} Nov 21 14:21:44 crc kubenswrapper[4897]: I1121 14:21:44.302086 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-index-gateway-0" Nov 21 14:21:44 crc kubenswrapper[4897]: I1121 14:21:44.306841 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-ingester-0" event={"ID":"69b9e549-607c-4707-9ff5-2dddbcc54b18","Type":"ContainerStarted","Data":"cb9ce0e6d6927d317ddc458987f7a613f8317c80932e919563bc816d8f9e03ef"} Nov 21 14:21:44 crc kubenswrapper[4897]: I1121 14:21:44.308354 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-ingester-0" Nov 21 14:21:44 crc kubenswrapper[4897]: I1121 14:21:44.311439 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-compactor-0" event={"ID":"9241fda3-de39-4efb-8825-a5d706a4dcc9","Type":"ContainerStarted","Data":"46d0ef6daa41b6772ee2f7815f0ed150f181976e62346a06e84a7fef5aff608e"} Nov 21 14:21:44 crc kubenswrapper[4897]: I1121 14:21:44.312139 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-compactor-0" Nov 21 14:21:44 crc kubenswrapper[4897]: I1121 14:21:44.333002 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-distributor-76cc67bf56-m64dx" podStartSLOduration=2.675362228 podStartE2EDuration="6.332971116s" podCreationTimestamp="2025-11-21 14:21:38 +0000 UTC" firstStartedPulling="2025-11-21 14:21:39.338954626 +0000 UTC m=+776.623548101" lastFinishedPulling="2025-11-21 14:21:42.996563484 +0000 UTC m=+780.281156989" observedRunningTime="2025-11-21 14:21:44.322287481 +0000 UTC m=+781.606880996" watchObservedRunningTime="2025-11-21 14:21:44.332971116 +0000 UTC m=+781.617564601" Nov 21 14:21:44 crc kubenswrapper[4897]: I1121 14:21:44.361302 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-ingester-0" podStartSLOduration=3.76395783 podStartE2EDuration="6.361278504s" podCreationTimestamp="2025-11-21 14:21:38 +0000 UTC" firstStartedPulling="2025-11-21 14:21:40.435816948 +0000 UTC m=+777.720410423" lastFinishedPulling="2025-11-21 14:21:43.033137612 +0000 UTC m=+780.317731097" observedRunningTime="2025-11-21 14:21:44.350312461 +0000 UTC m=+781.634905956" watchObservedRunningTime="2025-11-21 14:21:44.361278504 +0000 UTC m=+781.645871979" Nov 21 14:21:44 crc kubenswrapper[4897]: I1121 14:21:44.371167 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-compactor-0" podStartSLOduration=3.999852832 podStartE2EDuration="6.371144868s" podCreationTimestamp="2025-11-21 14:21:38 +0000 UTC" firstStartedPulling="2025-11-21 14:21:40.570095701 +0000 UTC m=+777.854689176" lastFinishedPulling="2025-11-21 14:21:42.941387697 +0000 UTC m=+780.225981212" observedRunningTime="2025-11-21 14:21:44.368103886 +0000 UTC m=+781.652697381" watchObservedRunningTime="2025-11-21 14:21:44.371144868 +0000 UTC m=+781.655738343" Nov 21 14:21:44 crc kubenswrapper[4897]: I1121 14:21:44.398639 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-index-gateway-0" podStartSLOduration=4.059110097 podStartE2EDuration="6.398617293s" podCreationTimestamp="2025-11-21 14:21:38 +0000 UTC" firstStartedPulling="2025-11-21 14:21:40.688916261 +0000 UTC m=+777.973509756" lastFinishedPulling="2025-11-21 14:21:43.028423437 +0000 UTC m=+780.313016952" observedRunningTime="2025-11-21 14:21:44.392677114 +0000 UTC m=+781.677270609" watchObservedRunningTime="2025-11-21 14:21:44.398617293 +0000 UTC m=+781.683210778" Nov 21 14:21:44 crc kubenswrapper[4897]: I1121 14:21:44.415443 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-vltt5" podStartSLOduration=3.003240003 podStartE2EDuration="6.415422483s" podCreationTimestamp="2025-11-21 14:21:38 +0000 UTC" firstStartedPulling="2025-11-21 14:21:39.618754553 +0000 UTC m=+776.903348028" lastFinishedPulling="2025-11-21 14:21:43.030937033 +0000 UTC m=+780.315530508" observedRunningTime="2025-11-21 14:21:44.414005645 +0000 UTC m=+781.698599140" watchObservedRunningTime="2025-11-21 14:21:44.415422483 +0000 UTC m=+781.700015968" Nov 21 14:21:46 crc kubenswrapper[4897]: I1121 14:21:46.328288 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-78b4f7fc55-dbkmd" event={"ID":"6dda2ab1-6304-4116-90f3-b9219aff2664","Type":"ContainerStarted","Data":"a8acaa05c825d9e056077577f40853f830830349a5f6f7f1565efe5465d77585"} Nov 21 14:21:46 crc kubenswrapper[4897]: I1121 14:21:46.329059 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-gateway-78b4f7fc55-dbkmd" Nov 21 14:21:46 crc kubenswrapper[4897]: I1121 14:21:46.329097 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-gateway-78b4f7fc55-dbkmd" Nov 21 14:21:46 crc kubenswrapper[4897]: I1121 14:21:46.331005 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-78b4f7fc55-txlqp" event={"ID":"268847f7-dd43-4094-a01c-e5d795722e82","Type":"ContainerStarted","Data":"ceb1ee306c0055e059eec20902b88b3414e6b569bc8c7186484c1f98586ab4a1"} Nov 21 14:21:46 crc kubenswrapper[4897]: I1121 14:21:46.331355 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-gateway-78b4f7fc55-txlqp" Nov 21 14:21:46 crc kubenswrapper[4897]: I1121 14:21:46.331383 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-gateway-78b4f7fc55-txlqp" Nov 21 14:21:46 crc kubenswrapper[4897]: I1121 14:21:46.341242 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-gateway-78b4f7fc55-txlqp" Nov 21 14:21:46 crc kubenswrapper[4897]: I1121 14:21:46.347452 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-gateway-78b4f7fc55-txlqp" Nov 21 14:21:46 crc kubenswrapper[4897]: I1121 14:21:46.348841 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-gateway-78b4f7fc55-dbkmd" Nov 21 14:21:46 crc kubenswrapper[4897]: I1121 14:21:46.358304 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-gateway-78b4f7fc55-dbkmd" Nov 21 14:21:46 crc kubenswrapper[4897]: I1121 14:21:46.359093 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-gateway-78b4f7fc55-dbkmd" podStartSLOduration=3.396930107 podStartE2EDuration="8.359063804s" podCreationTimestamp="2025-11-21 14:21:38 +0000 UTC" firstStartedPulling="2025-11-21 14:21:40.287072888 +0000 UTC m=+777.571666373" lastFinishedPulling="2025-11-21 14:21:45.249206585 +0000 UTC m=+782.533800070" observedRunningTime="2025-11-21 14:21:46.34920507 +0000 UTC m=+783.633798555" watchObservedRunningTime="2025-11-21 14:21:46.359063804 +0000 UTC m=+783.643657319" Nov 21 14:21:46 crc kubenswrapper[4897]: I1121 14:21:46.369170 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-gateway-78b4f7fc55-txlqp" podStartSLOduration=3.453206634 podStartE2EDuration="8.369149985s" podCreationTimestamp="2025-11-21 14:21:38 +0000 UTC" firstStartedPulling="2025-11-21 14:21:40.341639288 +0000 UTC m=+777.626232763" lastFinishedPulling="2025-11-21 14:21:45.257582639 +0000 UTC m=+782.542176114" observedRunningTime="2025-11-21 14:21:46.367396917 +0000 UTC m=+783.651990392" watchObservedRunningTime="2025-11-21 14:21:46.369149985 +0000 UTC m=+783.653743460" Nov 21 14:21:58 crc kubenswrapper[4897]: I1121 14:21:58.852636 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-distributor-76cc67bf56-m64dx" Nov 21 14:21:58 crc kubenswrapper[4897]: I1121 14:21:58.992840 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-querier-5895d59bb8-8lqvs" Nov 21 14:21:59 crc kubenswrapper[4897]: I1121 14:21:59.120994 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-vltt5" Nov 21 14:21:59 crc kubenswrapper[4897]: I1121 14:21:59.969302 4897 patch_prober.go:28] interesting pod/logging-loki-ingester-0 container/loki-ingester namespace/openshift-logging: Readiness probe status=failure output="HTTP probe failed with statuscode: 503" start-of-body=Ingester not ready: this instance owns no tokens Nov 21 14:21:59 crc kubenswrapper[4897]: I1121 14:21:59.969823 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-ingester-0" podUID="69b9e549-607c-4707-9ff5-2dddbcc54b18" containerName="loki-ingester" probeResult="failure" output="HTTP probe failed with statuscode: 503" Nov 21 14:22:00 crc kubenswrapper[4897]: I1121 14:22:00.119416 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-compactor-0" Nov 21 14:22:00 crc kubenswrapper[4897]: I1121 14:22:00.142911 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-index-gateway-0" Nov 21 14:22:01 crc kubenswrapper[4897]: I1121 14:22:01.732572 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-sdh5k"] Nov 21 14:22:01 crc kubenswrapper[4897]: I1121 14:22:01.734311 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sdh5k" Nov 21 14:22:01 crc kubenswrapper[4897]: I1121 14:22:01.759682 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8wxhd\" (UniqueName: \"kubernetes.io/projected/6ed64fe6-98d9-4f1c-8da9-e521c2604b16-kube-api-access-8wxhd\") pod \"redhat-operators-sdh5k\" (UID: \"6ed64fe6-98d9-4f1c-8da9-e521c2604b16\") " pod="openshift-marketplace/redhat-operators-sdh5k" Nov 21 14:22:01 crc kubenswrapper[4897]: I1121 14:22:01.759953 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ed64fe6-98d9-4f1c-8da9-e521c2604b16-utilities\") pod \"redhat-operators-sdh5k\" (UID: \"6ed64fe6-98d9-4f1c-8da9-e521c2604b16\") " pod="openshift-marketplace/redhat-operators-sdh5k" Nov 21 14:22:01 crc kubenswrapper[4897]: I1121 14:22:01.760095 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ed64fe6-98d9-4f1c-8da9-e521c2604b16-catalog-content\") pod \"redhat-operators-sdh5k\" (UID: \"6ed64fe6-98d9-4f1c-8da9-e521c2604b16\") " pod="openshift-marketplace/redhat-operators-sdh5k" Nov 21 14:22:01 crc kubenswrapper[4897]: I1121 14:22:01.796894 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-sdh5k"] Nov 21 14:22:01 crc kubenswrapper[4897]: I1121 14:22:01.862080 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8wxhd\" (UniqueName: \"kubernetes.io/projected/6ed64fe6-98d9-4f1c-8da9-e521c2604b16-kube-api-access-8wxhd\") pod \"redhat-operators-sdh5k\" (UID: \"6ed64fe6-98d9-4f1c-8da9-e521c2604b16\") " pod="openshift-marketplace/redhat-operators-sdh5k" Nov 21 14:22:01 crc kubenswrapper[4897]: I1121 14:22:01.862138 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ed64fe6-98d9-4f1c-8da9-e521c2604b16-utilities\") pod \"redhat-operators-sdh5k\" (UID: \"6ed64fe6-98d9-4f1c-8da9-e521c2604b16\") " pod="openshift-marketplace/redhat-operators-sdh5k" Nov 21 14:22:01 crc kubenswrapper[4897]: I1121 14:22:01.862171 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ed64fe6-98d9-4f1c-8da9-e521c2604b16-catalog-content\") pod \"redhat-operators-sdh5k\" (UID: \"6ed64fe6-98d9-4f1c-8da9-e521c2604b16\") " pod="openshift-marketplace/redhat-operators-sdh5k" Nov 21 14:22:01 crc kubenswrapper[4897]: I1121 14:22:01.862706 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ed64fe6-98d9-4f1c-8da9-e521c2604b16-catalog-content\") pod \"redhat-operators-sdh5k\" (UID: \"6ed64fe6-98d9-4f1c-8da9-e521c2604b16\") " pod="openshift-marketplace/redhat-operators-sdh5k" Nov 21 14:22:01 crc kubenswrapper[4897]: I1121 14:22:01.863023 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ed64fe6-98d9-4f1c-8da9-e521c2604b16-utilities\") pod \"redhat-operators-sdh5k\" (UID: \"6ed64fe6-98d9-4f1c-8da9-e521c2604b16\") " pod="openshift-marketplace/redhat-operators-sdh5k" Nov 21 14:22:01 crc kubenswrapper[4897]: I1121 14:22:01.880375 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8wxhd\" (UniqueName: \"kubernetes.io/projected/6ed64fe6-98d9-4f1c-8da9-e521c2604b16-kube-api-access-8wxhd\") pod \"redhat-operators-sdh5k\" (UID: \"6ed64fe6-98d9-4f1c-8da9-e521c2604b16\") " pod="openshift-marketplace/redhat-operators-sdh5k" Nov 21 14:22:02 crc kubenswrapper[4897]: I1121 14:22:02.050929 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sdh5k" Nov 21 14:22:02 crc kubenswrapper[4897]: W1121 14:22:02.504388 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6ed64fe6_98d9_4f1c_8da9_e521c2604b16.slice/crio-eb89c3c99ac6e18b815fc2491547d983463cf8df268ba72c64601718b6dc8b2f WatchSource:0}: Error finding container eb89c3c99ac6e18b815fc2491547d983463cf8df268ba72c64601718b6dc8b2f: Status 404 returned error can't find the container with id eb89c3c99ac6e18b815fc2491547d983463cf8df268ba72c64601718b6dc8b2f Nov 21 14:22:02 crc kubenswrapper[4897]: I1121 14:22:02.508328 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-sdh5k"] Nov 21 14:22:03 crc kubenswrapper[4897]: I1121 14:22:03.456480 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sdh5k" event={"ID":"6ed64fe6-98d9-4f1c-8da9-e521c2604b16","Type":"ContainerStarted","Data":"eb89c3c99ac6e18b815fc2491547d983463cf8df268ba72c64601718b6dc8b2f"} Nov 21 14:22:04 crc kubenswrapper[4897]: I1121 14:22:04.371294 4897 patch_prober.go:28] interesting pod/machine-config-daemon-krv5b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 21 14:22:04 crc kubenswrapper[4897]: I1121 14:22:04.371420 4897 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 21 14:22:04 crc kubenswrapper[4897]: I1121 14:22:04.371493 4897 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" Nov 21 14:22:04 crc kubenswrapper[4897]: I1121 14:22:04.372711 4897 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"cad6f00fa42dcd6129303195cd6590d896370bbb75a53d50c3950cca8fb66efa"} pod="openshift-machine-config-operator/machine-config-daemon-krv5b" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 21 14:22:04 crc kubenswrapper[4897]: I1121 14:22:04.372836 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" containerName="machine-config-daemon" containerID="cri-o://cad6f00fa42dcd6129303195cd6590d896370bbb75a53d50c3950cca8fb66efa" gracePeriod=600 Nov 21 14:22:08 crc kubenswrapper[4897]: I1121 14:22:08.007344 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-t9fnl"] Nov 21 14:22:08 crc kubenswrapper[4897]: I1121 14:22:08.009118 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-t9fnl" Nov 21 14:22:08 crc kubenswrapper[4897]: I1121 14:22:08.022459 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-t9fnl"] Nov 21 14:22:08 crc kubenswrapper[4897]: I1121 14:22:08.167214 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9920eebe-d0e6-4faa-9f8c-2dc8f1392770-catalog-content\") pod \"certified-operators-t9fnl\" (UID: \"9920eebe-d0e6-4faa-9f8c-2dc8f1392770\") " pod="openshift-marketplace/certified-operators-t9fnl" Nov 21 14:22:08 crc kubenswrapper[4897]: I1121 14:22:08.167333 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pp4wl\" (UniqueName: \"kubernetes.io/projected/9920eebe-d0e6-4faa-9f8c-2dc8f1392770-kube-api-access-pp4wl\") pod \"certified-operators-t9fnl\" (UID: \"9920eebe-d0e6-4faa-9f8c-2dc8f1392770\") " pod="openshift-marketplace/certified-operators-t9fnl" Nov 21 14:22:08 crc kubenswrapper[4897]: I1121 14:22:08.167696 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9920eebe-d0e6-4faa-9f8c-2dc8f1392770-utilities\") pod \"certified-operators-t9fnl\" (UID: \"9920eebe-d0e6-4faa-9f8c-2dc8f1392770\") " pod="openshift-marketplace/certified-operators-t9fnl" Nov 21 14:22:08 crc kubenswrapper[4897]: I1121 14:22:08.269795 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9920eebe-d0e6-4faa-9f8c-2dc8f1392770-catalog-content\") pod \"certified-operators-t9fnl\" (UID: \"9920eebe-d0e6-4faa-9f8c-2dc8f1392770\") " pod="openshift-marketplace/certified-operators-t9fnl" Nov 21 14:22:08 crc kubenswrapper[4897]: I1121 14:22:08.269928 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pp4wl\" (UniqueName: \"kubernetes.io/projected/9920eebe-d0e6-4faa-9f8c-2dc8f1392770-kube-api-access-pp4wl\") pod \"certified-operators-t9fnl\" (UID: \"9920eebe-d0e6-4faa-9f8c-2dc8f1392770\") " pod="openshift-marketplace/certified-operators-t9fnl" Nov 21 14:22:08 crc kubenswrapper[4897]: I1121 14:22:08.269972 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9920eebe-d0e6-4faa-9f8c-2dc8f1392770-utilities\") pod \"certified-operators-t9fnl\" (UID: \"9920eebe-d0e6-4faa-9f8c-2dc8f1392770\") " pod="openshift-marketplace/certified-operators-t9fnl" Nov 21 14:22:08 crc kubenswrapper[4897]: I1121 14:22:08.270369 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9920eebe-d0e6-4faa-9f8c-2dc8f1392770-utilities\") pod \"certified-operators-t9fnl\" (UID: \"9920eebe-d0e6-4faa-9f8c-2dc8f1392770\") " pod="openshift-marketplace/certified-operators-t9fnl" Nov 21 14:22:08 crc kubenswrapper[4897]: I1121 14:22:08.270397 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9920eebe-d0e6-4faa-9f8c-2dc8f1392770-catalog-content\") pod \"certified-operators-t9fnl\" (UID: \"9920eebe-d0e6-4faa-9f8c-2dc8f1392770\") " pod="openshift-marketplace/certified-operators-t9fnl" Nov 21 14:22:08 crc kubenswrapper[4897]: I1121 14:22:08.289871 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pp4wl\" (UniqueName: \"kubernetes.io/projected/9920eebe-d0e6-4faa-9f8c-2dc8f1392770-kube-api-access-pp4wl\") pod \"certified-operators-t9fnl\" (UID: \"9920eebe-d0e6-4faa-9f8c-2dc8f1392770\") " pod="openshift-marketplace/certified-operators-t9fnl" Nov 21 14:22:08 crc kubenswrapper[4897]: I1121 14:22:08.332023 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-t9fnl" Nov 21 14:22:08 crc kubenswrapper[4897]: I1121 14:22:08.819339 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-t9fnl"] Nov 21 14:22:08 crc kubenswrapper[4897]: W1121 14:22:08.836538 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9920eebe_d0e6_4faa_9f8c_2dc8f1392770.slice/crio-7a4c77ff9a577f093429c2a1b32ccdf6e00e2b77fafba32ed27ccb945d5fbe86 WatchSource:0}: Error finding container 7a4c77ff9a577f093429c2a1b32ccdf6e00e2b77fafba32ed27ccb945d5fbe86: Status 404 returned error can't find the container with id 7a4c77ff9a577f093429c2a1b32ccdf6e00e2b77fafba32ed27ccb945d5fbe86 Nov 21 14:22:09 crc kubenswrapper[4897]: I1121 14:22:09.969592 4897 patch_prober.go:28] interesting pod/logging-loki-ingester-0 container/loki-ingester namespace/openshift-logging: Readiness probe status=failure output="HTTP probe failed with statuscode: 503" start-of-body=Ingester not ready: this instance owns no tokens Nov 21 14:22:09 crc kubenswrapper[4897]: I1121 14:22:09.970094 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-ingester-0" podUID="69b9e549-607c-4707-9ff5-2dddbcc54b18" containerName="loki-ingester" probeResult="failure" output="HTTP probe failed with statuscode: 503" Nov 21 14:22:12 crc kubenswrapper[4897]: I1121 14:22:12.308248 4897 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-config-operator_machine-config-daemon-krv5b_e7670227-d280-4847-b882-754429f56b0f/machine-config-daemon/3.log" Nov 21 14:22:12 crc kubenswrapper[4897]: I1121 14:22:12.309774 4897 generic.go:334] "Generic (PLEG): container finished" podID="e7670227-d280-4847-b882-754429f56b0f" containerID="cad6f00fa42dcd6129303195cd6590d896370bbb75a53d50c3950cca8fb66efa" exitCode=-1 Nov 21 14:22:12 crc kubenswrapper[4897]: I1121 14:22:12.309828 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" event={"ID":"e7670227-d280-4847-b882-754429f56b0f","Type":"ContainerDied","Data":"cad6f00fa42dcd6129303195cd6590d896370bbb75a53d50c3950cca8fb66efa"} Nov 21 14:22:12 crc kubenswrapper[4897]: I1121 14:22:12.309862 4897 scope.go:117] "RemoveContainer" containerID="3ef2942b0389d06ca20fad79231df545f174b1eeae70986340588271efcd18e6" Nov 21 14:22:13 crc kubenswrapper[4897]: I1121 14:22:13.320339 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t9fnl" event={"ID":"9920eebe-d0e6-4faa-9f8c-2dc8f1392770","Type":"ContainerStarted","Data":"7a4c77ff9a577f093429c2a1b32ccdf6e00e2b77fafba32ed27ccb945d5fbe86"} Nov 21 14:22:14 crc kubenswrapper[4897]: I1121 14:22:14.330901 4897 generic.go:334] "Generic (PLEG): container finished" podID="6ed64fe6-98d9-4f1c-8da9-e521c2604b16" containerID="f346513a16dbce7ddbb97138b404f1ea60ca3fd82a6bc46c40fcea1f5663708a" exitCode=0 Nov 21 14:22:14 crc kubenswrapper[4897]: I1121 14:22:14.330951 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sdh5k" event={"ID":"6ed64fe6-98d9-4f1c-8da9-e521c2604b16","Type":"ContainerDied","Data":"f346513a16dbce7ddbb97138b404f1ea60ca3fd82a6bc46c40fcea1f5663708a"} Nov 21 14:22:14 crc kubenswrapper[4897]: I1121 14:22:14.333717 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t9fnl" event={"ID":"9920eebe-d0e6-4faa-9f8c-2dc8f1392770","Type":"ContainerStarted","Data":"a224fe6da8a6f7a0bfa666188c60d2aaf4a04cc8c25bb508a09ccdb68c32594d"} Nov 21 14:22:15 crc kubenswrapper[4897]: I1121 14:22:15.341489 4897 generic.go:334] "Generic (PLEG): container finished" podID="9920eebe-d0e6-4faa-9f8c-2dc8f1392770" containerID="a224fe6da8a6f7a0bfa666188c60d2aaf4a04cc8c25bb508a09ccdb68c32594d" exitCode=0 Nov 21 14:22:15 crc kubenswrapper[4897]: I1121 14:22:15.341568 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t9fnl" event={"ID":"9920eebe-d0e6-4faa-9f8c-2dc8f1392770","Type":"ContainerDied","Data":"a224fe6da8a6f7a0bfa666188c60d2aaf4a04cc8c25bb508a09ccdb68c32594d"} Nov 21 14:22:17 crc kubenswrapper[4897]: I1121 14:22:17.356801 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" event={"ID":"e7670227-d280-4847-b882-754429f56b0f","Type":"ContainerStarted","Data":"6f8d8438a69d75aad7ccb460842fd43be11b0196e6779f692d047f7ed8f74e83"} Nov 21 14:22:19 crc kubenswrapper[4897]: I1121 14:22:19.966721 4897 patch_prober.go:28] interesting pod/logging-loki-ingester-0 container/loki-ingester namespace/openshift-logging: Readiness probe status=failure output="HTTP probe failed with statuscode: 503" start-of-body=Ingester not ready: waiting for 15s after being ready Nov 21 14:22:19 crc kubenswrapper[4897]: I1121 14:22:19.967461 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-ingester-0" podUID="69b9e549-607c-4707-9ff5-2dddbcc54b18" containerName="loki-ingester" probeResult="failure" output="HTTP probe failed with statuscode: 503" Nov 21 14:22:29 crc kubenswrapper[4897]: I1121 14:22:29.969639 4897 patch_prober.go:28] interesting pod/logging-loki-ingester-0 container/loki-ingester namespace/openshift-logging: Readiness probe status=failure output="HTTP probe failed with statuscode: 503" start-of-body=Ingester not ready: waiting for 15s after being ready Nov 21 14:22:29 crc kubenswrapper[4897]: I1121 14:22:29.970685 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-ingester-0" podUID="69b9e549-607c-4707-9ff5-2dddbcc54b18" containerName="loki-ingester" probeResult="failure" output="HTTP probe failed with statuscode: 503" Nov 21 14:22:31 crc kubenswrapper[4897]: I1121 14:22:31.482447 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sdh5k" event={"ID":"6ed64fe6-98d9-4f1c-8da9-e521c2604b16","Type":"ContainerStarted","Data":"1582540efa85f2707892dcdbb9b10192ff5d7e87820b41e33297218eed037f3e"} Nov 21 14:22:32 crc kubenswrapper[4897]: I1121 14:22:32.497034 4897 generic.go:334] "Generic (PLEG): container finished" podID="6ed64fe6-98d9-4f1c-8da9-e521c2604b16" containerID="1582540efa85f2707892dcdbb9b10192ff5d7e87820b41e33297218eed037f3e" exitCode=0 Nov 21 14:22:32 crc kubenswrapper[4897]: I1121 14:22:32.497107 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sdh5k" event={"ID":"6ed64fe6-98d9-4f1c-8da9-e521c2604b16","Type":"ContainerDied","Data":"1582540efa85f2707892dcdbb9b10192ff5d7e87820b41e33297218eed037f3e"} Nov 21 14:22:32 crc kubenswrapper[4897]: I1121 14:22:32.504323 4897 generic.go:334] "Generic (PLEG): container finished" podID="9920eebe-d0e6-4faa-9f8c-2dc8f1392770" containerID="ce5fd0711c68e5c357a10c05e310e1f405ca32797f4158e68cd1f8f0483b9a28" exitCode=0 Nov 21 14:22:32 crc kubenswrapper[4897]: I1121 14:22:32.504362 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t9fnl" event={"ID":"9920eebe-d0e6-4faa-9f8c-2dc8f1392770","Type":"ContainerDied","Data":"ce5fd0711c68e5c357a10c05e310e1f405ca32797f4158e68cd1f8f0483b9a28"} Nov 21 14:22:34 crc kubenswrapper[4897]: I1121 14:22:34.521348 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sdh5k" event={"ID":"6ed64fe6-98d9-4f1c-8da9-e521c2604b16","Type":"ContainerStarted","Data":"258f84e0f83863a4b94aaa1e31acb431c9347c825add06ebd5a5685d78bed9b0"} Nov 21 14:22:34 crc kubenswrapper[4897]: I1121 14:22:34.526055 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t9fnl" event={"ID":"9920eebe-d0e6-4faa-9f8c-2dc8f1392770","Type":"ContainerStarted","Data":"a38c7b536e57d4fadeeed30f9f4678c14828ecd9799eb674f71c3c076712304b"} Nov 21 14:22:34 crc kubenswrapper[4897]: I1121 14:22:34.555431 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-sdh5k" podStartSLOduration=17.423953693 podStartE2EDuration="33.555413924s" podCreationTimestamp="2025-11-21 14:22:01 +0000 UTC" firstStartedPulling="2025-11-21 14:22:17.360479054 +0000 UTC m=+814.645072529" lastFinishedPulling="2025-11-21 14:22:33.491939275 +0000 UTC m=+830.776532760" observedRunningTime="2025-11-21 14:22:34.552232009 +0000 UTC m=+831.836825504" watchObservedRunningTime="2025-11-21 14:22:34.555413924 +0000 UTC m=+831.840007389" Nov 21 14:22:34 crc kubenswrapper[4897]: I1121 14:22:34.568612 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-t9fnl" podStartSLOduration=10.975678286 podStartE2EDuration="27.568597316s" podCreationTimestamp="2025-11-21 14:22:07 +0000 UTC" firstStartedPulling="2025-11-21 14:22:17.360558646 +0000 UTC m=+814.645152141" lastFinishedPulling="2025-11-21 14:22:33.953477696 +0000 UTC m=+831.238071171" observedRunningTime="2025-11-21 14:22:34.567744744 +0000 UTC m=+831.852338219" watchObservedRunningTime="2025-11-21 14:22:34.568597316 +0000 UTC m=+831.853190791" Nov 21 14:22:36 crc kubenswrapper[4897]: I1121 14:22:36.141960 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-7bh27"] Nov 21 14:22:36 crc kubenswrapper[4897]: I1121 14:22:36.145395 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7bh27" Nov 21 14:22:36 crc kubenswrapper[4897]: I1121 14:22:36.152269 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7bh27"] Nov 21 14:22:36 crc kubenswrapper[4897]: I1121 14:22:36.291336 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40f12519-bf8d-4792-a4ba-f4d64a3f3b93-catalog-content\") pod \"community-operators-7bh27\" (UID: \"40f12519-bf8d-4792-a4ba-f4d64a3f3b93\") " pod="openshift-marketplace/community-operators-7bh27" Nov 21 14:22:36 crc kubenswrapper[4897]: I1121 14:22:36.291412 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40f12519-bf8d-4792-a4ba-f4d64a3f3b93-utilities\") pod \"community-operators-7bh27\" (UID: \"40f12519-bf8d-4792-a4ba-f4d64a3f3b93\") " pod="openshift-marketplace/community-operators-7bh27" Nov 21 14:22:36 crc kubenswrapper[4897]: I1121 14:22:36.291449 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mpt8v\" (UniqueName: \"kubernetes.io/projected/40f12519-bf8d-4792-a4ba-f4d64a3f3b93-kube-api-access-mpt8v\") pod \"community-operators-7bh27\" (UID: \"40f12519-bf8d-4792-a4ba-f4d64a3f3b93\") " pod="openshift-marketplace/community-operators-7bh27" Nov 21 14:22:36 crc kubenswrapper[4897]: I1121 14:22:36.393369 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40f12519-bf8d-4792-a4ba-f4d64a3f3b93-utilities\") pod \"community-operators-7bh27\" (UID: \"40f12519-bf8d-4792-a4ba-f4d64a3f3b93\") " pod="openshift-marketplace/community-operators-7bh27" Nov 21 14:22:36 crc kubenswrapper[4897]: I1121 14:22:36.393494 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mpt8v\" (UniqueName: \"kubernetes.io/projected/40f12519-bf8d-4792-a4ba-f4d64a3f3b93-kube-api-access-mpt8v\") pod \"community-operators-7bh27\" (UID: \"40f12519-bf8d-4792-a4ba-f4d64a3f3b93\") " pod="openshift-marketplace/community-operators-7bh27" Nov 21 14:22:36 crc kubenswrapper[4897]: I1121 14:22:36.393783 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40f12519-bf8d-4792-a4ba-f4d64a3f3b93-catalog-content\") pod \"community-operators-7bh27\" (UID: \"40f12519-bf8d-4792-a4ba-f4d64a3f3b93\") " pod="openshift-marketplace/community-operators-7bh27" Nov 21 14:22:36 crc kubenswrapper[4897]: I1121 14:22:36.393839 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40f12519-bf8d-4792-a4ba-f4d64a3f3b93-utilities\") pod \"community-operators-7bh27\" (UID: \"40f12519-bf8d-4792-a4ba-f4d64a3f3b93\") " pod="openshift-marketplace/community-operators-7bh27" Nov 21 14:22:36 crc kubenswrapper[4897]: I1121 14:22:36.394202 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40f12519-bf8d-4792-a4ba-f4d64a3f3b93-catalog-content\") pod \"community-operators-7bh27\" (UID: \"40f12519-bf8d-4792-a4ba-f4d64a3f3b93\") " pod="openshift-marketplace/community-operators-7bh27" Nov 21 14:22:36 crc kubenswrapper[4897]: I1121 14:22:36.418571 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mpt8v\" (UniqueName: \"kubernetes.io/projected/40f12519-bf8d-4792-a4ba-f4d64a3f3b93-kube-api-access-mpt8v\") pod \"community-operators-7bh27\" (UID: \"40f12519-bf8d-4792-a4ba-f4d64a3f3b93\") " pod="openshift-marketplace/community-operators-7bh27" Nov 21 14:22:36 crc kubenswrapper[4897]: I1121 14:22:36.517552 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7bh27" Nov 21 14:22:36 crc kubenswrapper[4897]: I1121 14:22:36.992671 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7bh27"] Nov 21 14:22:36 crc kubenswrapper[4897]: W1121 14:22:36.996574 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod40f12519_bf8d_4792_a4ba_f4d64a3f3b93.slice/crio-7f428a58f9a5f23a45d6ab30352c35d6d1d0805b8a3b429842ea2bdd664b3dc4 WatchSource:0}: Error finding container 7f428a58f9a5f23a45d6ab30352c35d6d1d0805b8a3b429842ea2bdd664b3dc4: Status 404 returned error can't find the container with id 7f428a58f9a5f23a45d6ab30352c35d6d1d0805b8a3b429842ea2bdd664b3dc4 Nov 21 14:22:37 crc kubenswrapper[4897]: I1121 14:22:37.552812 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7bh27" event={"ID":"40f12519-bf8d-4792-a4ba-f4d64a3f3b93","Type":"ContainerStarted","Data":"7f428a58f9a5f23a45d6ab30352c35d6d1d0805b8a3b429842ea2bdd664b3dc4"} Nov 21 14:22:38 crc kubenswrapper[4897]: I1121 14:22:38.333789 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-t9fnl" Nov 21 14:22:38 crc kubenswrapper[4897]: I1121 14:22:38.333856 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-t9fnl" Nov 21 14:22:38 crc kubenswrapper[4897]: I1121 14:22:38.373989 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-t9fnl" Nov 21 14:22:39 crc kubenswrapper[4897]: I1121 14:22:39.565745 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7bh27" event={"ID":"40f12519-bf8d-4792-a4ba-f4d64a3f3b93","Type":"ContainerStarted","Data":"0e61b61751340c191384d392e6ce56c6768f85afa55788b1093d21e9b64e9fb5"} Nov 21 14:22:39 crc kubenswrapper[4897]: I1121 14:22:39.963908 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-ingester-0" Nov 21 14:22:40 crc kubenswrapper[4897]: I1121 14:22:40.573701 4897 generic.go:334] "Generic (PLEG): container finished" podID="40f12519-bf8d-4792-a4ba-f4d64a3f3b93" containerID="0e61b61751340c191384d392e6ce56c6768f85afa55788b1093d21e9b64e9fb5" exitCode=0 Nov 21 14:22:40 crc kubenswrapper[4897]: I1121 14:22:40.573747 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7bh27" event={"ID":"40f12519-bf8d-4792-a4ba-f4d64a3f3b93","Type":"ContainerDied","Data":"0e61b61751340c191384d392e6ce56c6768f85afa55788b1093d21e9b64e9fb5"} Nov 21 14:22:42 crc kubenswrapper[4897]: I1121 14:22:42.051953 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-sdh5k" Nov 21 14:22:42 crc kubenswrapper[4897]: I1121 14:22:42.052083 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-sdh5k" Nov 21 14:22:42 crc kubenswrapper[4897]: I1121 14:22:42.106870 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-sdh5k" Nov 21 14:22:42 crc kubenswrapper[4897]: I1121 14:22:42.641107 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-sdh5k" Nov 21 14:22:44 crc kubenswrapper[4897]: I1121 14:22:44.133939 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-sdh5k"] Nov 21 14:22:45 crc kubenswrapper[4897]: I1121 14:22:45.607156 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-sdh5k" podUID="6ed64fe6-98d9-4f1c-8da9-e521c2604b16" containerName="registry-server" containerID="cri-o://258f84e0f83863a4b94aaa1e31acb431c9347c825add06ebd5a5685d78bed9b0" gracePeriod=2 Nov 21 14:22:47 crc kubenswrapper[4897]: I1121 14:22:47.631295 4897 generic.go:334] "Generic (PLEG): container finished" podID="6ed64fe6-98d9-4f1c-8da9-e521c2604b16" containerID="258f84e0f83863a4b94aaa1e31acb431c9347c825add06ebd5a5685d78bed9b0" exitCode=0 Nov 21 14:22:47 crc kubenswrapper[4897]: I1121 14:22:47.631386 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sdh5k" event={"ID":"6ed64fe6-98d9-4f1c-8da9-e521c2604b16","Type":"ContainerDied","Data":"258f84e0f83863a4b94aaa1e31acb431c9347c825add06ebd5a5685d78bed9b0"} Nov 21 14:22:48 crc kubenswrapper[4897]: I1121 14:22:48.391582 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-t9fnl" Nov 21 14:22:48 crc kubenswrapper[4897]: I1121 14:22:48.437985 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-t9fnl"] Nov 21 14:22:48 crc kubenswrapper[4897]: I1121 14:22:48.638008 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-t9fnl" podUID="9920eebe-d0e6-4faa-9f8c-2dc8f1392770" containerName="registry-server" containerID="cri-o://a38c7b536e57d4fadeeed30f9f4678c14828ecd9799eb674f71c3c076712304b" gracePeriod=2 Nov 21 14:22:48 crc kubenswrapper[4897]: I1121 14:22:48.882653 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sdh5k" Nov 21 14:22:49 crc kubenswrapper[4897]: I1121 14:22:49.006748 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8wxhd\" (UniqueName: \"kubernetes.io/projected/6ed64fe6-98d9-4f1c-8da9-e521c2604b16-kube-api-access-8wxhd\") pod \"6ed64fe6-98d9-4f1c-8da9-e521c2604b16\" (UID: \"6ed64fe6-98d9-4f1c-8da9-e521c2604b16\") " Nov 21 14:22:49 crc kubenswrapper[4897]: I1121 14:22:49.006869 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ed64fe6-98d9-4f1c-8da9-e521c2604b16-catalog-content\") pod \"6ed64fe6-98d9-4f1c-8da9-e521c2604b16\" (UID: \"6ed64fe6-98d9-4f1c-8da9-e521c2604b16\") " Nov 21 14:22:49 crc kubenswrapper[4897]: I1121 14:22:49.007017 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ed64fe6-98d9-4f1c-8da9-e521c2604b16-utilities\") pod \"6ed64fe6-98d9-4f1c-8da9-e521c2604b16\" (UID: \"6ed64fe6-98d9-4f1c-8da9-e521c2604b16\") " Nov 21 14:22:49 crc kubenswrapper[4897]: I1121 14:22:49.009141 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6ed64fe6-98d9-4f1c-8da9-e521c2604b16-utilities" (OuterVolumeSpecName: "utilities") pod "6ed64fe6-98d9-4f1c-8da9-e521c2604b16" (UID: "6ed64fe6-98d9-4f1c-8da9-e521c2604b16"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:22:49 crc kubenswrapper[4897]: I1121 14:22:49.015069 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ed64fe6-98d9-4f1c-8da9-e521c2604b16-kube-api-access-8wxhd" (OuterVolumeSpecName: "kube-api-access-8wxhd") pod "6ed64fe6-98d9-4f1c-8da9-e521c2604b16" (UID: "6ed64fe6-98d9-4f1c-8da9-e521c2604b16"). InnerVolumeSpecName "kube-api-access-8wxhd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:22:49 crc kubenswrapper[4897]: I1121 14:22:49.041042 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-t9fnl" Nov 21 14:22:49 crc kubenswrapper[4897]: I1121 14:22:49.097662 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6ed64fe6-98d9-4f1c-8da9-e521c2604b16-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6ed64fe6-98d9-4f1c-8da9-e521c2604b16" (UID: "6ed64fe6-98d9-4f1c-8da9-e521c2604b16"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:22:49 crc kubenswrapper[4897]: I1121 14:22:49.111030 4897 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ed64fe6-98d9-4f1c-8da9-e521c2604b16-utilities\") on node \"crc\" DevicePath \"\"" Nov 21 14:22:49 crc kubenswrapper[4897]: I1121 14:22:49.111101 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8wxhd\" (UniqueName: \"kubernetes.io/projected/6ed64fe6-98d9-4f1c-8da9-e521c2604b16-kube-api-access-8wxhd\") on node \"crc\" DevicePath \"\"" Nov 21 14:22:49 crc kubenswrapper[4897]: I1121 14:22:49.111147 4897 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ed64fe6-98d9-4f1c-8da9-e521c2604b16-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 21 14:22:49 crc kubenswrapper[4897]: I1121 14:22:49.212885 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9920eebe-d0e6-4faa-9f8c-2dc8f1392770-utilities\") pod \"9920eebe-d0e6-4faa-9f8c-2dc8f1392770\" (UID: \"9920eebe-d0e6-4faa-9f8c-2dc8f1392770\") " Nov 21 14:22:49 crc kubenswrapper[4897]: I1121 14:22:49.213019 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9920eebe-d0e6-4faa-9f8c-2dc8f1392770-catalog-content\") pod \"9920eebe-d0e6-4faa-9f8c-2dc8f1392770\" (UID: \"9920eebe-d0e6-4faa-9f8c-2dc8f1392770\") " Nov 21 14:22:49 crc kubenswrapper[4897]: I1121 14:22:49.213056 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pp4wl\" (UniqueName: \"kubernetes.io/projected/9920eebe-d0e6-4faa-9f8c-2dc8f1392770-kube-api-access-pp4wl\") pod \"9920eebe-d0e6-4faa-9f8c-2dc8f1392770\" (UID: \"9920eebe-d0e6-4faa-9f8c-2dc8f1392770\") " Nov 21 14:22:49 crc kubenswrapper[4897]: I1121 14:22:49.213759 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9920eebe-d0e6-4faa-9f8c-2dc8f1392770-utilities" (OuterVolumeSpecName: "utilities") pod "9920eebe-d0e6-4faa-9f8c-2dc8f1392770" (UID: "9920eebe-d0e6-4faa-9f8c-2dc8f1392770"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:22:49 crc kubenswrapper[4897]: I1121 14:22:49.219020 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9920eebe-d0e6-4faa-9f8c-2dc8f1392770-kube-api-access-pp4wl" (OuterVolumeSpecName: "kube-api-access-pp4wl") pod "9920eebe-d0e6-4faa-9f8c-2dc8f1392770" (UID: "9920eebe-d0e6-4faa-9f8c-2dc8f1392770"). InnerVolumeSpecName "kube-api-access-pp4wl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:22:49 crc kubenswrapper[4897]: I1121 14:22:49.275924 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9920eebe-d0e6-4faa-9f8c-2dc8f1392770-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9920eebe-d0e6-4faa-9f8c-2dc8f1392770" (UID: "9920eebe-d0e6-4faa-9f8c-2dc8f1392770"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:22:49 crc kubenswrapper[4897]: I1121 14:22:49.315171 4897 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9920eebe-d0e6-4faa-9f8c-2dc8f1392770-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 21 14:22:49 crc kubenswrapper[4897]: I1121 14:22:49.315221 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pp4wl\" (UniqueName: \"kubernetes.io/projected/9920eebe-d0e6-4faa-9f8c-2dc8f1392770-kube-api-access-pp4wl\") on node \"crc\" DevicePath \"\"" Nov 21 14:22:49 crc kubenswrapper[4897]: I1121 14:22:49.315238 4897 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9920eebe-d0e6-4faa-9f8c-2dc8f1392770-utilities\") on node \"crc\" DevicePath \"\"" Nov 21 14:22:49 crc kubenswrapper[4897]: I1121 14:22:49.646618 4897 generic.go:334] "Generic (PLEG): container finished" podID="9920eebe-d0e6-4faa-9f8c-2dc8f1392770" containerID="a38c7b536e57d4fadeeed30f9f4678c14828ecd9799eb674f71c3c076712304b" exitCode=0 Nov 21 14:22:49 crc kubenswrapper[4897]: I1121 14:22:49.646690 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-t9fnl" Nov 21 14:22:49 crc kubenswrapper[4897]: I1121 14:22:49.646733 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t9fnl" event={"ID":"9920eebe-d0e6-4faa-9f8c-2dc8f1392770","Type":"ContainerDied","Data":"a38c7b536e57d4fadeeed30f9f4678c14828ecd9799eb674f71c3c076712304b"} Nov 21 14:22:49 crc kubenswrapper[4897]: I1121 14:22:49.646874 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t9fnl" event={"ID":"9920eebe-d0e6-4faa-9f8c-2dc8f1392770","Type":"ContainerDied","Data":"7a4c77ff9a577f093429c2a1b32ccdf6e00e2b77fafba32ed27ccb945d5fbe86"} Nov 21 14:22:49 crc kubenswrapper[4897]: I1121 14:22:49.646943 4897 scope.go:117] "RemoveContainer" containerID="a38c7b536e57d4fadeeed30f9f4678c14828ecd9799eb674f71c3c076712304b" Nov 21 14:22:49 crc kubenswrapper[4897]: I1121 14:22:49.653945 4897 generic.go:334] "Generic (PLEG): container finished" podID="40f12519-bf8d-4792-a4ba-f4d64a3f3b93" containerID="0c0c9b2aca5dcd69d12d194b7dc2fdc181992481faf70ed283b72e5629fd2d71" exitCode=0 Nov 21 14:22:49 crc kubenswrapper[4897]: I1121 14:22:49.654042 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7bh27" event={"ID":"40f12519-bf8d-4792-a4ba-f4d64a3f3b93","Type":"ContainerDied","Data":"0c0c9b2aca5dcd69d12d194b7dc2fdc181992481faf70ed283b72e5629fd2d71"} Nov 21 14:22:49 crc kubenswrapper[4897]: I1121 14:22:49.660284 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sdh5k" event={"ID":"6ed64fe6-98d9-4f1c-8da9-e521c2604b16","Type":"ContainerDied","Data":"eb89c3c99ac6e18b815fc2491547d983463cf8df268ba72c64601718b6dc8b2f"} Nov 21 14:22:49 crc kubenswrapper[4897]: I1121 14:22:49.660461 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sdh5k" Nov 21 14:22:49 crc kubenswrapper[4897]: I1121 14:22:49.677135 4897 scope.go:117] "RemoveContainer" containerID="ce5fd0711c68e5c357a10c05e310e1f405ca32797f4158e68cd1f8f0483b9a28" Nov 21 14:22:49 crc kubenswrapper[4897]: I1121 14:22:49.699003 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-t9fnl"] Nov 21 14:22:49 crc kubenswrapper[4897]: I1121 14:22:49.713667 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-t9fnl"] Nov 21 14:22:49 crc kubenswrapper[4897]: I1121 14:22:49.720897 4897 scope.go:117] "RemoveContainer" containerID="a224fe6da8a6f7a0bfa666188c60d2aaf4a04cc8c25bb508a09ccdb68c32594d" Nov 21 14:22:49 crc kubenswrapper[4897]: I1121 14:22:49.722895 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-sdh5k"] Nov 21 14:22:49 crc kubenswrapper[4897]: I1121 14:22:49.731519 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-sdh5k"] Nov 21 14:22:49 crc kubenswrapper[4897]: I1121 14:22:49.736499 4897 scope.go:117] "RemoveContainer" containerID="a38c7b536e57d4fadeeed30f9f4678c14828ecd9799eb674f71c3c076712304b" Nov 21 14:22:49 crc kubenswrapper[4897]: E1121 14:22:49.736856 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a38c7b536e57d4fadeeed30f9f4678c14828ecd9799eb674f71c3c076712304b\": container with ID starting with a38c7b536e57d4fadeeed30f9f4678c14828ecd9799eb674f71c3c076712304b not found: ID does not exist" containerID="a38c7b536e57d4fadeeed30f9f4678c14828ecd9799eb674f71c3c076712304b" Nov 21 14:22:49 crc kubenswrapper[4897]: I1121 14:22:49.736899 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a38c7b536e57d4fadeeed30f9f4678c14828ecd9799eb674f71c3c076712304b"} err="failed to get container status \"a38c7b536e57d4fadeeed30f9f4678c14828ecd9799eb674f71c3c076712304b\": rpc error: code = NotFound desc = could not find container \"a38c7b536e57d4fadeeed30f9f4678c14828ecd9799eb674f71c3c076712304b\": container with ID starting with a38c7b536e57d4fadeeed30f9f4678c14828ecd9799eb674f71c3c076712304b not found: ID does not exist" Nov 21 14:22:49 crc kubenswrapper[4897]: I1121 14:22:49.736931 4897 scope.go:117] "RemoveContainer" containerID="ce5fd0711c68e5c357a10c05e310e1f405ca32797f4158e68cd1f8f0483b9a28" Nov 21 14:22:49 crc kubenswrapper[4897]: E1121 14:22:49.737194 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce5fd0711c68e5c357a10c05e310e1f405ca32797f4158e68cd1f8f0483b9a28\": container with ID starting with ce5fd0711c68e5c357a10c05e310e1f405ca32797f4158e68cd1f8f0483b9a28 not found: ID does not exist" containerID="ce5fd0711c68e5c357a10c05e310e1f405ca32797f4158e68cd1f8f0483b9a28" Nov 21 14:22:49 crc kubenswrapper[4897]: I1121 14:22:49.737221 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce5fd0711c68e5c357a10c05e310e1f405ca32797f4158e68cd1f8f0483b9a28"} err="failed to get container status \"ce5fd0711c68e5c357a10c05e310e1f405ca32797f4158e68cd1f8f0483b9a28\": rpc error: code = NotFound desc = could not find container \"ce5fd0711c68e5c357a10c05e310e1f405ca32797f4158e68cd1f8f0483b9a28\": container with ID starting with ce5fd0711c68e5c357a10c05e310e1f405ca32797f4158e68cd1f8f0483b9a28 not found: ID does not exist" Nov 21 14:22:49 crc kubenswrapper[4897]: I1121 14:22:49.737237 4897 scope.go:117] "RemoveContainer" containerID="a224fe6da8a6f7a0bfa666188c60d2aaf4a04cc8c25bb508a09ccdb68c32594d" Nov 21 14:22:49 crc kubenswrapper[4897]: E1121 14:22:49.737493 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a224fe6da8a6f7a0bfa666188c60d2aaf4a04cc8c25bb508a09ccdb68c32594d\": container with ID starting with a224fe6da8a6f7a0bfa666188c60d2aaf4a04cc8c25bb508a09ccdb68c32594d not found: ID does not exist" containerID="a224fe6da8a6f7a0bfa666188c60d2aaf4a04cc8c25bb508a09ccdb68c32594d" Nov 21 14:22:49 crc kubenswrapper[4897]: I1121 14:22:49.737549 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a224fe6da8a6f7a0bfa666188c60d2aaf4a04cc8c25bb508a09ccdb68c32594d"} err="failed to get container status \"a224fe6da8a6f7a0bfa666188c60d2aaf4a04cc8c25bb508a09ccdb68c32594d\": rpc error: code = NotFound desc = could not find container \"a224fe6da8a6f7a0bfa666188c60d2aaf4a04cc8c25bb508a09ccdb68c32594d\": container with ID starting with a224fe6da8a6f7a0bfa666188c60d2aaf4a04cc8c25bb508a09ccdb68c32594d not found: ID does not exist" Nov 21 14:22:49 crc kubenswrapper[4897]: I1121 14:22:49.737575 4897 scope.go:117] "RemoveContainer" containerID="258f84e0f83863a4b94aaa1e31acb431c9347c825add06ebd5a5685d78bed9b0" Nov 21 14:22:49 crc kubenswrapper[4897]: I1121 14:22:49.753788 4897 scope.go:117] "RemoveContainer" containerID="1582540efa85f2707892dcdbb9b10192ff5d7e87820b41e33297218eed037f3e" Nov 21 14:22:49 crc kubenswrapper[4897]: I1121 14:22:49.805610 4897 scope.go:117] "RemoveContainer" containerID="f346513a16dbce7ddbb97138b404f1ea60ca3fd82a6bc46c40fcea1f5663708a" Nov 21 14:22:50 crc kubenswrapper[4897]: I1121 14:22:50.122612 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ed64fe6-98d9-4f1c-8da9-e521c2604b16" path="/var/lib/kubelet/pods/6ed64fe6-98d9-4f1c-8da9-e521c2604b16/volumes" Nov 21 14:22:50 crc kubenswrapper[4897]: I1121 14:22:50.123678 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9920eebe-d0e6-4faa-9f8c-2dc8f1392770" path="/var/lib/kubelet/pods/9920eebe-d0e6-4faa-9f8c-2dc8f1392770/volumes" Nov 21 14:22:51 crc kubenswrapper[4897]: I1121 14:22:51.682030 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7bh27" event={"ID":"40f12519-bf8d-4792-a4ba-f4d64a3f3b93","Type":"ContainerStarted","Data":"e2f8433b9450638ad1fd0afe959acfab63dd3c88d5a12c5b3c590dae389dbecc"} Nov 21 14:22:51 crc kubenswrapper[4897]: I1121 14:22:51.704197 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-7bh27" podStartSLOduration=7.715882662 podStartE2EDuration="15.704180521s" podCreationTimestamp="2025-11-21 14:22:36 +0000 UTC" firstStartedPulling="2025-11-21 14:22:42.591573892 +0000 UTC m=+839.876167377" lastFinishedPulling="2025-11-21 14:22:50.579871761 +0000 UTC m=+847.864465236" observedRunningTime="2025-11-21 14:22:51.698304371 +0000 UTC m=+848.982897886" watchObservedRunningTime="2025-11-21 14:22:51.704180521 +0000 UTC m=+848.988773996" Nov 21 14:22:56 crc kubenswrapper[4897]: I1121 14:22:56.035223 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-dvr4v"] Nov 21 14:22:56 crc kubenswrapper[4897]: E1121 14:22:56.035883 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ed64fe6-98d9-4f1c-8da9-e521c2604b16" containerName="registry-server" Nov 21 14:22:56 crc kubenswrapper[4897]: I1121 14:22:56.035898 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ed64fe6-98d9-4f1c-8da9-e521c2604b16" containerName="registry-server" Nov 21 14:22:56 crc kubenswrapper[4897]: E1121 14:22:56.035912 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ed64fe6-98d9-4f1c-8da9-e521c2604b16" containerName="extract-utilities" Nov 21 14:22:56 crc kubenswrapper[4897]: I1121 14:22:56.035920 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ed64fe6-98d9-4f1c-8da9-e521c2604b16" containerName="extract-utilities" Nov 21 14:22:56 crc kubenswrapper[4897]: E1121 14:22:56.035931 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9920eebe-d0e6-4faa-9f8c-2dc8f1392770" containerName="extract-utilities" Nov 21 14:22:56 crc kubenswrapper[4897]: I1121 14:22:56.035939 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="9920eebe-d0e6-4faa-9f8c-2dc8f1392770" containerName="extract-utilities" Nov 21 14:22:56 crc kubenswrapper[4897]: E1121 14:22:56.035961 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9920eebe-d0e6-4faa-9f8c-2dc8f1392770" containerName="registry-server" Nov 21 14:22:56 crc kubenswrapper[4897]: I1121 14:22:56.035968 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="9920eebe-d0e6-4faa-9f8c-2dc8f1392770" containerName="registry-server" Nov 21 14:22:56 crc kubenswrapper[4897]: E1121 14:22:56.035980 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9920eebe-d0e6-4faa-9f8c-2dc8f1392770" containerName="extract-content" Nov 21 14:22:56 crc kubenswrapper[4897]: I1121 14:22:56.035986 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="9920eebe-d0e6-4faa-9f8c-2dc8f1392770" containerName="extract-content" Nov 21 14:22:56 crc kubenswrapper[4897]: E1121 14:22:56.035995 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ed64fe6-98d9-4f1c-8da9-e521c2604b16" containerName="extract-content" Nov 21 14:22:56 crc kubenswrapper[4897]: I1121 14:22:56.036002 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ed64fe6-98d9-4f1c-8da9-e521c2604b16" containerName="extract-content" Nov 21 14:22:56 crc kubenswrapper[4897]: I1121 14:22:56.036152 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="9920eebe-d0e6-4faa-9f8c-2dc8f1392770" containerName="registry-server" Nov 21 14:22:56 crc kubenswrapper[4897]: I1121 14:22:56.036169 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ed64fe6-98d9-4f1c-8da9-e521c2604b16" containerName="registry-server" Nov 21 14:22:56 crc kubenswrapper[4897]: I1121 14:22:56.037393 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dvr4v" Nov 21 14:22:56 crc kubenswrapper[4897]: I1121 14:22:56.056133 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dvr4v"] Nov 21 14:22:56 crc kubenswrapper[4897]: I1121 14:22:56.126244 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ebc2d11c-7430-4ccc-ba13-ad3635215760-catalog-content\") pod \"redhat-marketplace-dvr4v\" (UID: \"ebc2d11c-7430-4ccc-ba13-ad3635215760\") " pod="openshift-marketplace/redhat-marketplace-dvr4v" Nov 21 14:22:56 crc kubenswrapper[4897]: I1121 14:22:56.126457 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hmm46\" (UniqueName: \"kubernetes.io/projected/ebc2d11c-7430-4ccc-ba13-ad3635215760-kube-api-access-hmm46\") pod \"redhat-marketplace-dvr4v\" (UID: \"ebc2d11c-7430-4ccc-ba13-ad3635215760\") " pod="openshift-marketplace/redhat-marketplace-dvr4v" Nov 21 14:22:56 crc kubenswrapper[4897]: I1121 14:22:56.126769 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ebc2d11c-7430-4ccc-ba13-ad3635215760-utilities\") pod \"redhat-marketplace-dvr4v\" (UID: \"ebc2d11c-7430-4ccc-ba13-ad3635215760\") " pod="openshift-marketplace/redhat-marketplace-dvr4v" Nov 21 14:22:56 crc kubenswrapper[4897]: I1121 14:22:56.228060 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hmm46\" (UniqueName: \"kubernetes.io/projected/ebc2d11c-7430-4ccc-ba13-ad3635215760-kube-api-access-hmm46\") pod \"redhat-marketplace-dvr4v\" (UID: \"ebc2d11c-7430-4ccc-ba13-ad3635215760\") " pod="openshift-marketplace/redhat-marketplace-dvr4v" Nov 21 14:22:56 crc kubenswrapper[4897]: I1121 14:22:56.228141 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ebc2d11c-7430-4ccc-ba13-ad3635215760-utilities\") pod \"redhat-marketplace-dvr4v\" (UID: \"ebc2d11c-7430-4ccc-ba13-ad3635215760\") " pod="openshift-marketplace/redhat-marketplace-dvr4v" Nov 21 14:22:56 crc kubenswrapper[4897]: I1121 14:22:56.228196 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ebc2d11c-7430-4ccc-ba13-ad3635215760-catalog-content\") pod \"redhat-marketplace-dvr4v\" (UID: \"ebc2d11c-7430-4ccc-ba13-ad3635215760\") " pod="openshift-marketplace/redhat-marketplace-dvr4v" Nov 21 14:22:56 crc kubenswrapper[4897]: I1121 14:22:56.228821 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ebc2d11c-7430-4ccc-ba13-ad3635215760-catalog-content\") pod \"redhat-marketplace-dvr4v\" (UID: \"ebc2d11c-7430-4ccc-ba13-ad3635215760\") " pod="openshift-marketplace/redhat-marketplace-dvr4v" Nov 21 14:22:56 crc kubenswrapper[4897]: I1121 14:22:56.228973 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ebc2d11c-7430-4ccc-ba13-ad3635215760-utilities\") pod \"redhat-marketplace-dvr4v\" (UID: \"ebc2d11c-7430-4ccc-ba13-ad3635215760\") " pod="openshift-marketplace/redhat-marketplace-dvr4v" Nov 21 14:22:56 crc kubenswrapper[4897]: I1121 14:22:56.249094 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hmm46\" (UniqueName: \"kubernetes.io/projected/ebc2d11c-7430-4ccc-ba13-ad3635215760-kube-api-access-hmm46\") pod \"redhat-marketplace-dvr4v\" (UID: \"ebc2d11c-7430-4ccc-ba13-ad3635215760\") " pod="openshift-marketplace/redhat-marketplace-dvr4v" Nov 21 14:22:56 crc kubenswrapper[4897]: I1121 14:22:56.360437 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dvr4v" Nov 21 14:22:56 crc kubenswrapper[4897]: I1121 14:22:56.518738 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-7bh27" Nov 21 14:22:56 crc kubenswrapper[4897]: I1121 14:22:56.519160 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-7bh27" Nov 21 14:22:56 crc kubenswrapper[4897]: I1121 14:22:56.570664 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-7bh27" Nov 21 14:22:56 crc kubenswrapper[4897]: I1121 14:22:56.756394 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-7bh27" Nov 21 14:22:56 crc kubenswrapper[4897]: I1121 14:22:56.822167 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dvr4v"] Nov 21 14:22:57 crc kubenswrapper[4897]: I1121 14:22:57.729167 4897 generic.go:334] "Generic (PLEG): container finished" podID="ebc2d11c-7430-4ccc-ba13-ad3635215760" containerID="e406ac4dcc45af4000049ce215d96afdd379d7b41b10e28e6c3afc870adcda27" exitCode=0 Nov 21 14:22:57 crc kubenswrapper[4897]: I1121 14:22:57.729288 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dvr4v" event={"ID":"ebc2d11c-7430-4ccc-ba13-ad3635215760","Type":"ContainerDied","Data":"e406ac4dcc45af4000049ce215d96afdd379d7b41b10e28e6c3afc870adcda27"} Nov 21 14:22:57 crc kubenswrapper[4897]: I1121 14:22:57.729624 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dvr4v" event={"ID":"ebc2d11c-7430-4ccc-ba13-ad3635215760","Type":"ContainerStarted","Data":"2773b77a864672e303dbb1d1ae91cbd40bb1e2f918cb36b14eb217a77307965b"} Nov 21 14:22:58 crc kubenswrapper[4897]: I1121 14:22:58.826859 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7bh27"] Nov 21 14:22:58 crc kubenswrapper[4897]: I1121 14:22:58.827124 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-7bh27" podUID="40f12519-bf8d-4792-a4ba-f4d64a3f3b93" containerName="registry-server" containerID="cri-o://e2f8433b9450638ad1fd0afe959acfab63dd3c88d5a12c5b3c590dae389dbecc" gracePeriod=2 Nov 21 14:22:59 crc kubenswrapper[4897]: I1121 14:22:59.743613 4897 generic.go:334] "Generic (PLEG): container finished" podID="ebc2d11c-7430-4ccc-ba13-ad3635215760" containerID="b602e769bf7b9dfe3032546f152a21e027f614b425be184a0b55c652035266e4" exitCode=0 Nov 21 14:22:59 crc kubenswrapper[4897]: I1121 14:22:59.743680 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dvr4v" event={"ID":"ebc2d11c-7430-4ccc-ba13-ad3635215760","Type":"ContainerDied","Data":"b602e769bf7b9dfe3032546f152a21e027f614b425be184a0b55c652035266e4"} Nov 21 14:22:59 crc kubenswrapper[4897]: I1121 14:22:59.748550 4897 generic.go:334] "Generic (PLEG): container finished" podID="40f12519-bf8d-4792-a4ba-f4d64a3f3b93" containerID="e2f8433b9450638ad1fd0afe959acfab63dd3c88d5a12c5b3c590dae389dbecc" exitCode=0 Nov 21 14:22:59 crc kubenswrapper[4897]: I1121 14:22:59.748587 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7bh27" event={"ID":"40f12519-bf8d-4792-a4ba-f4d64a3f3b93","Type":"ContainerDied","Data":"e2f8433b9450638ad1fd0afe959acfab63dd3c88d5a12c5b3c590dae389dbecc"} Nov 21 14:22:59 crc kubenswrapper[4897]: I1121 14:22:59.835836 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7bh27" Nov 21 14:22:59 crc kubenswrapper[4897]: I1121 14:22:59.884111 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40f12519-bf8d-4792-a4ba-f4d64a3f3b93-catalog-content\") pod \"40f12519-bf8d-4792-a4ba-f4d64a3f3b93\" (UID: \"40f12519-bf8d-4792-a4ba-f4d64a3f3b93\") " Nov 21 14:22:59 crc kubenswrapper[4897]: I1121 14:22:59.884187 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mpt8v\" (UniqueName: \"kubernetes.io/projected/40f12519-bf8d-4792-a4ba-f4d64a3f3b93-kube-api-access-mpt8v\") pod \"40f12519-bf8d-4792-a4ba-f4d64a3f3b93\" (UID: \"40f12519-bf8d-4792-a4ba-f4d64a3f3b93\") " Nov 21 14:22:59 crc kubenswrapper[4897]: I1121 14:22:59.884216 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40f12519-bf8d-4792-a4ba-f4d64a3f3b93-utilities\") pod \"40f12519-bf8d-4792-a4ba-f4d64a3f3b93\" (UID: \"40f12519-bf8d-4792-a4ba-f4d64a3f3b93\") " Nov 21 14:22:59 crc kubenswrapper[4897]: I1121 14:22:59.885169 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/40f12519-bf8d-4792-a4ba-f4d64a3f3b93-utilities" (OuterVolumeSpecName: "utilities") pod "40f12519-bf8d-4792-a4ba-f4d64a3f3b93" (UID: "40f12519-bf8d-4792-a4ba-f4d64a3f3b93"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:22:59 crc kubenswrapper[4897]: I1121 14:22:59.889833 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/40f12519-bf8d-4792-a4ba-f4d64a3f3b93-kube-api-access-mpt8v" (OuterVolumeSpecName: "kube-api-access-mpt8v") pod "40f12519-bf8d-4792-a4ba-f4d64a3f3b93" (UID: "40f12519-bf8d-4792-a4ba-f4d64a3f3b93"). InnerVolumeSpecName "kube-api-access-mpt8v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:22:59 crc kubenswrapper[4897]: I1121 14:22:59.936732 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/40f12519-bf8d-4792-a4ba-f4d64a3f3b93-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "40f12519-bf8d-4792-a4ba-f4d64a3f3b93" (UID: "40f12519-bf8d-4792-a4ba-f4d64a3f3b93"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:22:59 crc kubenswrapper[4897]: I1121 14:22:59.985554 4897 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40f12519-bf8d-4792-a4ba-f4d64a3f3b93-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 21 14:22:59 crc kubenswrapper[4897]: I1121 14:22:59.985595 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mpt8v\" (UniqueName: \"kubernetes.io/projected/40f12519-bf8d-4792-a4ba-f4d64a3f3b93-kube-api-access-mpt8v\") on node \"crc\" DevicePath \"\"" Nov 21 14:22:59 crc kubenswrapper[4897]: I1121 14:22:59.985606 4897 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40f12519-bf8d-4792-a4ba-f4d64a3f3b93-utilities\") on node \"crc\" DevicePath \"\"" Nov 21 14:23:00 crc kubenswrapper[4897]: I1121 14:23:00.385269 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/collector-9fggc"] Nov 21 14:23:00 crc kubenswrapper[4897]: E1121 14:23:00.385876 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40f12519-bf8d-4792-a4ba-f4d64a3f3b93" containerName="registry-server" Nov 21 14:23:00 crc kubenswrapper[4897]: I1121 14:23:00.386359 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="40f12519-bf8d-4792-a4ba-f4d64a3f3b93" containerName="registry-server" Nov 21 14:23:00 crc kubenswrapper[4897]: E1121 14:23:00.386386 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40f12519-bf8d-4792-a4ba-f4d64a3f3b93" containerName="extract-content" Nov 21 14:23:00 crc kubenswrapper[4897]: I1121 14:23:00.386394 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="40f12519-bf8d-4792-a4ba-f4d64a3f3b93" containerName="extract-content" Nov 21 14:23:00 crc kubenswrapper[4897]: E1121 14:23:00.386420 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40f12519-bf8d-4792-a4ba-f4d64a3f3b93" containerName="extract-utilities" Nov 21 14:23:00 crc kubenswrapper[4897]: I1121 14:23:00.386429 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="40f12519-bf8d-4792-a4ba-f4d64a3f3b93" containerName="extract-utilities" Nov 21 14:23:00 crc kubenswrapper[4897]: I1121 14:23:00.386635 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="40f12519-bf8d-4792-a4ba-f4d64a3f3b93" containerName="registry-server" Nov 21 14:23:00 crc kubenswrapper[4897]: I1121 14:23:00.387254 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-9fggc" Nov 21 14:23:00 crc kubenswrapper[4897]: I1121 14:23:00.393555 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-syslog-receiver" Nov 21 14:23:00 crc kubenswrapper[4897]: I1121 14:23:00.393802 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-metrics" Nov 21 14:23:00 crc kubenswrapper[4897]: I1121 14:23:00.393967 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-dockercfg-kqrh2" Nov 21 14:23:00 crc kubenswrapper[4897]: I1121 14:23:00.394131 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-token" Nov 21 14:23:00 crc kubenswrapper[4897]: I1121 14:23:00.394268 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"collector-config" Nov 21 14:23:00 crc kubenswrapper[4897]: I1121 14:23:00.397197 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"collector-trustbundle" Nov 21 14:23:00 crc kubenswrapper[4897]: I1121 14:23:00.399845 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/collector-9fggc"] Nov 21 14:23:00 crc kubenswrapper[4897]: I1121 14:23:00.463848 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-logging/collector-9fggc"] Nov 21 14:23:00 crc kubenswrapper[4897]: E1121 14:23:00.464484 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[collector-syslog-receiver collector-token config config-openshift-service-cacrt datadir entrypoint kube-api-access-4gmqx metrics sa-token tmp trusted-ca], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openshift-logging/collector-9fggc" podUID="8e92cb91-808f-4981-9b49-6564bfcd408f" Nov 21 14:23:00 crc kubenswrapper[4897]: I1121 14:23:00.500469 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/8e92cb91-808f-4981-9b49-6564bfcd408f-datadir\") pod \"collector-9fggc\" (UID: \"8e92cb91-808f-4981-9b49-6564bfcd408f\") " pod="openshift-logging/collector-9fggc" Nov 21 14:23:00 crc kubenswrapper[4897]: I1121 14:23:00.500613 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e92cb91-808f-4981-9b49-6564bfcd408f-config\") pod \"collector-9fggc\" (UID: \"8e92cb91-808f-4981-9b49-6564bfcd408f\") " pod="openshift-logging/collector-9fggc" Nov 21 14:23:00 crc kubenswrapper[4897]: I1121 14:23:00.500639 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/8e92cb91-808f-4981-9b49-6564bfcd408f-collector-token\") pod \"collector-9fggc\" (UID: \"8e92cb91-808f-4981-9b49-6564bfcd408f\") " pod="openshift-logging/collector-9fggc" Nov 21 14:23:00 crc kubenswrapper[4897]: I1121 14:23:00.501089 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/8e92cb91-808f-4981-9b49-6564bfcd408f-metrics\") pod \"collector-9fggc\" (UID: \"8e92cb91-808f-4981-9b49-6564bfcd408f\") " pod="openshift-logging/collector-9fggc" Nov 21 14:23:00 crc kubenswrapper[4897]: I1121 14:23:00.501115 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/8e92cb91-808f-4981-9b49-6564bfcd408f-tmp\") pod \"collector-9fggc\" (UID: \"8e92cb91-808f-4981-9b49-6564bfcd408f\") " pod="openshift-logging/collector-9fggc" Nov 21 14:23:00 crc kubenswrapper[4897]: I1121 14:23:00.501176 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8e92cb91-808f-4981-9b49-6564bfcd408f-trusted-ca\") pod \"collector-9fggc\" (UID: \"8e92cb91-808f-4981-9b49-6564bfcd408f\") " pod="openshift-logging/collector-9fggc" Nov 21 14:23:00 crc kubenswrapper[4897]: I1121 14:23:00.501199 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/8e92cb91-808f-4981-9b49-6564bfcd408f-sa-token\") pod \"collector-9fggc\" (UID: \"8e92cb91-808f-4981-9b49-6564bfcd408f\") " pod="openshift-logging/collector-9fggc" Nov 21 14:23:00 crc kubenswrapper[4897]: I1121 14:23:00.501360 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/8e92cb91-808f-4981-9b49-6564bfcd408f-entrypoint\") pod \"collector-9fggc\" (UID: \"8e92cb91-808f-4981-9b49-6564bfcd408f\") " pod="openshift-logging/collector-9fggc" Nov 21 14:23:00 crc kubenswrapper[4897]: I1121 14:23:00.501394 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/8e92cb91-808f-4981-9b49-6564bfcd408f-config-openshift-service-cacrt\") pod \"collector-9fggc\" (UID: \"8e92cb91-808f-4981-9b49-6564bfcd408f\") " pod="openshift-logging/collector-9fggc" Nov 21 14:23:00 crc kubenswrapper[4897]: I1121 14:23:00.501427 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4gmqx\" (UniqueName: \"kubernetes.io/projected/8e92cb91-808f-4981-9b49-6564bfcd408f-kube-api-access-4gmqx\") pod \"collector-9fggc\" (UID: \"8e92cb91-808f-4981-9b49-6564bfcd408f\") " pod="openshift-logging/collector-9fggc" Nov 21 14:23:00 crc kubenswrapper[4897]: I1121 14:23:00.501458 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/8e92cb91-808f-4981-9b49-6564bfcd408f-collector-syslog-receiver\") pod \"collector-9fggc\" (UID: \"8e92cb91-808f-4981-9b49-6564bfcd408f\") " pod="openshift-logging/collector-9fggc" Nov 21 14:23:00 crc kubenswrapper[4897]: I1121 14:23:00.602832 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/8e92cb91-808f-4981-9b49-6564bfcd408f-metrics\") pod \"collector-9fggc\" (UID: \"8e92cb91-808f-4981-9b49-6564bfcd408f\") " pod="openshift-logging/collector-9fggc" Nov 21 14:23:00 crc kubenswrapper[4897]: I1121 14:23:00.602868 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/8e92cb91-808f-4981-9b49-6564bfcd408f-tmp\") pod \"collector-9fggc\" (UID: \"8e92cb91-808f-4981-9b49-6564bfcd408f\") " pod="openshift-logging/collector-9fggc" Nov 21 14:23:00 crc kubenswrapper[4897]: I1121 14:23:00.602891 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8e92cb91-808f-4981-9b49-6564bfcd408f-trusted-ca\") pod \"collector-9fggc\" (UID: \"8e92cb91-808f-4981-9b49-6564bfcd408f\") " pod="openshift-logging/collector-9fggc" Nov 21 14:23:00 crc kubenswrapper[4897]: I1121 14:23:00.602911 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/8e92cb91-808f-4981-9b49-6564bfcd408f-sa-token\") pod \"collector-9fggc\" (UID: \"8e92cb91-808f-4981-9b49-6564bfcd408f\") " pod="openshift-logging/collector-9fggc" Nov 21 14:23:00 crc kubenswrapper[4897]: I1121 14:23:00.602940 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/8e92cb91-808f-4981-9b49-6564bfcd408f-entrypoint\") pod \"collector-9fggc\" (UID: \"8e92cb91-808f-4981-9b49-6564bfcd408f\") " pod="openshift-logging/collector-9fggc" Nov 21 14:23:00 crc kubenswrapper[4897]: I1121 14:23:00.602960 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/8e92cb91-808f-4981-9b49-6564bfcd408f-config-openshift-service-cacrt\") pod \"collector-9fggc\" (UID: \"8e92cb91-808f-4981-9b49-6564bfcd408f\") " pod="openshift-logging/collector-9fggc" Nov 21 14:23:00 crc kubenswrapper[4897]: I1121 14:23:00.602992 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4gmqx\" (UniqueName: \"kubernetes.io/projected/8e92cb91-808f-4981-9b49-6564bfcd408f-kube-api-access-4gmqx\") pod \"collector-9fggc\" (UID: \"8e92cb91-808f-4981-9b49-6564bfcd408f\") " pod="openshift-logging/collector-9fggc" Nov 21 14:23:00 crc kubenswrapper[4897]: I1121 14:23:00.603015 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/8e92cb91-808f-4981-9b49-6564bfcd408f-collector-syslog-receiver\") pod \"collector-9fggc\" (UID: \"8e92cb91-808f-4981-9b49-6564bfcd408f\") " pod="openshift-logging/collector-9fggc" Nov 21 14:23:00 crc kubenswrapper[4897]: E1121 14:23:00.603033 4897 secret.go:188] Couldn't get secret openshift-logging/collector-metrics: secret "collector-metrics" not found Nov 21 14:23:00 crc kubenswrapper[4897]: I1121 14:23:00.603081 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/8e92cb91-808f-4981-9b49-6564bfcd408f-datadir\") pod \"collector-9fggc\" (UID: \"8e92cb91-808f-4981-9b49-6564bfcd408f\") " pod="openshift-logging/collector-9fggc" Nov 21 14:23:00 crc kubenswrapper[4897]: E1121 14:23:00.603113 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8e92cb91-808f-4981-9b49-6564bfcd408f-metrics podName:8e92cb91-808f-4981-9b49-6564bfcd408f nodeName:}" failed. No retries permitted until 2025-11-21 14:23:01.103092906 +0000 UTC m=+858.387686421 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics" (UniqueName: "kubernetes.io/secret/8e92cb91-808f-4981-9b49-6564bfcd408f-metrics") pod "collector-9fggc" (UID: "8e92cb91-808f-4981-9b49-6564bfcd408f") : secret "collector-metrics" not found Nov 21 14:23:00 crc kubenswrapper[4897]: I1121 14:23:00.603051 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/8e92cb91-808f-4981-9b49-6564bfcd408f-datadir\") pod \"collector-9fggc\" (UID: \"8e92cb91-808f-4981-9b49-6564bfcd408f\") " pod="openshift-logging/collector-9fggc" Nov 21 14:23:00 crc kubenswrapper[4897]: I1121 14:23:00.603192 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e92cb91-808f-4981-9b49-6564bfcd408f-config\") pod \"collector-9fggc\" (UID: \"8e92cb91-808f-4981-9b49-6564bfcd408f\") " pod="openshift-logging/collector-9fggc" Nov 21 14:23:00 crc kubenswrapper[4897]: I1121 14:23:00.603223 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/8e92cb91-808f-4981-9b49-6564bfcd408f-collector-token\") pod \"collector-9fggc\" (UID: \"8e92cb91-808f-4981-9b49-6564bfcd408f\") " pod="openshift-logging/collector-9fggc" Nov 21 14:23:00 crc kubenswrapper[4897]: I1121 14:23:00.603878 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8e92cb91-808f-4981-9b49-6564bfcd408f-trusted-ca\") pod \"collector-9fggc\" (UID: \"8e92cb91-808f-4981-9b49-6564bfcd408f\") " pod="openshift-logging/collector-9fggc" Nov 21 14:23:00 crc kubenswrapper[4897]: I1121 14:23:00.604177 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/8e92cb91-808f-4981-9b49-6564bfcd408f-config-openshift-service-cacrt\") pod \"collector-9fggc\" (UID: \"8e92cb91-808f-4981-9b49-6564bfcd408f\") " pod="openshift-logging/collector-9fggc" Nov 21 14:23:00 crc kubenswrapper[4897]: I1121 14:23:00.604661 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e92cb91-808f-4981-9b49-6564bfcd408f-config\") pod \"collector-9fggc\" (UID: \"8e92cb91-808f-4981-9b49-6564bfcd408f\") " pod="openshift-logging/collector-9fggc" Nov 21 14:23:00 crc kubenswrapper[4897]: I1121 14:23:00.605029 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/8e92cb91-808f-4981-9b49-6564bfcd408f-entrypoint\") pod \"collector-9fggc\" (UID: \"8e92cb91-808f-4981-9b49-6564bfcd408f\") " pod="openshift-logging/collector-9fggc" Nov 21 14:23:00 crc kubenswrapper[4897]: I1121 14:23:00.613902 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/8e92cb91-808f-4981-9b49-6564bfcd408f-tmp\") pod \"collector-9fggc\" (UID: \"8e92cb91-808f-4981-9b49-6564bfcd408f\") " pod="openshift-logging/collector-9fggc" Nov 21 14:23:00 crc kubenswrapper[4897]: I1121 14:23:00.614275 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/8e92cb91-808f-4981-9b49-6564bfcd408f-collector-syslog-receiver\") pod \"collector-9fggc\" (UID: \"8e92cb91-808f-4981-9b49-6564bfcd408f\") " pod="openshift-logging/collector-9fggc" Nov 21 14:23:00 crc kubenswrapper[4897]: I1121 14:23:00.614415 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/8e92cb91-808f-4981-9b49-6564bfcd408f-collector-token\") pod \"collector-9fggc\" (UID: \"8e92cb91-808f-4981-9b49-6564bfcd408f\") " pod="openshift-logging/collector-9fggc" Nov 21 14:23:00 crc kubenswrapper[4897]: I1121 14:23:00.629254 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4gmqx\" (UniqueName: \"kubernetes.io/projected/8e92cb91-808f-4981-9b49-6564bfcd408f-kube-api-access-4gmqx\") pod \"collector-9fggc\" (UID: \"8e92cb91-808f-4981-9b49-6564bfcd408f\") " pod="openshift-logging/collector-9fggc" Nov 21 14:23:00 crc kubenswrapper[4897]: I1121 14:23:00.633919 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/8e92cb91-808f-4981-9b49-6564bfcd408f-sa-token\") pod \"collector-9fggc\" (UID: \"8e92cb91-808f-4981-9b49-6564bfcd408f\") " pod="openshift-logging/collector-9fggc" Nov 21 14:23:00 crc kubenswrapper[4897]: I1121 14:23:00.755645 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-9fggc" Nov 21 14:23:00 crc kubenswrapper[4897]: I1121 14:23:00.756207 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7bh27" Nov 21 14:23:00 crc kubenswrapper[4897]: I1121 14:23:00.756607 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7bh27" event={"ID":"40f12519-bf8d-4792-a4ba-f4d64a3f3b93","Type":"ContainerDied","Data":"7f428a58f9a5f23a45d6ab30352c35d6d1d0805b8a3b429842ea2bdd664b3dc4"} Nov 21 14:23:00 crc kubenswrapper[4897]: I1121 14:23:00.756662 4897 scope.go:117] "RemoveContainer" containerID="e2f8433b9450638ad1fd0afe959acfab63dd3c88d5a12c5b3c590dae389dbecc" Nov 21 14:23:00 crc kubenswrapper[4897]: I1121 14:23:00.764256 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-9fggc" Nov 21 14:23:00 crc kubenswrapper[4897]: I1121 14:23:00.773619 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7bh27"] Nov 21 14:23:00 crc kubenswrapper[4897]: I1121 14:23:00.777825 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-7bh27"] Nov 21 14:23:00 crc kubenswrapper[4897]: I1121 14:23:00.838791 4897 scope.go:117] "RemoveContainer" containerID="0c0c9b2aca5dcd69d12d194b7dc2fdc181992481faf70ed283b72e5629fd2d71" Nov 21 14:23:00 crc kubenswrapper[4897]: I1121 14:23:00.907247 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/8e92cb91-808f-4981-9b49-6564bfcd408f-collector-syslog-receiver\") pod \"8e92cb91-808f-4981-9b49-6564bfcd408f\" (UID: \"8e92cb91-808f-4981-9b49-6564bfcd408f\") " Nov 21 14:23:00 crc kubenswrapper[4897]: I1121 14:23:00.907647 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8e92cb91-808f-4981-9b49-6564bfcd408f-trusted-ca\") pod \"8e92cb91-808f-4981-9b49-6564bfcd408f\" (UID: \"8e92cb91-808f-4981-9b49-6564bfcd408f\") " Nov 21 14:23:00 crc kubenswrapper[4897]: I1121 14:23:00.907716 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/8e92cb91-808f-4981-9b49-6564bfcd408f-tmp\") pod \"8e92cb91-808f-4981-9b49-6564bfcd408f\" (UID: \"8e92cb91-808f-4981-9b49-6564bfcd408f\") " Nov 21 14:23:00 crc kubenswrapper[4897]: I1121 14:23:00.907768 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4gmqx\" (UniqueName: \"kubernetes.io/projected/8e92cb91-808f-4981-9b49-6564bfcd408f-kube-api-access-4gmqx\") pod \"8e92cb91-808f-4981-9b49-6564bfcd408f\" (UID: \"8e92cb91-808f-4981-9b49-6564bfcd408f\") " Nov 21 14:23:00 crc kubenswrapper[4897]: I1121 14:23:00.907889 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/8e92cb91-808f-4981-9b49-6564bfcd408f-config-openshift-service-cacrt\") pod \"8e92cb91-808f-4981-9b49-6564bfcd408f\" (UID: \"8e92cb91-808f-4981-9b49-6564bfcd408f\") " Nov 21 14:23:00 crc kubenswrapper[4897]: I1121 14:23:00.907970 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/8e92cb91-808f-4981-9b49-6564bfcd408f-sa-token\") pod \"8e92cb91-808f-4981-9b49-6564bfcd408f\" (UID: \"8e92cb91-808f-4981-9b49-6564bfcd408f\") " Nov 21 14:23:00 crc kubenswrapper[4897]: I1121 14:23:00.908747 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/8e92cb91-808f-4981-9b49-6564bfcd408f-datadir\") pod \"8e92cb91-808f-4981-9b49-6564bfcd408f\" (UID: \"8e92cb91-808f-4981-9b49-6564bfcd408f\") " Nov 21 14:23:00 crc kubenswrapper[4897]: I1121 14:23:00.908813 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e92cb91-808f-4981-9b49-6564bfcd408f-config-openshift-service-cacrt" (OuterVolumeSpecName: "config-openshift-service-cacrt") pod "8e92cb91-808f-4981-9b49-6564bfcd408f" (UID: "8e92cb91-808f-4981-9b49-6564bfcd408f"). InnerVolumeSpecName "config-openshift-service-cacrt". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:23:00 crc kubenswrapper[4897]: I1121 14:23:00.908840 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e92cb91-808f-4981-9b49-6564bfcd408f-config\") pod \"8e92cb91-808f-4981-9b49-6564bfcd408f\" (UID: \"8e92cb91-808f-4981-9b49-6564bfcd408f\") " Nov 21 14:23:00 crc kubenswrapper[4897]: I1121 14:23:00.908877 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8e92cb91-808f-4981-9b49-6564bfcd408f-datadir" (OuterVolumeSpecName: "datadir") pod "8e92cb91-808f-4981-9b49-6564bfcd408f" (UID: "8e92cb91-808f-4981-9b49-6564bfcd408f"). InnerVolumeSpecName "datadir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 21 14:23:00 crc kubenswrapper[4897]: I1121 14:23:00.908904 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/8e92cb91-808f-4981-9b49-6564bfcd408f-collector-token\") pod \"8e92cb91-808f-4981-9b49-6564bfcd408f\" (UID: \"8e92cb91-808f-4981-9b49-6564bfcd408f\") " Nov 21 14:23:00 crc kubenswrapper[4897]: I1121 14:23:00.908963 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/8e92cb91-808f-4981-9b49-6564bfcd408f-entrypoint\") pod \"8e92cb91-808f-4981-9b49-6564bfcd408f\" (UID: \"8e92cb91-808f-4981-9b49-6564bfcd408f\") " Nov 21 14:23:00 crc kubenswrapper[4897]: I1121 14:23:00.909369 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e92cb91-808f-4981-9b49-6564bfcd408f-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8e92cb91-808f-4981-9b49-6564bfcd408f" (UID: "8e92cb91-808f-4981-9b49-6564bfcd408f"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:23:00 crc kubenswrapper[4897]: I1121 14:23:00.909375 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e92cb91-808f-4981-9b49-6564bfcd408f-config" (OuterVolumeSpecName: "config") pod "8e92cb91-808f-4981-9b49-6564bfcd408f" (UID: "8e92cb91-808f-4981-9b49-6564bfcd408f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:23:00 crc kubenswrapper[4897]: I1121 14:23:00.909754 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e92cb91-808f-4981-9b49-6564bfcd408f-entrypoint" (OuterVolumeSpecName: "entrypoint") pod "8e92cb91-808f-4981-9b49-6564bfcd408f" (UID: "8e92cb91-808f-4981-9b49-6564bfcd408f"). InnerVolumeSpecName "entrypoint". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:23:00 crc kubenswrapper[4897]: I1121 14:23:00.910433 4897 reconciler_common.go:293] "Volume detached for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/8e92cb91-808f-4981-9b49-6564bfcd408f-datadir\") on node \"crc\" DevicePath \"\"" Nov 21 14:23:00 crc kubenswrapper[4897]: I1121 14:23:00.910456 4897 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e92cb91-808f-4981-9b49-6564bfcd408f-config\") on node \"crc\" DevicePath \"\"" Nov 21 14:23:00 crc kubenswrapper[4897]: I1121 14:23:00.910468 4897 reconciler_common.go:293] "Volume detached for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/8e92cb91-808f-4981-9b49-6564bfcd408f-entrypoint\") on node \"crc\" DevicePath \"\"" Nov 21 14:23:00 crc kubenswrapper[4897]: I1121 14:23:00.910486 4897 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8e92cb91-808f-4981-9b49-6564bfcd408f-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 21 14:23:00 crc kubenswrapper[4897]: I1121 14:23:00.910500 4897 reconciler_common.go:293] "Volume detached for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/8e92cb91-808f-4981-9b49-6564bfcd408f-config-openshift-service-cacrt\") on node \"crc\" DevicePath \"\"" Nov 21 14:23:00 crc kubenswrapper[4897]: I1121 14:23:00.911284 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e92cb91-808f-4981-9b49-6564bfcd408f-collector-syslog-receiver" (OuterVolumeSpecName: "collector-syslog-receiver") pod "8e92cb91-808f-4981-9b49-6564bfcd408f" (UID: "8e92cb91-808f-4981-9b49-6564bfcd408f"). InnerVolumeSpecName "collector-syslog-receiver". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:23:00 crc kubenswrapper[4897]: I1121 14:23:00.912981 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e92cb91-808f-4981-9b49-6564bfcd408f-sa-token" (OuterVolumeSpecName: "sa-token") pod "8e92cb91-808f-4981-9b49-6564bfcd408f" (UID: "8e92cb91-808f-4981-9b49-6564bfcd408f"). InnerVolumeSpecName "sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:23:00 crc kubenswrapper[4897]: I1121 14:23:00.913663 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8e92cb91-808f-4981-9b49-6564bfcd408f-tmp" (OuterVolumeSpecName: "tmp") pod "8e92cb91-808f-4981-9b49-6564bfcd408f" (UID: "8e92cb91-808f-4981-9b49-6564bfcd408f"). InnerVolumeSpecName "tmp". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:23:00 crc kubenswrapper[4897]: I1121 14:23:00.922157 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e92cb91-808f-4981-9b49-6564bfcd408f-collector-token" (OuterVolumeSpecName: "collector-token") pod "8e92cb91-808f-4981-9b49-6564bfcd408f" (UID: "8e92cb91-808f-4981-9b49-6564bfcd408f"). InnerVolumeSpecName "collector-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:23:00 crc kubenswrapper[4897]: I1121 14:23:00.922478 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e92cb91-808f-4981-9b49-6564bfcd408f-kube-api-access-4gmqx" (OuterVolumeSpecName: "kube-api-access-4gmqx") pod "8e92cb91-808f-4981-9b49-6564bfcd408f" (UID: "8e92cb91-808f-4981-9b49-6564bfcd408f"). InnerVolumeSpecName "kube-api-access-4gmqx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:23:01 crc kubenswrapper[4897]: I1121 14:23:01.011184 4897 reconciler_common.go:293] "Volume detached for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/8e92cb91-808f-4981-9b49-6564bfcd408f-collector-token\") on node \"crc\" DevicePath \"\"" Nov 21 14:23:01 crc kubenswrapper[4897]: I1121 14:23:01.011229 4897 reconciler_common.go:293] "Volume detached for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/8e92cb91-808f-4981-9b49-6564bfcd408f-collector-syslog-receiver\") on node \"crc\" DevicePath \"\"" Nov 21 14:23:01 crc kubenswrapper[4897]: I1121 14:23:01.011243 4897 reconciler_common.go:293] "Volume detached for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/8e92cb91-808f-4981-9b49-6564bfcd408f-tmp\") on node \"crc\" DevicePath \"\"" Nov 21 14:23:01 crc kubenswrapper[4897]: I1121 14:23:01.011253 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4gmqx\" (UniqueName: \"kubernetes.io/projected/8e92cb91-808f-4981-9b49-6564bfcd408f-kube-api-access-4gmqx\") on node \"crc\" DevicePath \"\"" Nov 21 14:23:01 crc kubenswrapper[4897]: I1121 14:23:01.011264 4897 reconciler_common.go:293] "Volume detached for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/8e92cb91-808f-4981-9b49-6564bfcd408f-sa-token\") on node \"crc\" DevicePath \"\"" Nov 21 14:23:01 crc kubenswrapper[4897]: I1121 14:23:01.111797 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/8e92cb91-808f-4981-9b49-6564bfcd408f-metrics\") pod \"collector-9fggc\" (UID: \"8e92cb91-808f-4981-9b49-6564bfcd408f\") " pod="openshift-logging/collector-9fggc" Nov 21 14:23:01 crc kubenswrapper[4897]: I1121 14:23:01.115259 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/8e92cb91-808f-4981-9b49-6564bfcd408f-metrics\") pod \"collector-9fggc\" (UID: \"8e92cb91-808f-4981-9b49-6564bfcd408f\") " pod="openshift-logging/collector-9fggc" Nov 21 14:23:01 crc kubenswrapper[4897]: I1121 14:23:01.212858 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/8e92cb91-808f-4981-9b49-6564bfcd408f-metrics\") pod \"8e92cb91-808f-4981-9b49-6564bfcd408f\" (UID: \"8e92cb91-808f-4981-9b49-6564bfcd408f\") " Nov 21 14:23:01 crc kubenswrapper[4897]: I1121 14:23:01.217344 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e92cb91-808f-4981-9b49-6564bfcd408f-metrics" (OuterVolumeSpecName: "metrics") pod "8e92cb91-808f-4981-9b49-6564bfcd408f" (UID: "8e92cb91-808f-4981-9b49-6564bfcd408f"). InnerVolumeSpecName "metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:23:01 crc kubenswrapper[4897]: I1121 14:23:01.243544 4897 scope.go:117] "RemoveContainer" containerID="0e61b61751340c191384d392e6ce56c6768f85afa55788b1093d21e9b64e9fb5" Nov 21 14:23:01 crc kubenswrapper[4897]: I1121 14:23:01.314957 4897 reconciler_common.go:293] "Volume detached for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/8e92cb91-808f-4981-9b49-6564bfcd408f-metrics\") on node \"crc\" DevicePath \"\"" Nov 21 14:23:01 crc kubenswrapper[4897]: I1121 14:23:01.763468 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-9fggc" Nov 21 14:23:01 crc kubenswrapper[4897]: I1121 14:23:01.763497 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dvr4v" event={"ID":"ebc2d11c-7430-4ccc-ba13-ad3635215760","Type":"ContainerStarted","Data":"e3fccd58e25908e8e4a818810c89a3e7c7e9269e86904b1103dca5d120e9f2fe"} Nov 21 14:23:01 crc kubenswrapper[4897]: I1121 14:23:01.786666 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-dvr4v" podStartSLOduration=2.275342637 podStartE2EDuration="5.786649915s" podCreationTimestamp="2025-11-21 14:22:56 +0000 UTC" firstStartedPulling="2025-11-21 14:22:57.732419229 +0000 UTC m=+855.017012714" lastFinishedPulling="2025-11-21 14:23:01.243726517 +0000 UTC m=+858.528319992" observedRunningTime="2025-11-21 14:23:01.781758832 +0000 UTC m=+859.066352317" watchObservedRunningTime="2025-11-21 14:23:01.786649915 +0000 UTC m=+859.071243390" Nov 21 14:23:01 crc kubenswrapper[4897]: I1121 14:23:01.837283 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-logging/collector-9fggc"] Nov 21 14:23:01 crc kubenswrapper[4897]: I1121 14:23:01.846330 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-logging/collector-9fggc"] Nov 21 14:23:01 crc kubenswrapper[4897]: I1121 14:23:01.853352 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/collector-h5wk2"] Nov 21 14:23:01 crc kubenswrapper[4897]: I1121 14:23:01.854347 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-h5wk2" Nov 21 14:23:01 crc kubenswrapper[4897]: I1121 14:23:01.858977 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"collector-config" Nov 21 14:23:01 crc kubenswrapper[4897]: I1121 14:23:01.861145 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-token" Nov 21 14:23:01 crc kubenswrapper[4897]: I1121 14:23:01.861219 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-syslog-receiver" Nov 21 14:23:01 crc kubenswrapper[4897]: I1121 14:23:01.861579 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-dockercfg-kqrh2" Nov 21 14:23:01 crc kubenswrapper[4897]: I1121 14:23:01.861711 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-metrics" Nov 21 14:23:01 crc kubenswrapper[4897]: I1121 14:23:01.865126 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"collector-trustbundle" Nov 21 14:23:01 crc kubenswrapper[4897]: I1121 14:23:01.873754 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/collector-h5wk2"] Nov 21 14:23:02 crc kubenswrapper[4897]: I1121 14:23:02.026542 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7e4a6043-b73b-44a2-b7f8-f10ee63da05f-trusted-ca\") pod \"collector-h5wk2\" (UID: \"7e4a6043-b73b-44a2-b7f8-f10ee63da05f\") " pod="openshift-logging/collector-h5wk2" Nov 21 14:23:02 crc kubenswrapper[4897]: I1121 14:23:02.026697 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e4a6043-b73b-44a2-b7f8-f10ee63da05f-config\") pod \"collector-h5wk2\" (UID: \"7e4a6043-b73b-44a2-b7f8-f10ee63da05f\") " pod="openshift-logging/collector-h5wk2" Nov 21 14:23:02 crc kubenswrapper[4897]: I1121 14:23:02.026734 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/7e4a6043-b73b-44a2-b7f8-f10ee63da05f-config-openshift-service-cacrt\") pod \"collector-h5wk2\" (UID: \"7e4a6043-b73b-44a2-b7f8-f10ee63da05f\") " pod="openshift-logging/collector-h5wk2" Nov 21 14:23:02 crc kubenswrapper[4897]: I1121 14:23:02.026789 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/7e4a6043-b73b-44a2-b7f8-f10ee63da05f-collector-syslog-receiver\") pod \"collector-h5wk2\" (UID: \"7e4a6043-b73b-44a2-b7f8-f10ee63da05f\") " pod="openshift-logging/collector-h5wk2" Nov 21 14:23:02 crc kubenswrapper[4897]: I1121 14:23:02.026868 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/7e4a6043-b73b-44a2-b7f8-f10ee63da05f-sa-token\") pod \"collector-h5wk2\" (UID: \"7e4a6043-b73b-44a2-b7f8-f10ee63da05f\") " pod="openshift-logging/collector-h5wk2" Nov 21 14:23:02 crc kubenswrapper[4897]: I1121 14:23:02.026910 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/7e4a6043-b73b-44a2-b7f8-f10ee63da05f-metrics\") pod \"collector-h5wk2\" (UID: \"7e4a6043-b73b-44a2-b7f8-f10ee63da05f\") " pod="openshift-logging/collector-h5wk2" Nov 21 14:23:02 crc kubenswrapper[4897]: I1121 14:23:02.026928 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/7e4a6043-b73b-44a2-b7f8-f10ee63da05f-entrypoint\") pod \"collector-h5wk2\" (UID: \"7e4a6043-b73b-44a2-b7f8-f10ee63da05f\") " pod="openshift-logging/collector-h5wk2" Nov 21 14:23:02 crc kubenswrapper[4897]: I1121 14:23:02.027045 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/7e4a6043-b73b-44a2-b7f8-f10ee63da05f-datadir\") pod \"collector-h5wk2\" (UID: \"7e4a6043-b73b-44a2-b7f8-f10ee63da05f\") " pod="openshift-logging/collector-h5wk2" Nov 21 14:23:02 crc kubenswrapper[4897]: I1121 14:23:02.027081 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/7e4a6043-b73b-44a2-b7f8-f10ee63da05f-collector-token\") pod \"collector-h5wk2\" (UID: \"7e4a6043-b73b-44a2-b7f8-f10ee63da05f\") " pod="openshift-logging/collector-h5wk2" Nov 21 14:23:02 crc kubenswrapper[4897]: I1121 14:23:02.027149 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t8fvj\" (UniqueName: \"kubernetes.io/projected/7e4a6043-b73b-44a2-b7f8-f10ee63da05f-kube-api-access-t8fvj\") pod \"collector-h5wk2\" (UID: \"7e4a6043-b73b-44a2-b7f8-f10ee63da05f\") " pod="openshift-logging/collector-h5wk2" Nov 21 14:23:02 crc kubenswrapper[4897]: I1121 14:23:02.027196 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/7e4a6043-b73b-44a2-b7f8-f10ee63da05f-tmp\") pod \"collector-h5wk2\" (UID: \"7e4a6043-b73b-44a2-b7f8-f10ee63da05f\") " pod="openshift-logging/collector-h5wk2" Nov 21 14:23:02 crc kubenswrapper[4897]: I1121 14:23:02.099724 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="40f12519-bf8d-4792-a4ba-f4d64a3f3b93" path="/var/lib/kubelet/pods/40f12519-bf8d-4792-a4ba-f4d64a3f3b93/volumes" Nov 21 14:23:02 crc kubenswrapper[4897]: I1121 14:23:02.100489 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e92cb91-808f-4981-9b49-6564bfcd408f" path="/var/lib/kubelet/pods/8e92cb91-808f-4981-9b49-6564bfcd408f/volumes" Nov 21 14:23:02 crc kubenswrapper[4897]: I1121 14:23:02.128466 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/7e4a6043-b73b-44a2-b7f8-f10ee63da05f-sa-token\") pod \"collector-h5wk2\" (UID: \"7e4a6043-b73b-44a2-b7f8-f10ee63da05f\") " pod="openshift-logging/collector-h5wk2" Nov 21 14:23:02 crc kubenswrapper[4897]: I1121 14:23:02.128547 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/7e4a6043-b73b-44a2-b7f8-f10ee63da05f-metrics\") pod \"collector-h5wk2\" (UID: \"7e4a6043-b73b-44a2-b7f8-f10ee63da05f\") " pod="openshift-logging/collector-h5wk2" Nov 21 14:23:02 crc kubenswrapper[4897]: I1121 14:23:02.128571 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/7e4a6043-b73b-44a2-b7f8-f10ee63da05f-entrypoint\") pod \"collector-h5wk2\" (UID: \"7e4a6043-b73b-44a2-b7f8-f10ee63da05f\") " pod="openshift-logging/collector-h5wk2" Nov 21 14:23:02 crc kubenswrapper[4897]: I1121 14:23:02.128613 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/7e4a6043-b73b-44a2-b7f8-f10ee63da05f-datadir\") pod \"collector-h5wk2\" (UID: \"7e4a6043-b73b-44a2-b7f8-f10ee63da05f\") " pod="openshift-logging/collector-h5wk2" Nov 21 14:23:02 crc kubenswrapper[4897]: I1121 14:23:02.128638 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/7e4a6043-b73b-44a2-b7f8-f10ee63da05f-collector-token\") pod \"collector-h5wk2\" (UID: \"7e4a6043-b73b-44a2-b7f8-f10ee63da05f\") " pod="openshift-logging/collector-h5wk2" Nov 21 14:23:02 crc kubenswrapper[4897]: I1121 14:23:02.128657 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t8fvj\" (UniqueName: \"kubernetes.io/projected/7e4a6043-b73b-44a2-b7f8-f10ee63da05f-kube-api-access-t8fvj\") pod \"collector-h5wk2\" (UID: \"7e4a6043-b73b-44a2-b7f8-f10ee63da05f\") " pod="openshift-logging/collector-h5wk2" Nov 21 14:23:02 crc kubenswrapper[4897]: I1121 14:23:02.128698 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/7e4a6043-b73b-44a2-b7f8-f10ee63da05f-tmp\") pod \"collector-h5wk2\" (UID: \"7e4a6043-b73b-44a2-b7f8-f10ee63da05f\") " pod="openshift-logging/collector-h5wk2" Nov 21 14:23:02 crc kubenswrapper[4897]: I1121 14:23:02.128717 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7e4a6043-b73b-44a2-b7f8-f10ee63da05f-trusted-ca\") pod \"collector-h5wk2\" (UID: \"7e4a6043-b73b-44a2-b7f8-f10ee63da05f\") " pod="openshift-logging/collector-h5wk2" Nov 21 14:23:02 crc kubenswrapper[4897]: I1121 14:23:02.128747 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e4a6043-b73b-44a2-b7f8-f10ee63da05f-config\") pod \"collector-h5wk2\" (UID: \"7e4a6043-b73b-44a2-b7f8-f10ee63da05f\") " pod="openshift-logging/collector-h5wk2" Nov 21 14:23:02 crc kubenswrapper[4897]: I1121 14:23:02.128778 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/7e4a6043-b73b-44a2-b7f8-f10ee63da05f-config-openshift-service-cacrt\") pod \"collector-h5wk2\" (UID: \"7e4a6043-b73b-44a2-b7f8-f10ee63da05f\") " pod="openshift-logging/collector-h5wk2" Nov 21 14:23:02 crc kubenswrapper[4897]: I1121 14:23:02.128798 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/7e4a6043-b73b-44a2-b7f8-f10ee63da05f-collector-syslog-receiver\") pod \"collector-h5wk2\" (UID: \"7e4a6043-b73b-44a2-b7f8-f10ee63da05f\") " pod="openshift-logging/collector-h5wk2" Nov 21 14:23:02 crc kubenswrapper[4897]: I1121 14:23:02.128799 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/7e4a6043-b73b-44a2-b7f8-f10ee63da05f-datadir\") pod \"collector-h5wk2\" (UID: \"7e4a6043-b73b-44a2-b7f8-f10ee63da05f\") " pod="openshift-logging/collector-h5wk2" Nov 21 14:23:02 crc kubenswrapper[4897]: I1121 14:23:02.129809 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/7e4a6043-b73b-44a2-b7f8-f10ee63da05f-config-openshift-service-cacrt\") pod \"collector-h5wk2\" (UID: \"7e4a6043-b73b-44a2-b7f8-f10ee63da05f\") " pod="openshift-logging/collector-h5wk2" Nov 21 14:23:02 crc kubenswrapper[4897]: I1121 14:23:02.129947 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7e4a6043-b73b-44a2-b7f8-f10ee63da05f-trusted-ca\") pod \"collector-h5wk2\" (UID: \"7e4a6043-b73b-44a2-b7f8-f10ee63da05f\") " pod="openshift-logging/collector-h5wk2" Nov 21 14:23:02 crc kubenswrapper[4897]: I1121 14:23:02.130000 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/7e4a6043-b73b-44a2-b7f8-f10ee63da05f-entrypoint\") pod \"collector-h5wk2\" (UID: \"7e4a6043-b73b-44a2-b7f8-f10ee63da05f\") " pod="openshift-logging/collector-h5wk2" Nov 21 14:23:02 crc kubenswrapper[4897]: I1121 14:23:02.130091 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e4a6043-b73b-44a2-b7f8-f10ee63da05f-config\") pod \"collector-h5wk2\" (UID: \"7e4a6043-b73b-44a2-b7f8-f10ee63da05f\") " pod="openshift-logging/collector-h5wk2" Nov 21 14:23:02 crc kubenswrapper[4897]: I1121 14:23:02.133703 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/7e4a6043-b73b-44a2-b7f8-f10ee63da05f-tmp\") pod \"collector-h5wk2\" (UID: \"7e4a6043-b73b-44a2-b7f8-f10ee63da05f\") " pod="openshift-logging/collector-h5wk2" Nov 21 14:23:02 crc kubenswrapper[4897]: I1121 14:23:02.134016 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/7e4a6043-b73b-44a2-b7f8-f10ee63da05f-metrics\") pod \"collector-h5wk2\" (UID: \"7e4a6043-b73b-44a2-b7f8-f10ee63da05f\") " pod="openshift-logging/collector-h5wk2" Nov 21 14:23:02 crc kubenswrapper[4897]: I1121 14:23:02.134083 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/7e4a6043-b73b-44a2-b7f8-f10ee63da05f-collector-syslog-receiver\") pod \"collector-h5wk2\" (UID: \"7e4a6043-b73b-44a2-b7f8-f10ee63da05f\") " pod="openshift-logging/collector-h5wk2" Nov 21 14:23:02 crc kubenswrapper[4897]: I1121 14:23:02.134558 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/7e4a6043-b73b-44a2-b7f8-f10ee63da05f-collector-token\") pod \"collector-h5wk2\" (UID: \"7e4a6043-b73b-44a2-b7f8-f10ee63da05f\") " pod="openshift-logging/collector-h5wk2" Nov 21 14:23:02 crc kubenswrapper[4897]: I1121 14:23:02.151012 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/7e4a6043-b73b-44a2-b7f8-f10ee63da05f-sa-token\") pod \"collector-h5wk2\" (UID: \"7e4a6043-b73b-44a2-b7f8-f10ee63da05f\") " pod="openshift-logging/collector-h5wk2" Nov 21 14:23:02 crc kubenswrapper[4897]: I1121 14:23:02.151219 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t8fvj\" (UniqueName: \"kubernetes.io/projected/7e4a6043-b73b-44a2-b7f8-f10ee63da05f-kube-api-access-t8fvj\") pod \"collector-h5wk2\" (UID: \"7e4a6043-b73b-44a2-b7f8-f10ee63da05f\") " pod="openshift-logging/collector-h5wk2" Nov 21 14:23:02 crc kubenswrapper[4897]: I1121 14:23:02.174999 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-h5wk2" Nov 21 14:23:02 crc kubenswrapper[4897]: I1121 14:23:02.615129 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/collector-h5wk2"] Nov 21 14:23:02 crc kubenswrapper[4897]: W1121 14:23:02.621956 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7e4a6043_b73b_44a2_b7f8_f10ee63da05f.slice/crio-7c29edb5f010b32302d9745063f019350e7ec558c13743a6ad1674097634386a WatchSource:0}: Error finding container 7c29edb5f010b32302d9745063f019350e7ec558c13743a6ad1674097634386a: Status 404 returned error can't find the container with id 7c29edb5f010b32302d9745063f019350e7ec558c13743a6ad1674097634386a Nov 21 14:23:02 crc kubenswrapper[4897]: I1121 14:23:02.771968 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/collector-h5wk2" event={"ID":"7e4a6043-b73b-44a2-b7f8-f10ee63da05f","Type":"ContainerStarted","Data":"7c29edb5f010b32302d9745063f019350e7ec558c13743a6ad1674097634386a"} Nov 21 14:23:06 crc kubenswrapper[4897]: I1121 14:23:06.361540 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-dvr4v" Nov 21 14:23:06 crc kubenswrapper[4897]: I1121 14:23:06.362839 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-dvr4v" Nov 21 14:23:06 crc kubenswrapper[4897]: I1121 14:23:06.409635 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-dvr4v" Nov 21 14:23:06 crc kubenswrapper[4897]: I1121 14:23:06.852532 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-dvr4v" Nov 21 14:23:07 crc kubenswrapper[4897]: I1121 14:23:07.427269 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dvr4v"] Nov 21 14:23:08 crc kubenswrapper[4897]: I1121 14:23:08.822788 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-dvr4v" podUID="ebc2d11c-7430-4ccc-ba13-ad3635215760" containerName="registry-server" containerID="cri-o://e3fccd58e25908e8e4a818810c89a3e7c7e9269e86904b1103dca5d120e9f2fe" gracePeriod=2 Nov 21 14:23:09 crc kubenswrapper[4897]: I1121 14:23:09.310585 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dvr4v" Nov 21 14:23:09 crc kubenswrapper[4897]: I1121 14:23:09.445138 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ebc2d11c-7430-4ccc-ba13-ad3635215760-catalog-content\") pod \"ebc2d11c-7430-4ccc-ba13-ad3635215760\" (UID: \"ebc2d11c-7430-4ccc-ba13-ad3635215760\") " Nov 21 14:23:09 crc kubenswrapper[4897]: I1121 14:23:09.445194 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ebc2d11c-7430-4ccc-ba13-ad3635215760-utilities\") pod \"ebc2d11c-7430-4ccc-ba13-ad3635215760\" (UID: \"ebc2d11c-7430-4ccc-ba13-ad3635215760\") " Nov 21 14:23:09 crc kubenswrapper[4897]: I1121 14:23:09.445288 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hmm46\" (UniqueName: \"kubernetes.io/projected/ebc2d11c-7430-4ccc-ba13-ad3635215760-kube-api-access-hmm46\") pod \"ebc2d11c-7430-4ccc-ba13-ad3635215760\" (UID: \"ebc2d11c-7430-4ccc-ba13-ad3635215760\") " Nov 21 14:23:09 crc kubenswrapper[4897]: I1121 14:23:09.446084 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ebc2d11c-7430-4ccc-ba13-ad3635215760-utilities" (OuterVolumeSpecName: "utilities") pod "ebc2d11c-7430-4ccc-ba13-ad3635215760" (UID: "ebc2d11c-7430-4ccc-ba13-ad3635215760"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:23:09 crc kubenswrapper[4897]: I1121 14:23:09.452352 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ebc2d11c-7430-4ccc-ba13-ad3635215760-kube-api-access-hmm46" (OuterVolumeSpecName: "kube-api-access-hmm46") pod "ebc2d11c-7430-4ccc-ba13-ad3635215760" (UID: "ebc2d11c-7430-4ccc-ba13-ad3635215760"). InnerVolumeSpecName "kube-api-access-hmm46". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:23:09 crc kubenswrapper[4897]: I1121 14:23:09.467091 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ebc2d11c-7430-4ccc-ba13-ad3635215760-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ebc2d11c-7430-4ccc-ba13-ad3635215760" (UID: "ebc2d11c-7430-4ccc-ba13-ad3635215760"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:23:09 crc kubenswrapper[4897]: I1121 14:23:09.547367 4897 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ebc2d11c-7430-4ccc-ba13-ad3635215760-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 21 14:23:09 crc kubenswrapper[4897]: I1121 14:23:09.547404 4897 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ebc2d11c-7430-4ccc-ba13-ad3635215760-utilities\") on node \"crc\" DevicePath \"\"" Nov 21 14:23:09 crc kubenswrapper[4897]: I1121 14:23:09.547413 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hmm46\" (UniqueName: \"kubernetes.io/projected/ebc2d11c-7430-4ccc-ba13-ad3635215760-kube-api-access-hmm46\") on node \"crc\" DevicePath \"\"" Nov 21 14:23:09 crc kubenswrapper[4897]: I1121 14:23:09.836142 4897 generic.go:334] "Generic (PLEG): container finished" podID="ebc2d11c-7430-4ccc-ba13-ad3635215760" containerID="e3fccd58e25908e8e4a818810c89a3e7c7e9269e86904b1103dca5d120e9f2fe" exitCode=0 Nov 21 14:23:09 crc kubenswrapper[4897]: I1121 14:23:09.836203 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dvr4v" Nov 21 14:23:09 crc kubenswrapper[4897]: I1121 14:23:09.836230 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dvr4v" event={"ID":"ebc2d11c-7430-4ccc-ba13-ad3635215760","Type":"ContainerDied","Data":"e3fccd58e25908e8e4a818810c89a3e7c7e9269e86904b1103dca5d120e9f2fe"} Nov 21 14:23:09 crc kubenswrapper[4897]: I1121 14:23:09.836288 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dvr4v" event={"ID":"ebc2d11c-7430-4ccc-ba13-ad3635215760","Type":"ContainerDied","Data":"2773b77a864672e303dbb1d1ae91cbd40bb1e2f918cb36b14eb217a77307965b"} Nov 21 14:23:09 crc kubenswrapper[4897]: I1121 14:23:09.836313 4897 scope.go:117] "RemoveContainer" containerID="e3fccd58e25908e8e4a818810c89a3e7c7e9269e86904b1103dca5d120e9f2fe" Nov 21 14:23:09 crc kubenswrapper[4897]: I1121 14:23:09.838866 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/collector-h5wk2" event={"ID":"7e4a6043-b73b-44a2-b7f8-f10ee63da05f","Type":"ContainerStarted","Data":"c4bce076e91935fc4b6aa0929ddbd8f5ca62602513febba2a68bbf6321340ea8"} Nov 21 14:23:09 crc kubenswrapper[4897]: I1121 14:23:09.859717 4897 scope.go:117] "RemoveContainer" containerID="b602e769bf7b9dfe3032546f152a21e027f614b425be184a0b55c652035266e4" Nov 21 14:23:09 crc kubenswrapper[4897]: I1121 14:23:09.877945 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/collector-h5wk2" podStartSLOduration=2.723995107 podStartE2EDuration="8.877922571s" podCreationTimestamp="2025-11-21 14:23:01 +0000 UTC" firstStartedPulling="2025-11-21 14:23:02.626066192 +0000 UTC m=+859.910659657" lastFinishedPulling="2025-11-21 14:23:08.779993606 +0000 UTC m=+866.064587121" observedRunningTime="2025-11-21 14:23:09.868914366 +0000 UTC m=+867.153507881" watchObservedRunningTime="2025-11-21 14:23:09.877922571 +0000 UTC m=+867.162516056" Nov 21 14:23:09 crc kubenswrapper[4897]: I1121 14:23:09.878799 4897 scope.go:117] "RemoveContainer" containerID="e406ac4dcc45af4000049ce215d96afdd379d7b41b10e28e6c3afc870adcda27" Nov 21 14:23:09 crc kubenswrapper[4897]: I1121 14:23:09.946810 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dvr4v"] Nov 21 14:23:09 crc kubenswrapper[4897]: I1121 14:23:09.948922 4897 scope.go:117] "RemoveContainer" containerID="e3fccd58e25908e8e4a818810c89a3e7c7e9269e86904b1103dca5d120e9f2fe" Nov 21 14:23:09 crc kubenswrapper[4897]: E1121 14:23:09.949543 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e3fccd58e25908e8e4a818810c89a3e7c7e9269e86904b1103dca5d120e9f2fe\": container with ID starting with e3fccd58e25908e8e4a818810c89a3e7c7e9269e86904b1103dca5d120e9f2fe not found: ID does not exist" containerID="e3fccd58e25908e8e4a818810c89a3e7c7e9269e86904b1103dca5d120e9f2fe" Nov 21 14:23:09 crc kubenswrapper[4897]: I1121 14:23:09.949672 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e3fccd58e25908e8e4a818810c89a3e7c7e9269e86904b1103dca5d120e9f2fe"} err="failed to get container status \"e3fccd58e25908e8e4a818810c89a3e7c7e9269e86904b1103dca5d120e9f2fe\": rpc error: code = NotFound desc = could not find container \"e3fccd58e25908e8e4a818810c89a3e7c7e9269e86904b1103dca5d120e9f2fe\": container with ID starting with e3fccd58e25908e8e4a818810c89a3e7c7e9269e86904b1103dca5d120e9f2fe not found: ID does not exist" Nov 21 14:23:09 crc kubenswrapper[4897]: I1121 14:23:09.949829 4897 scope.go:117] "RemoveContainer" containerID="b602e769bf7b9dfe3032546f152a21e027f614b425be184a0b55c652035266e4" Nov 21 14:23:09 crc kubenswrapper[4897]: E1121 14:23:09.950274 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b602e769bf7b9dfe3032546f152a21e027f614b425be184a0b55c652035266e4\": container with ID starting with b602e769bf7b9dfe3032546f152a21e027f614b425be184a0b55c652035266e4 not found: ID does not exist" containerID="b602e769bf7b9dfe3032546f152a21e027f614b425be184a0b55c652035266e4" Nov 21 14:23:09 crc kubenswrapper[4897]: I1121 14:23:09.950314 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b602e769bf7b9dfe3032546f152a21e027f614b425be184a0b55c652035266e4"} err="failed to get container status \"b602e769bf7b9dfe3032546f152a21e027f614b425be184a0b55c652035266e4\": rpc error: code = NotFound desc = could not find container \"b602e769bf7b9dfe3032546f152a21e027f614b425be184a0b55c652035266e4\": container with ID starting with b602e769bf7b9dfe3032546f152a21e027f614b425be184a0b55c652035266e4 not found: ID does not exist" Nov 21 14:23:09 crc kubenswrapper[4897]: I1121 14:23:09.950344 4897 scope.go:117] "RemoveContainer" containerID="e406ac4dcc45af4000049ce215d96afdd379d7b41b10e28e6c3afc870adcda27" Nov 21 14:23:09 crc kubenswrapper[4897]: E1121 14:23:09.950650 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e406ac4dcc45af4000049ce215d96afdd379d7b41b10e28e6c3afc870adcda27\": container with ID starting with e406ac4dcc45af4000049ce215d96afdd379d7b41b10e28e6c3afc870adcda27 not found: ID does not exist" containerID="e406ac4dcc45af4000049ce215d96afdd379d7b41b10e28e6c3afc870adcda27" Nov 21 14:23:09 crc kubenswrapper[4897]: I1121 14:23:09.950763 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e406ac4dcc45af4000049ce215d96afdd379d7b41b10e28e6c3afc870adcda27"} err="failed to get container status \"e406ac4dcc45af4000049ce215d96afdd379d7b41b10e28e6c3afc870adcda27\": rpc error: code = NotFound desc = could not find container \"e406ac4dcc45af4000049ce215d96afdd379d7b41b10e28e6c3afc870adcda27\": container with ID starting with e406ac4dcc45af4000049ce215d96afdd379d7b41b10e28e6c3afc870adcda27 not found: ID does not exist" Nov 21 14:23:09 crc kubenswrapper[4897]: I1121 14:23:09.954685 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-dvr4v"] Nov 21 14:23:10 crc kubenswrapper[4897]: I1121 14:23:10.099281 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ebc2d11c-7430-4ccc-ba13-ad3635215760" path="/var/lib/kubelet/pods/ebc2d11c-7430-4ccc-ba13-ad3635215760/volumes" Nov 21 14:23:40 crc kubenswrapper[4897]: I1121 14:23:40.258810 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edfc8j"] Nov 21 14:23:40 crc kubenswrapper[4897]: E1121 14:23:40.259701 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebc2d11c-7430-4ccc-ba13-ad3635215760" containerName="registry-server" Nov 21 14:23:40 crc kubenswrapper[4897]: I1121 14:23:40.259719 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebc2d11c-7430-4ccc-ba13-ad3635215760" containerName="registry-server" Nov 21 14:23:40 crc kubenswrapper[4897]: E1121 14:23:40.259749 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebc2d11c-7430-4ccc-ba13-ad3635215760" containerName="extract-utilities" Nov 21 14:23:40 crc kubenswrapper[4897]: I1121 14:23:40.259758 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebc2d11c-7430-4ccc-ba13-ad3635215760" containerName="extract-utilities" Nov 21 14:23:40 crc kubenswrapper[4897]: E1121 14:23:40.259774 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebc2d11c-7430-4ccc-ba13-ad3635215760" containerName="extract-content" Nov 21 14:23:40 crc kubenswrapper[4897]: I1121 14:23:40.259782 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebc2d11c-7430-4ccc-ba13-ad3635215760" containerName="extract-content" Nov 21 14:23:40 crc kubenswrapper[4897]: I1121 14:23:40.259952 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="ebc2d11c-7430-4ccc-ba13-ad3635215760" containerName="registry-server" Nov 21 14:23:40 crc kubenswrapper[4897]: I1121 14:23:40.261174 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edfc8j" Nov 21 14:23:40 crc kubenswrapper[4897]: I1121 14:23:40.271699 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edfc8j"] Nov 21 14:23:40 crc kubenswrapper[4897]: I1121 14:23:40.274015 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 21 14:23:40 crc kubenswrapper[4897]: I1121 14:23:40.454684 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/fddf9293-f983-44ee-a3dc-18e93fda7061-bundle\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edfc8j\" (UID: \"fddf9293-f983-44ee-a3dc-18e93fda7061\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edfc8j" Nov 21 14:23:40 crc kubenswrapper[4897]: I1121 14:23:40.454744 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/fddf9293-f983-44ee-a3dc-18e93fda7061-util\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edfc8j\" (UID: \"fddf9293-f983-44ee-a3dc-18e93fda7061\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edfc8j" Nov 21 14:23:40 crc kubenswrapper[4897]: I1121 14:23:40.454983 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-98x62\" (UniqueName: \"kubernetes.io/projected/fddf9293-f983-44ee-a3dc-18e93fda7061-kube-api-access-98x62\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edfc8j\" (UID: \"fddf9293-f983-44ee-a3dc-18e93fda7061\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edfc8j" Nov 21 14:23:40 crc kubenswrapper[4897]: I1121 14:23:40.556445 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/fddf9293-f983-44ee-a3dc-18e93fda7061-bundle\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edfc8j\" (UID: \"fddf9293-f983-44ee-a3dc-18e93fda7061\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edfc8j" Nov 21 14:23:40 crc kubenswrapper[4897]: I1121 14:23:40.556485 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/fddf9293-f983-44ee-a3dc-18e93fda7061-util\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edfc8j\" (UID: \"fddf9293-f983-44ee-a3dc-18e93fda7061\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edfc8j" Nov 21 14:23:40 crc kubenswrapper[4897]: I1121 14:23:40.556624 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-98x62\" (UniqueName: \"kubernetes.io/projected/fddf9293-f983-44ee-a3dc-18e93fda7061-kube-api-access-98x62\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edfc8j\" (UID: \"fddf9293-f983-44ee-a3dc-18e93fda7061\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edfc8j" Nov 21 14:23:40 crc kubenswrapper[4897]: I1121 14:23:40.556868 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/fddf9293-f983-44ee-a3dc-18e93fda7061-bundle\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edfc8j\" (UID: \"fddf9293-f983-44ee-a3dc-18e93fda7061\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edfc8j" Nov 21 14:23:40 crc kubenswrapper[4897]: I1121 14:23:40.557051 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/fddf9293-f983-44ee-a3dc-18e93fda7061-util\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edfc8j\" (UID: \"fddf9293-f983-44ee-a3dc-18e93fda7061\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edfc8j" Nov 21 14:23:40 crc kubenswrapper[4897]: I1121 14:23:40.586864 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-98x62\" (UniqueName: \"kubernetes.io/projected/fddf9293-f983-44ee-a3dc-18e93fda7061-kube-api-access-98x62\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edfc8j\" (UID: \"fddf9293-f983-44ee-a3dc-18e93fda7061\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edfc8j" Nov 21 14:23:40 crc kubenswrapper[4897]: I1121 14:23:40.885844 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edfc8j" Nov 21 14:23:41 crc kubenswrapper[4897]: I1121 14:23:41.301378 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edfc8j"] Nov 21 14:23:42 crc kubenswrapper[4897]: I1121 14:23:42.124418 4897 generic.go:334] "Generic (PLEG): container finished" podID="fddf9293-f983-44ee-a3dc-18e93fda7061" containerID="a4263830b4d1ece31ccb800a68cd811698eee00e40622cc62138be097eae037c" exitCode=0 Nov 21 14:23:42 crc kubenswrapper[4897]: I1121 14:23:42.124497 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edfc8j" event={"ID":"fddf9293-f983-44ee-a3dc-18e93fda7061","Type":"ContainerDied","Data":"a4263830b4d1ece31ccb800a68cd811698eee00e40622cc62138be097eae037c"} Nov 21 14:23:42 crc kubenswrapper[4897]: I1121 14:23:42.125019 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edfc8j" event={"ID":"fddf9293-f983-44ee-a3dc-18e93fda7061","Type":"ContainerStarted","Data":"1d70a0b72e0d433c8631503329f94fb1d7b4c647457e7b540d816b8e0c232b61"} Nov 21 14:23:44 crc kubenswrapper[4897]: I1121 14:23:44.137617 4897 generic.go:334] "Generic (PLEG): container finished" podID="fddf9293-f983-44ee-a3dc-18e93fda7061" containerID="1f5c7cc73016608802f6ffaa74c91ed847d6e9954225cf5e3a050366448e9a78" exitCode=0 Nov 21 14:23:44 crc kubenswrapper[4897]: I1121 14:23:44.138140 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edfc8j" event={"ID":"fddf9293-f983-44ee-a3dc-18e93fda7061","Type":"ContainerDied","Data":"1f5c7cc73016608802f6ffaa74c91ed847d6e9954225cf5e3a050366448e9a78"} Nov 21 14:23:45 crc kubenswrapper[4897]: I1121 14:23:45.153629 4897 generic.go:334] "Generic (PLEG): container finished" podID="fddf9293-f983-44ee-a3dc-18e93fda7061" containerID="de961f51386c78f921e942c16084c6e050d83051cc64bbd8061f58961faf212c" exitCode=0 Nov 21 14:23:45 crc kubenswrapper[4897]: I1121 14:23:45.153714 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edfc8j" event={"ID":"fddf9293-f983-44ee-a3dc-18e93fda7061","Type":"ContainerDied","Data":"de961f51386c78f921e942c16084c6e050d83051cc64bbd8061f58961faf212c"} Nov 21 14:23:46 crc kubenswrapper[4897]: I1121 14:23:46.456190 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edfc8j" Nov 21 14:23:46 crc kubenswrapper[4897]: I1121 14:23:46.544520 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-98x62\" (UniqueName: \"kubernetes.io/projected/fddf9293-f983-44ee-a3dc-18e93fda7061-kube-api-access-98x62\") pod \"fddf9293-f983-44ee-a3dc-18e93fda7061\" (UID: \"fddf9293-f983-44ee-a3dc-18e93fda7061\") " Nov 21 14:23:46 crc kubenswrapper[4897]: I1121 14:23:46.544617 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/fddf9293-f983-44ee-a3dc-18e93fda7061-util\") pod \"fddf9293-f983-44ee-a3dc-18e93fda7061\" (UID: \"fddf9293-f983-44ee-a3dc-18e93fda7061\") " Nov 21 14:23:46 crc kubenswrapper[4897]: I1121 14:23:46.544808 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/fddf9293-f983-44ee-a3dc-18e93fda7061-bundle\") pod \"fddf9293-f983-44ee-a3dc-18e93fda7061\" (UID: \"fddf9293-f983-44ee-a3dc-18e93fda7061\") " Nov 21 14:23:46 crc kubenswrapper[4897]: I1121 14:23:46.545687 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fddf9293-f983-44ee-a3dc-18e93fda7061-bundle" (OuterVolumeSpecName: "bundle") pod "fddf9293-f983-44ee-a3dc-18e93fda7061" (UID: "fddf9293-f983-44ee-a3dc-18e93fda7061"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:23:46 crc kubenswrapper[4897]: I1121 14:23:46.551019 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fddf9293-f983-44ee-a3dc-18e93fda7061-kube-api-access-98x62" (OuterVolumeSpecName: "kube-api-access-98x62") pod "fddf9293-f983-44ee-a3dc-18e93fda7061" (UID: "fddf9293-f983-44ee-a3dc-18e93fda7061"). InnerVolumeSpecName "kube-api-access-98x62". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:23:46 crc kubenswrapper[4897]: I1121 14:23:46.577343 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fddf9293-f983-44ee-a3dc-18e93fda7061-util" (OuterVolumeSpecName: "util") pod "fddf9293-f983-44ee-a3dc-18e93fda7061" (UID: "fddf9293-f983-44ee-a3dc-18e93fda7061"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:23:46 crc kubenswrapper[4897]: I1121 14:23:46.646373 4897 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/fddf9293-f983-44ee-a3dc-18e93fda7061-bundle\") on node \"crc\" DevicePath \"\"" Nov 21 14:23:46 crc kubenswrapper[4897]: I1121 14:23:46.646412 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-98x62\" (UniqueName: \"kubernetes.io/projected/fddf9293-f983-44ee-a3dc-18e93fda7061-kube-api-access-98x62\") on node \"crc\" DevicePath \"\"" Nov 21 14:23:46 crc kubenswrapper[4897]: I1121 14:23:46.646423 4897 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/fddf9293-f983-44ee-a3dc-18e93fda7061-util\") on node \"crc\" DevicePath \"\"" Nov 21 14:23:47 crc kubenswrapper[4897]: I1121 14:23:47.171937 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edfc8j" event={"ID":"fddf9293-f983-44ee-a3dc-18e93fda7061","Type":"ContainerDied","Data":"1d70a0b72e0d433c8631503329f94fb1d7b4c647457e7b540d816b8e0c232b61"} Nov 21 14:23:47 crc kubenswrapper[4897]: I1121 14:23:47.172170 4897 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1d70a0b72e0d433c8631503329f94fb1d7b4c647457e7b540d816b8e0c232b61" Nov 21 14:23:47 crc kubenswrapper[4897]: I1121 14:23:47.172068 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772edfc8j" Nov 21 14:23:52 crc kubenswrapper[4897]: I1121 14:23:52.241551 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-557fdffb88-6q8p4"] Nov 21 14:23:52 crc kubenswrapper[4897]: E1121 14:23:52.242488 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fddf9293-f983-44ee-a3dc-18e93fda7061" containerName="extract" Nov 21 14:23:52 crc kubenswrapper[4897]: I1121 14:23:52.242533 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="fddf9293-f983-44ee-a3dc-18e93fda7061" containerName="extract" Nov 21 14:23:52 crc kubenswrapper[4897]: E1121 14:23:52.242552 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fddf9293-f983-44ee-a3dc-18e93fda7061" containerName="util" Nov 21 14:23:52 crc kubenswrapper[4897]: I1121 14:23:52.242558 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="fddf9293-f983-44ee-a3dc-18e93fda7061" containerName="util" Nov 21 14:23:52 crc kubenswrapper[4897]: E1121 14:23:52.242576 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fddf9293-f983-44ee-a3dc-18e93fda7061" containerName="pull" Nov 21 14:23:52 crc kubenswrapper[4897]: I1121 14:23:52.242582 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="fddf9293-f983-44ee-a3dc-18e93fda7061" containerName="pull" Nov 21 14:23:52 crc kubenswrapper[4897]: I1121 14:23:52.242859 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="fddf9293-f983-44ee-a3dc-18e93fda7061" containerName="extract" Nov 21 14:23:52 crc kubenswrapper[4897]: I1121 14:23:52.243659 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-557fdffb88-6q8p4" Nov 21 14:23:52 crc kubenswrapper[4897]: I1121 14:23:52.246273 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Nov 21 14:23:52 crc kubenswrapper[4897]: I1121 14:23:52.246703 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-mfz56" Nov 21 14:23:52 crc kubenswrapper[4897]: I1121 14:23:52.246492 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Nov 21 14:23:52 crc kubenswrapper[4897]: I1121 14:23:52.264802 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-557fdffb88-6q8p4"] Nov 21 14:23:52 crc kubenswrapper[4897]: I1121 14:23:52.345120 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s7dc6\" (UniqueName: \"kubernetes.io/projected/693cda1f-685f-4ee0-bde2-260fed47d852-kube-api-access-s7dc6\") pod \"nmstate-operator-557fdffb88-6q8p4\" (UID: \"693cda1f-685f-4ee0-bde2-260fed47d852\") " pod="openshift-nmstate/nmstate-operator-557fdffb88-6q8p4" Nov 21 14:23:52 crc kubenswrapper[4897]: I1121 14:23:52.446530 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s7dc6\" (UniqueName: \"kubernetes.io/projected/693cda1f-685f-4ee0-bde2-260fed47d852-kube-api-access-s7dc6\") pod \"nmstate-operator-557fdffb88-6q8p4\" (UID: \"693cda1f-685f-4ee0-bde2-260fed47d852\") " pod="openshift-nmstate/nmstate-operator-557fdffb88-6q8p4" Nov 21 14:23:52 crc kubenswrapper[4897]: I1121 14:23:52.466298 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s7dc6\" (UniqueName: \"kubernetes.io/projected/693cda1f-685f-4ee0-bde2-260fed47d852-kube-api-access-s7dc6\") pod \"nmstate-operator-557fdffb88-6q8p4\" (UID: \"693cda1f-685f-4ee0-bde2-260fed47d852\") " pod="openshift-nmstate/nmstate-operator-557fdffb88-6q8p4" Nov 21 14:23:52 crc kubenswrapper[4897]: I1121 14:23:52.593667 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-557fdffb88-6q8p4" Nov 21 14:23:53 crc kubenswrapper[4897]: I1121 14:23:53.019100 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-557fdffb88-6q8p4"] Nov 21 14:23:53 crc kubenswrapper[4897]: I1121 14:23:53.219314 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-557fdffb88-6q8p4" event={"ID":"693cda1f-685f-4ee0-bde2-260fed47d852","Type":"ContainerStarted","Data":"15d808557b884cf5468d37398d78ce8931a3ef43b6e0c3d4f20feffe1f28ae8a"} Nov 21 14:23:57 crc kubenswrapper[4897]: I1121 14:23:57.246832 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-557fdffb88-6q8p4" event={"ID":"693cda1f-685f-4ee0-bde2-260fed47d852","Type":"ContainerStarted","Data":"446650e4d552765b6c66ab53091092f09241310eb96ddbd6324553a3facabfc6"} Nov 21 14:23:57 crc kubenswrapper[4897]: I1121 14:23:57.267390 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-557fdffb88-6q8p4" podStartSLOduration=2.17197495 podStartE2EDuration="5.267371248s" podCreationTimestamp="2025-11-21 14:23:52 +0000 UTC" firstStartedPulling="2025-11-21 14:23:53.031384057 +0000 UTC m=+910.315977532" lastFinishedPulling="2025-11-21 14:23:56.126780345 +0000 UTC m=+913.411373830" observedRunningTime="2025-11-21 14:23:57.263764349 +0000 UTC m=+914.548357824" watchObservedRunningTime="2025-11-21 14:23:57.267371248 +0000 UTC m=+914.551964723" Nov 21 14:24:01 crc kubenswrapper[4897]: I1121 14:24:01.470436 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-5dcf9c57c5-wmjz5"] Nov 21 14:24:01 crc kubenswrapper[4897]: I1121 14:24:01.472215 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-wmjz5" Nov 21 14:24:01 crc kubenswrapper[4897]: I1121 14:24:01.475000 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-qfbvt" Nov 21 14:24:01 crc kubenswrapper[4897]: I1121 14:24:01.478993 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-6b89b748d8-lp4lc"] Nov 21 14:24:01 crc kubenswrapper[4897]: I1121 14:24:01.479881 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-lp4lc" Nov 21 14:24:01 crc kubenswrapper[4897]: I1121 14:24:01.483621 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Nov 21 14:24:01 crc kubenswrapper[4897]: I1121 14:24:01.485868 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5k96q\" (UniqueName: \"kubernetes.io/projected/e41a325d-405a-4d0d-bf21-c8dbd7abcd14-kube-api-access-5k96q\") pod \"nmstate-webhook-6b89b748d8-lp4lc\" (UID: \"e41a325d-405a-4d0d-bf21-c8dbd7abcd14\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-lp4lc" Nov 21 14:24:01 crc kubenswrapper[4897]: I1121 14:24:01.485957 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/e41a325d-405a-4d0d-bf21-c8dbd7abcd14-tls-key-pair\") pod \"nmstate-webhook-6b89b748d8-lp4lc\" (UID: \"e41a325d-405a-4d0d-bf21-c8dbd7abcd14\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-lp4lc" Nov 21 14:24:01 crc kubenswrapper[4897]: I1121 14:24:01.485994 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dblzw\" (UniqueName: \"kubernetes.io/projected/cef2e7b5-90a4-4bd2-b8bc-e0513aa09e6d-kube-api-access-dblzw\") pod \"nmstate-metrics-5dcf9c57c5-wmjz5\" (UID: \"cef2e7b5-90a4-4bd2-b8bc-e0513aa09e6d\") " pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-wmjz5" Nov 21 14:24:01 crc kubenswrapper[4897]: I1121 14:24:01.489806 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-5dcf9c57c5-wmjz5"] Nov 21 14:24:01 crc kubenswrapper[4897]: I1121 14:24:01.501743 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6b89b748d8-lp4lc"] Nov 21 14:24:01 crc kubenswrapper[4897]: I1121 14:24:01.538845 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-896b2"] Nov 21 14:24:01 crc kubenswrapper[4897]: I1121 14:24:01.539738 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-896b2" Nov 21 14:24:01 crc kubenswrapper[4897]: I1121 14:24:01.589305 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/e41a325d-405a-4d0d-bf21-c8dbd7abcd14-tls-key-pair\") pod \"nmstate-webhook-6b89b748d8-lp4lc\" (UID: \"e41a325d-405a-4d0d-bf21-c8dbd7abcd14\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-lp4lc" Nov 21 14:24:01 crc kubenswrapper[4897]: I1121 14:24:01.589580 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dblzw\" (UniqueName: \"kubernetes.io/projected/cef2e7b5-90a4-4bd2-b8bc-e0513aa09e6d-kube-api-access-dblzw\") pod \"nmstate-metrics-5dcf9c57c5-wmjz5\" (UID: \"cef2e7b5-90a4-4bd2-b8bc-e0513aa09e6d\") " pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-wmjz5" Nov 21 14:24:01 crc kubenswrapper[4897]: I1121 14:24:01.589712 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fll2p\" (UniqueName: \"kubernetes.io/projected/7594e6c8-4476-4ceb-b518-500f0f93b3e3-kube-api-access-fll2p\") pod \"nmstate-handler-896b2\" (UID: \"7594e6c8-4476-4ceb-b518-500f0f93b3e3\") " pod="openshift-nmstate/nmstate-handler-896b2" Nov 21 14:24:01 crc kubenswrapper[4897]: I1121 14:24:01.589790 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/7594e6c8-4476-4ceb-b518-500f0f93b3e3-dbus-socket\") pod \"nmstate-handler-896b2\" (UID: \"7594e6c8-4476-4ceb-b518-500f0f93b3e3\") " pod="openshift-nmstate/nmstate-handler-896b2" Nov 21 14:24:01 crc kubenswrapper[4897]: I1121 14:24:01.589871 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/7594e6c8-4476-4ceb-b518-500f0f93b3e3-ovs-socket\") pod \"nmstate-handler-896b2\" (UID: \"7594e6c8-4476-4ceb-b518-500f0f93b3e3\") " pod="openshift-nmstate/nmstate-handler-896b2" Nov 21 14:24:01 crc kubenswrapper[4897]: I1121 14:24:01.590092 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/7594e6c8-4476-4ceb-b518-500f0f93b3e3-nmstate-lock\") pod \"nmstate-handler-896b2\" (UID: \"7594e6c8-4476-4ceb-b518-500f0f93b3e3\") " pod="openshift-nmstate/nmstate-handler-896b2" Nov 21 14:24:01 crc kubenswrapper[4897]: I1121 14:24:01.590173 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5k96q\" (UniqueName: \"kubernetes.io/projected/e41a325d-405a-4d0d-bf21-c8dbd7abcd14-kube-api-access-5k96q\") pod \"nmstate-webhook-6b89b748d8-lp4lc\" (UID: \"e41a325d-405a-4d0d-bf21-c8dbd7abcd14\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-lp4lc" Nov 21 14:24:01 crc kubenswrapper[4897]: E1121 14:24:01.589475 4897 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Nov 21 14:24:01 crc kubenswrapper[4897]: E1121 14:24:01.590653 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e41a325d-405a-4d0d-bf21-c8dbd7abcd14-tls-key-pair podName:e41a325d-405a-4d0d-bf21-c8dbd7abcd14 nodeName:}" failed. No retries permitted until 2025-11-21 14:24:02.090633426 +0000 UTC m=+919.375226901 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/e41a325d-405a-4d0d-bf21-c8dbd7abcd14-tls-key-pair") pod "nmstate-webhook-6b89b748d8-lp4lc" (UID: "e41a325d-405a-4d0d-bf21-c8dbd7abcd14") : secret "openshift-nmstate-webhook" not found Nov 21 14:24:01 crc kubenswrapper[4897]: I1121 14:24:01.610764 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5k96q\" (UniqueName: \"kubernetes.io/projected/e41a325d-405a-4d0d-bf21-c8dbd7abcd14-kube-api-access-5k96q\") pod \"nmstate-webhook-6b89b748d8-lp4lc\" (UID: \"e41a325d-405a-4d0d-bf21-c8dbd7abcd14\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-lp4lc" Nov 21 14:24:01 crc kubenswrapper[4897]: I1121 14:24:01.611535 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dblzw\" (UniqueName: \"kubernetes.io/projected/cef2e7b5-90a4-4bd2-b8bc-e0513aa09e6d-kube-api-access-dblzw\") pod \"nmstate-metrics-5dcf9c57c5-wmjz5\" (UID: \"cef2e7b5-90a4-4bd2-b8bc-e0513aa09e6d\") " pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-wmjz5" Nov 21 14:24:01 crc kubenswrapper[4897]: I1121 14:24:01.645567 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5874bd7bc5-6tvtd"] Nov 21 14:24:01 crc kubenswrapper[4897]: I1121 14:24:01.646948 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-6tvtd" Nov 21 14:24:01 crc kubenswrapper[4897]: I1121 14:24:01.651908 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Nov 21 14:24:01 crc kubenswrapper[4897]: I1121 14:24:01.651985 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-lnnb2" Nov 21 14:24:01 crc kubenswrapper[4897]: I1121 14:24:01.656607 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Nov 21 14:24:01 crc kubenswrapper[4897]: I1121 14:24:01.660661 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5874bd7bc5-6tvtd"] Nov 21 14:24:01 crc kubenswrapper[4897]: I1121 14:24:01.691250 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/3eb1faee-c7f5-4910-9382-46f6e5a7c5f7-nginx-conf\") pod \"nmstate-console-plugin-5874bd7bc5-6tvtd\" (UID: \"3eb1faee-c7f5-4910-9382-46f6e5a7c5f7\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-6tvtd" Nov 21 14:24:01 crc kubenswrapper[4897]: I1121 14:24:01.691302 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fll2p\" (UniqueName: \"kubernetes.io/projected/7594e6c8-4476-4ceb-b518-500f0f93b3e3-kube-api-access-fll2p\") pod \"nmstate-handler-896b2\" (UID: \"7594e6c8-4476-4ceb-b518-500f0f93b3e3\") " pod="openshift-nmstate/nmstate-handler-896b2" Nov 21 14:24:01 crc kubenswrapper[4897]: I1121 14:24:01.691332 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/7594e6c8-4476-4ceb-b518-500f0f93b3e3-dbus-socket\") pod \"nmstate-handler-896b2\" (UID: \"7594e6c8-4476-4ceb-b518-500f0f93b3e3\") " pod="openshift-nmstate/nmstate-handler-896b2" Nov 21 14:24:01 crc kubenswrapper[4897]: I1121 14:24:01.691359 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/7594e6c8-4476-4ceb-b518-500f0f93b3e3-ovs-socket\") pod \"nmstate-handler-896b2\" (UID: \"7594e6c8-4476-4ceb-b518-500f0f93b3e3\") " pod="openshift-nmstate/nmstate-handler-896b2" Nov 21 14:24:01 crc kubenswrapper[4897]: I1121 14:24:01.691397 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/3eb1faee-c7f5-4910-9382-46f6e5a7c5f7-plugin-serving-cert\") pod \"nmstate-console-plugin-5874bd7bc5-6tvtd\" (UID: \"3eb1faee-c7f5-4910-9382-46f6e5a7c5f7\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-6tvtd" Nov 21 14:24:01 crc kubenswrapper[4897]: I1121 14:24:01.691433 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qvkjn\" (UniqueName: \"kubernetes.io/projected/3eb1faee-c7f5-4910-9382-46f6e5a7c5f7-kube-api-access-qvkjn\") pod \"nmstate-console-plugin-5874bd7bc5-6tvtd\" (UID: \"3eb1faee-c7f5-4910-9382-46f6e5a7c5f7\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-6tvtd" Nov 21 14:24:01 crc kubenswrapper[4897]: I1121 14:24:01.691493 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/7594e6c8-4476-4ceb-b518-500f0f93b3e3-nmstate-lock\") pod \"nmstate-handler-896b2\" (UID: \"7594e6c8-4476-4ceb-b518-500f0f93b3e3\") " pod="openshift-nmstate/nmstate-handler-896b2" Nov 21 14:24:01 crc kubenswrapper[4897]: I1121 14:24:01.692124 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/7594e6c8-4476-4ceb-b518-500f0f93b3e3-nmstate-lock\") pod \"nmstate-handler-896b2\" (UID: \"7594e6c8-4476-4ceb-b518-500f0f93b3e3\") " pod="openshift-nmstate/nmstate-handler-896b2" Nov 21 14:24:01 crc kubenswrapper[4897]: I1121 14:24:01.693090 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/7594e6c8-4476-4ceb-b518-500f0f93b3e3-dbus-socket\") pod \"nmstate-handler-896b2\" (UID: \"7594e6c8-4476-4ceb-b518-500f0f93b3e3\") " pod="openshift-nmstate/nmstate-handler-896b2" Nov 21 14:24:01 crc kubenswrapper[4897]: I1121 14:24:01.693141 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/7594e6c8-4476-4ceb-b518-500f0f93b3e3-ovs-socket\") pod \"nmstate-handler-896b2\" (UID: \"7594e6c8-4476-4ceb-b518-500f0f93b3e3\") " pod="openshift-nmstate/nmstate-handler-896b2" Nov 21 14:24:01 crc kubenswrapper[4897]: I1121 14:24:01.719071 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fll2p\" (UniqueName: \"kubernetes.io/projected/7594e6c8-4476-4ceb-b518-500f0f93b3e3-kube-api-access-fll2p\") pod \"nmstate-handler-896b2\" (UID: \"7594e6c8-4476-4ceb-b518-500f0f93b3e3\") " pod="openshift-nmstate/nmstate-handler-896b2" Nov 21 14:24:01 crc kubenswrapper[4897]: I1121 14:24:01.792444 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/3eb1faee-c7f5-4910-9382-46f6e5a7c5f7-plugin-serving-cert\") pod \"nmstate-console-plugin-5874bd7bc5-6tvtd\" (UID: \"3eb1faee-c7f5-4910-9382-46f6e5a7c5f7\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-6tvtd" Nov 21 14:24:01 crc kubenswrapper[4897]: I1121 14:24:01.792496 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qvkjn\" (UniqueName: \"kubernetes.io/projected/3eb1faee-c7f5-4910-9382-46f6e5a7c5f7-kube-api-access-qvkjn\") pod \"nmstate-console-plugin-5874bd7bc5-6tvtd\" (UID: \"3eb1faee-c7f5-4910-9382-46f6e5a7c5f7\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-6tvtd" Nov 21 14:24:01 crc kubenswrapper[4897]: I1121 14:24:01.792607 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/3eb1faee-c7f5-4910-9382-46f6e5a7c5f7-nginx-conf\") pod \"nmstate-console-plugin-5874bd7bc5-6tvtd\" (UID: \"3eb1faee-c7f5-4910-9382-46f6e5a7c5f7\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-6tvtd" Nov 21 14:24:01 crc kubenswrapper[4897]: E1121 14:24:01.792634 4897 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Nov 21 14:24:01 crc kubenswrapper[4897]: E1121 14:24:01.792704 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3eb1faee-c7f5-4910-9382-46f6e5a7c5f7-plugin-serving-cert podName:3eb1faee-c7f5-4910-9382-46f6e5a7c5f7 nodeName:}" failed. No retries permitted until 2025-11-21 14:24:02.292687031 +0000 UTC m=+919.577280506 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/3eb1faee-c7f5-4910-9382-46f6e5a7c5f7-plugin-serving-cert") pod "nmstate-console-plugin-5874bd7bc5-6tvtd" (UID: "3eb1faee-c7f5-4910-9382-46f6e5a7c5f7") : secret "plugin-serving-cert" not found Nov 21 14:24:01 crc kubenswrapper[4897]: I1121 14:24:01.793523 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/3eb1faee-c7f5-4910-9382-46f6e5a7c5f7-nginx-conf\") pod \"nmstate-console-plugin-5874bd7bc5-6tvtd\" (UID: \"3eb1faee-c7f5-4910-9382-46f6e5a7c5f7\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-6tvtd" Nov 21 14:24:01 crc kubenswrapper[4897]: I1121 14:24:01.802706 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-wmjz5" Nov 21 14:24:01 crc kubenswrapper[4897]: I1121 14:24:01.816863 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qvkjn\" (UniqueName: \"kubernetes.io/projected/3eb1faee-c7f5-4910-9382-46f6e5a7c5f7-kube-api-access-qvkjn\") pod \"nmstate-console-plugin-5874bd7bc5-6tvtd\" (UID: \"3eb1faee-c7f5-4910-9382-46f6e5a7c5f7\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-6tvtd" Nov 21 14:24:01 crc kubenswrapper[4897]: I1121 14:24:01.827184 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f58d5cd65-2pc2p"] Nov 21 14:24:01 crc kubenswrapper[4897]: I1121 14:24:01.832625 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f58d5cd65-2pc2p" Nov 21 14:24:01 crc kubenswrapper[4897]: I1121 14:24:01.838732 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f58d5cd65-2pc2p"] Nov 21 14:24:01 crc kubenswrapper[4897]: I1121 14:24:01.861838 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-896b2" Nov 21 14:24:01 crc kubenswrapper[4897]: I1121 14:24:01.999852 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ab48e37a-819f-4f63-8e67-9a2cbeaaeb82-service-ca\") pod \"console-f58d5cd65-2pc2p\" (UID: \"ab48e37a-819f-4f63-8e67-9a2cbeaaeb82\") " pod="openshift-console/console-f58d5cd65-2pc2p" Nov 21 14:24:01 crc kubenswrapper[4897]: I1121 14:24:01.999928 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ab48e37a-819f-4f63-8e67-9a2cbeaaeb82-trusted-ca-bundle\") pod \"console-f58d5cd65-2pc2p\" (UID: \"ab48e37a-819f-4f63-8e67-9a2cbeaaeb82\") " pod="openshift-console/console-f58d5cd65-2pc2p" Nov 21 14:24:02 crc kubenswrapper[4897]: I1121 14:24:01.999994 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ab48e37a-819f-4f63-8e67-9a2cbeaaeb82-console-oauth-config\") pod \"console-f58d5cd65-2pc2p\" (UID: \"ab48e37a-819f-4f63-8e67-9a2cbeaaeb82\") " pod="openshift-console/console-f58d5cd65-2pc2p" Nov 21 14:24:02 crc kubenswrapper[4897]: I1121 14:24:02.000037 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xdtcg\" (UniqueName: \"kubernetes.io/projected/ab48e37a-819f-4f63-8e67-9a2cbeaaeb82-kube-api-access-xdtcg\") pod \"console-f58d5cd65-2pc2p\" (UID: \"ab48e37a-819f-4f63-8e67-9a2cbeaaeb82\") " pod="openshift-console/console-f58d5cd65-2pc2p" Nov 21 14:24:02 crc kubenswrapper[4897]: I1121 14:24:02.000064 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ab48e37a-819f-4f63-8e67-9a2cbeaaeb82-oauth-serving-cert\") pod \"console-f58d5cd65-2pc2p\" (UID: \"ab48e37a-819f-4f63-8e67-9a2cbeaaeb82\") " pod="openshift-console/console-f58d5cd65-2pc2p" Nov 21 14:24:02 crc kubenswrapper[4897]: I1121 14:24:02.000123 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ab48e37a-819f-4f63-8e67-9a2cbeaaeb82-console-serving-cert\") pod \"console-f58d5cd65-2pc2p\" (UID: \"ab48e37a-819f-4f63-8e67-9a2cbeaaeb82\") " pod="openshift-console/console-f58d5cd65-2pc2p" Nov 21 14:24:02 crc kubenswrapper[4897]: I1121 14:24:02.000152 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ab48e37a-819f-4f63-8e67-9a2cbeaaeb82-console-config\") pod \"console-f58d5cd65-2pc2p\" (UID: \"ab48e37a-819f-4f63-8e67-9a2cbeaaeb82\") " pod="openshift-console/console-f58d5cd65-2pc2p" Nov 21 14:24:02 crc kubenswrapper[4897]: I1121 14:24:02.102176 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ab48e37a-819f-4f63-8e67-9a2cbeaaeb82-trusted-ca-bundle\") pod \"console-f58d5cd65-2pc2p\" (UID: \"ab48e37a-819f-4f63-8e67-9a2cbeaaeb82\") " pod="openshift-console/console-f58d5cd65-2pc2p" Nov 21 14:24:02 crc kubenswrapper[4897]: I1121 14:24:02.102271 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ab48e37a-819f-4f63-8e67-9a2cbeaaeb82-console-oauth-config\") pod \"console-f58d5cd65-2pc2p\" (UID: \"ab48e37a-819f-4f63-8e67-9a2cbeaaeb82\") " pod="openshift-console/console-f58d5cd65-2pc2p" Nov 21 14:24:02 crc kubenswrapper[4897]: I1121 14:24:02.103487 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ab48e37a-819f-4f63-8e67-9a2cbeaaeb82-trusted-ca-bundle\") pod \"console-f58d5cd65-2pc2p\" (UID: \"ab48e37a-819f-4f63-8e67-9a2cbeaaeb82\") " pod="openshift-console/console-f58d5cd65-2pc2p" Nov 21 14:24:02 crc kubenswrapper[4897]: I1121 14:24:02.103563 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/e41a325d-405a-4d0d-bf21-c8dbd7abcd14-tls-key-pair\") pod \"nmstate-webhook-6b89b748d8-lp4lc\" (UID: \"e41a325d-405a-4d0d-bf21-c8dbd7abcd14\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-lp4lc" Nov 21 14:24:02 crc kubenswrapper[4897]: I1121 14:24:02.104096 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xdtcg\" (UniqueName: \"kubernetes.io/projected/ab48e37a-819f-4f63-8e67-9a2cbeaaeb82-kube-api-access-xdtcg\") pod \"console-f58d5cd65-2pc2p\" (UID: \"ab48e37a-819f-4f63-8e67-9a2cbeaaeb82\") " pod="openshift-console/console-f58d5cd65-2pc2p" Nov 21 14:24:02 crc kubenswrapper[4897]: I1121 14:24:02.104149 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ab48e37a-819f-4f63-8e67-9a2cbeaaeb82-oauth-serving-cert\") pod \"console-f58d5cd65-2pc2p\" (UID: \"ab48e37a-819f-4f63-8e67-9a2cbeaaeb82\") " pod="openshift-console/console-f58d5cd65-2pc2p" Nov 21 14:24:02 crc kubenswrapper[4897]: I1121 14:24:02.104197 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ab48e37a-819f-4f63-8e67-9a2cbeaaeb82-console-serving-cert\") pod \"console-f58d5cd65-2pc2p\" (UID: \"ab48e37a-819f-4f63-8e67-9a2cbeaaeb82\") " pod="openshift-console/console-f58d5cd65-2pc2p" Nov 21 14:24:02 crc kubenswrapper[4897]: I1121 14:24:02.104228 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ab48e37a-819f-4f63-8e67-9a2cbeaaeb82-console-config\") pod \"console-f58d5cd65-2pc2p\" (UID: \"ab48e37a-819f-4f63-8e67-9a2cbeaaeb82\") " pod="openshift-console/console-f58d5cd65-2pc2p" Nov 21 14:24:02 crc kubenswrapper[4897]: I1121 14:24:02.104269 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ab48e37a-819f-4f63-8e67-9a2cbeaaeb82-service-ca\") pod \"console-f58d5cd65-2pc2p\" (UID: \"ab48e37a-819f-4f63-8e67-9a2cbeaaeb82\") " pod="openshift-console/console-f58d5cd65-2pc2p" Nov 21 14:24:02 crc kubenswrapper[4897]: I1121 14:24:02.104903 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ab48e37a-819f-4f63-8e67-9a2cbeaaeb82-console-config\") pod \"console-f58d5cd65-2pc2p\" (UID: \"ab48e37a-819f-4f63-8e67-9a2cbeaaeb82\") " pod="openshift-console/console-f58d5cd65-2pc2p" Nov 21 14:24:02 crc kubenswrapper[4897]: I1121 14:24:02.104971 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ab48e37a-819f-4f63-8e67-9a2cbeaaeb82-service-ca\") pod \"console-f58d5cd65-2pc2p\" (UID: \"ab48e37a-819f-4f63-8e67-9a2cbeaaeb82\") " pod="openshift-console/console-f58d5cd65-2pc2p" Nov 21 14:24:02 crc kubenswrapper[4897]: I1121 14:24:02.105457 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ab48e37a-819f-4f63-8e67-9a2cbeaaeb82-oauth-serving-cert\") pod \"console-f58d5cd65-2pc2p\" (UID: \"ab48e37a-819f-4f63-8e67-9a2cbeaaeb82\") " pod="openshift-console/console-f58d5cd65-2pc2p" Nov 21 14:24:02 crc kubenswrapper[4897]: I1121 14:24:02.105849 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ab48e37a-819f-4f63-8e67-9a2cbeaaeb82-console-oauth-config\") pod \"console-f58d5cd65-2pc2p\" (UID: \"ab48e37a-819f-4f63-8e67-9a2cbeaaeb82\") " pod="openshift-console/console-f58d5cd65-2pc2p" Nov 21 14:24:02 crc kubenswrapper[4897]: I1121 14:24:02.108326 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/e41a325d-405a-4d0d-bf21-c8dbd7abcd14-tls-key-pair\") pod \"nmstate-webhook-6b89b748d8-lp4lc\" (UID: \"e41a325d-405a-4d0d-bf21-c8dbd7abcd14\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-lp4lc" Nov 21 14:24:02 crc kubenswrapper[4897]: I1121 14:24:02.109156 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ab48e37a-819f-4f63-8e67-9a2cbeaaeb82-console-serving-cert\") pod \"console-f58d5cd65-2pc2p\" (UID: \"ab48e37a-819f-4f63-8e67-9a2cbeaaeb82\") " pod="openshift-console/console-f58d5cd65-2pc2p" Nov 21 14:24:02 crc kubenswrapper[4897]: I1121 14:24:02.115579 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-lp4lc" Nov 21 14:24:02 crc kubenswrapper[4897]: I1121 14:24:02.121059 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-5dcf9c57c5-wmjz5"] Nov 21 14:24:02 crc kubenswrapper[4897]: I1121 14:24:02.129980 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xdtcg\" (UniqueName: \"kubernetes.io/projected/ab48e37a-819f-4f63-8e67-9a2cbeaaeb82-kube-api-access-xdtcg\") pod \"console-f58d5cd65-2pc2p\" (UID: \"ab48e37a-819f-4f63-8e67-9a2cbeaaeb82\") " pod="openshift-console/console-f58d5cd65-2pc2p" Nov 21 14:24:02 crc kubenswrapper[4897]: W1121 14:24:02.132955 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcef2e7b5_90a4_4bd2_b8bc_e0513aa09e6d.slice/crio-d337af18f7444804257a7ea78d4bdfa1b8e4ae9f4b6c06f51ebd33ba268c0b6c WatchSource:0}: Error finding container d337af18f7444804257a7ea78d4bdfa1b8e4ae9f4b6c06f51ebd33ba268c0b6c: Status 404 returned error can't find the container with id d337af18f7444804257a7ea78d4bdfa1b8e4ae9f4b6c06f51ebd33ba268c0b6c Nov 21 14:24:02 crc kubenswrapper[4897]: I1121 14:24:02.188640 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f58d5cd65-2pc2p" Nov 21 14:24:02 crc kubenswrapper[4897]: I1121 14:24:02.289078 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-896b2" event={"ID":"7594e6c8-4476-4ceb-b518-500f0f93b3e3","Type":"ContainerStarted","Data":"89a16090b6908b0c0622a315cabc96fffc287dfa728fb4b361925f65b9ab174d"} Nov 21 14:24:02 crc kubenswrapper[4897]: I1121 14:24:02.290575 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-wmjz5" event={"ID":"cef2e7b5-90a4-4bd2-b8bc-e0513aa09e6d","Type":"ContainerStarted","Data":"d337af18f7444804257a7ea78d4bdfa1b8e4ae9f4b6c06f51ebd33ba268c0b6c"} Nov 21 14:24:02 crc kubenswrapper[4897]: I1121 14:24:02.305746 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/3eb1faee-c7f5-4910-9382-46f6e5a7c5f7-plugin-serving-cert\") pod \"nmstate-console-plugin-5874bd7bc5-6tvtd\" (UID: \"3eb1faee-c7f5-4910-9382-46f6e5a7c5f7\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-6tvtd" Nov 21 14:24:02 crc kubenswrapper[4897]: I1121 14:24:02.310149 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/3eb1faee-c7f5-4910-9382-46f6e5a7c5f7-plugin-serving-cert\") pod \"nmstate-console-plugin-5874bd7bc5-6tvtd\" (UID: \"3eb1faee-c7f5-4910-9382-46f6e5a7c5f7\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-6tvtd" Nov 21 14:24:02 crc kubenswrapper[4897]: I1121 14:24:02.551818 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6b89b748d8-lp4lc"] Nov 21 14:24:02 crc kubenswrapper[4897]: W1121 14:24:02.553935 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode41a325d_405a_4d0d_bf21_c8dbd7abcd14.slice/crio-ce6647e845702c4ccab0171416005c75bff0ded6b330a25c323f51f9d58c9165 WatchSource:0}: Error finding container ce6647e845702c4ccab0171416005c75bff0ded6b330a25c323f51f9d58c9165: Status 404 returned error can't find the container with id ce6647e845702c4ccab0171416005c75bff0ded6b330a25c323f51f9d58c9165 Nov 21 14:24:02 crc kubenswrapper[4897]: I1121 14:24:02.590575 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-6tvtd" Nov 21 14:24:02 crc kubenswrapper[4897]: I1121 14:24:02.646175 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f58d5cd65-2pc2p"] Nov 21 14:24:02 crc kubenswrapper[4897]: W1121 14:24:02.670477 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podab48e37a_819f_4f63_8e67_9a2cbeaaeb82.slice/crio-8dfce227bc2bae973871ee709fef9ec9da02fa04602c7a887424930ec0991bb9 WatchSource:0}: Error finding container 8dfce227bc2bae973871ee709fef9ec9da02fa04602c7a887424930ec0991bb9: Status 404 returned error can't find the container with id 8dfce227bc2bae973871ee709fef9ec9da02fa04602c7a887424930ec0991bb9 Nov 21 14:24:03 crc kubenswrapper[4897]: I1121 14:24:03.088462 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5874bd7bc5-6tvtd"] Nov 21 14:24:03 crc kubenswrapper[4897]: I1121 14:24:03.335330 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-6tvtd" event={"ID":"3eb1faee-c7f5-4910-9382-46f6e5a7c5f7","Type":"ContainerStarted","Data":"35880646a524956daef3ebbcf5af07bf6644ff7e162a72fc89256e2e44882146"} Nov 21 14:24:03 crc kubenswrapper[4897]: I1121 14:24:03.348684 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-lp4lc" event={"ID":"e41a325d-405a-4d0d-bf21-c8dbd7abcd14","Type":"ContainerStarted","Data":"ce6647e845702c4ccab0171416005c75bff0ded6b330a25c323f51f9d58c9165"} Nov 21 14:24:03 crc kubenswrapper[4897]: I1121 14:24:03.353588 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f58d5cd65-2pc2p" event={"ID":"ab48e37a-819f-4f63-8e67-9a2cbeaaeb82","Type":"ContainerStarted","Data":"8219d3138b4d11b7091ac5ca0fff7e8fe08b2afd7ab553ce6e30d332a3d5d333"} Nov 21 14:24:03 crc kubenswrapper[4897]: I1121 14:24:03.353628 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f58d5cd65-2pc2p" event={"ID":"ab48e37a-819f-4f63-8e67-9a2cbeaaeb82","Type":"ContainerStarted","Data":"8dfce227bc2bae973871ee709fef9ec9da02fa04602c7a887424930ec0991bb9"} Nov 21 14:24:03 crc kubenswrapper[4897]: I1121 14:24:03.377159 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f58d5cd65-2pc2p" podStartSLOduration=2.377144332 podStartE2EDuration="2.377144332s" podCreationTimestamp="2025-11-21 14:24:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:24:03.376173867 +0000 UTC m=+920.660767352" watchObservedRunningTime="2025-11-21 14:24:03.377144332 +0000 UTC m=+920.661737807" Nov 21 14:24:06 crc kubenswrapper[4897]: I1121 14:24:06.377029 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-6tvtd" event={"ID":"3eb1faee-c7f5-4910-9382-46f6e5a7c5f7","Type":"ContainerStarted","Data":"681f334e527ef7523a989a1ce39b7146c45938f387f391d1ecdea60bd0bad471"} Nov 21 14:24:06 crc kubenswrapper[4897]: I1121 14:24:06.379541 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-896b2" event={"ID":"7594e6c8-4476-4ceb-b518-500f0f93b3e3","Type":"ContainerStarted","Data":"2fdb293895bc8e8335a73c5269333fa4a2479389ea41d1164cd4189d1a31dde2"} Nov 21 14:24:06 crc kubenswrapper[4897]: I1121 14:24:06.379765 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-896b2" Nov 21 14:24:06 crc kubenswrapper[4897]: I1121 14:24:06.381197 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-lp4lc" event={"ID":"e41a325d-405a-4d0d-bf21-c8dbd7abcd14","Type":"ContainerStarted","Data":"f6c15d65580bcd646d2df160deca26bc7d18b3dffeeaf8cad3cc688e896af3b2"} Nov 21 14:24:06 crc kubenswrapper[4897]: I1121 14:24:06.381302 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-lp4lc" Nov 21 14:24:06 crc kubenswrapper[4897]: I1121 14:24:06.382644 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-wmjz5" event={"ID":"cef2e7b5-90a4-4bd2-b8bc-e0513aa09e6d","Type":"ContainerStarted","Data":"17c69eeea43e02e0d693bc180ae1508274391e47c3b878c40fb39578d3b82298"} Nov 21 14:24:06 crc kubenswrapper[4897]: I1121 14:24:06.401275 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-6tvtd" podStartSLOduration=2.749315136 podStartE2EDuration="5.401255545s" podCreationTimestamp="2025-11-21 14:24:01 +0000 UTC" firstStartedPulling="2025-11-21 14:24:03.089813043 +0000 UTC m=+920.374406518" lastFinishedPulling="2025-11-21 14:24:05.741753452 +0000 UTC m=+923.026346927" observedRunningTime="2025-11-21 14:24:06.397572386 +0000 UTC m=+923.682165871" watchObservedRunningTime="2025-11-21 14:24:06.401255545 +0000 UTC m=+923.685849020" Nov 21 14:24:06 crc kubenswrapper[4897]: I1121 14:24:06.437136 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-lp4lc" podStartSLOduration=2.250954388 podStartE2EDuration="5.437114929s" podCreationTimestamp="2025-11-21 14:24:01 +0000 UTC" firstStartedPulling="2025-11-21 14:24:02.555854408 +0000 UTC m=+919.840447883" lastFinishedPulling="2025-11-21 14:24:05.742014929 +0000 UTC m=+923.026608424" observedRunningTime="2025-11-21 14:24:06.418970316 +0000 UTC m=+923.703563791" watchObservedRunningTime="2025-11-21 14:24:06.437114929 +0000 UTC m=+923.721708404" Nov 21 14:24:10 crc kubenswrapper[4897]: I1121 14:24:10.425262 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-wmjz5" event={"ID":"cef2e7b5-90a4-4bd2-b8bc-e0513aa09e6d","Type":"ContainerStarted","Data":"4a439e4c10d9c96a488cadbee7223587538a1e2ed18747eb3cfb2b29435568da"} Nov 21 14:24:10 crc kubenswrapper[4897]: I1121 14:24:10.443383 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-wmjz5" podStartSLOduration=2.253327992 podStartE2EDuration="9.443368201s" podCreationTimestamp="2025-11-21 14:24:01 +0000 UTC" firstStartedPulling="2025-11-21 14:24:02.137096401 +0000 UTC m=+919.421689876" lastFinishedPulling="2025-11-21 14:24:09.32713661 +0000 UTC m=+926.611730085" observedRunningTime="2025-11-21 14:24:10.438145789 +0000 UTC m=+927.722739264" watchObservedRunningTime="2025-11-21 14:24:10.443368201 +0000 UTC m=+927.727961676" Nov 21 14:24:10 crc kubenswrapper[4897]: I1121 14:24:10.443991 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-896b2" podStartSLOduration=5.618861751 podStartE2EDuration="9.443986418s" podCreationTimestamp="2025-11-21 14:24:01 +0000 UTC" firstStartedPulling="2025-11-21 14:24:01.916629965 +0000 UTC m=+919.201223440" lastFinishedPulling="2025-11-21 14:24:05.741754632 +0000 UTC m=+923.026348107" observedRunningTime="2025-11-21 14:24:06.440630624 +0000 UTC m=+923.725224119" watchObservedRunningTime="2025-11-21 14:24:10.443986418 +0000 UTC m=+927.728579883" Nov 21 14:24:11 crc kubenswrapper[4897]: I1121 14:24:11.895388 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-896b2" Nov 21 14:24:12 crc kubenswrapper[4897]: I1121 14:24:12.190568 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f58d5cd65-2pc2p" Nov 21 14:24:12 crc kubenswrapper[4897]: I1121 14:24:12.190653 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f58d5cd65-2pc2p" Nov 21 14:24:12 crc kubenswrapper[4897]: I1121 14:24:12.196799 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f58d5cd65-2pc2p" Nov 21 14:24:12 crc kubenswrapper[4897]: I1121 14:24:12.444761 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f58d5cd65-2pc2p" Nov 21 14:24:12 crc kubenswrapper[4897]: I1121 14:24:12.510128 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-95c6d8978-l44nd"] Nov 21 14:24:22 crc kubenswrapper[4897]: I1121 14:24:22.123282 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-lp4lc" Nov 21 14:24:34 crc kubenswrapper[4897]: I1121 14:24:34.371892 4897 patch_prober.go:28] interesting pod/machine-config-daemon-krv5b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 21 14:24:34 crc kubenswrapper[4897]: I1121 14:24:34.372476 4897 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 21 14:24:37 crc kubenswrapper[4897]: I1121 14:24:37.564335 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-95c6d8978-l44nd" podUID="bb7f0923-4b62-4114-8119-34c39e469323" containerName="console" containerID="cri-o://0f15d6842e5cb4a63dd02145ff446366682c58b091cbf480855eac870aec7321" gracePeriod=15 Nov 21 14:24:38 crc kubenswrapper[4897]: I1121 14:24:38.019239 4897 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-95c6d8978-l44nd_bb7f0923-4b62-4114-8119-34c39e469323/console/0.log" Nov 21 14:24:38 crc kubenswrapper[4897]: I1121 14:24:38.019690 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-95c6d8978-l44nd" Nov 21 14:24:38 crc kubenswrapper[4897]: I1121 14:24:38.165736 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/bb7f0923-4b62-4114-8119-34c39e469323-oauth-serving-cert\") pod \"bb7f0923-4b62-4114-8119-34c39e469323\" (UID: \"bb7f0923-4b62-4114-8119-34c39e469323\") " Nov 21 14:24:38 crc kubenswrapper[4897]: I1121 14:24:38.165828 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xktjm\" (UniqueName: \"kubernetes.io/projected/bb7f0923-4b62-4114-8119-34c39e469323-kube-api-access-xktjm\") pod \"bb7f0923-4b62-4114-8119-34c39e469323\" (UID: \"bb7f0923-4b62-4114-8119-34c39e469323\") " Nov 21 14:24:38 crc kubenswrapper[4897]: I1121 14:24:38.165849 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bb7f0923-4b62-4114-8119-34c39e469323-trusted-ca-bundle\") pod \"bb7f0923-4b62-4114-8119-34c39e469323\" (UID: \"bb7f0923-4b62-4114-8119-34c39e469323\") " Nov 21 14:24:38 crc kubenswrapper[4897]: I1121 14:24:38.165961 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/bb7f0923-4b62-4114-8119-34c39e469323-console-oauth-config\") pod \"bb7f0923-4b62-4114-8119-34c39e469323\" (UID: \"bb7f0923-4b62-4114-8119-34c39e469323\") " Nov 21 14:24:38 crc kubenswrapper[4897]: I1121 14:24:38.166420 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb7f0923-4b62-4114-8119-34c39e469323-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "bb7f0923-4b62-4114-8119-34c39e469323" (UID: "bb7f0923-4b62-4114-8119-34c39e469323"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:24:38 crc kubenswrapper[4897]: I1121 14:24:38.166438 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb7f0923-4b62-4114-8119-34c39e469323-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "bb7f0923-4b62-4114-8119-34c39e469323" (UID: "bb7f0923-4b62-4114-8119-34c39e469323"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:24:38 crc kubenswrapper[4897]: I1121 14:24:38.166534 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/bb7f0923-4b62-4114-8119-34c39e469323-console-serving-cert\") pod \"bb7f0923-4b62-4114-8119-34c39e469323\" (UID: \"bb7f0923-4b62-4114-8119-34c39e469323\") " Nov 21 14:24:38 crc kubenswrapper[4897]: I1121 14:24:38.166555 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/bb7f0923-4b62-4114-8119-34c39e469323-service-ca\") pod \"bb7f0923-4b62-4114-8119-34c39e469323\" (UID: \"bb7f0923-4b62-4114-8119-34c39e469323\") " Nov 21 14:24:38 crc kubenswrapper[4897]: I1121 14:24:38.167030 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/bb7f0923-4b62-4114-8119-34c39e469323-console-config\") pod \"bb7f0923-4b62-4114-8119-34c39e469323\" (UID: \"bb7f0923-4b62-4114-8119-34c39e469323\") " Nov 21 14:24:38 crc kubenswrapper[4897]: I1121 14:24:38.167124 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb7f0923-4b62-4114-8119-34c39e469323-service-ca" (OuterVolumeSpecName: "service-ca") pod "bb7f0923-4b62-4114-8119-34c39e469323" (UID: "bb7f0923-4b62-4114-8119-34c39e469323"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:24:38 crc kubenswrapper[4897]: I1121 14:24:38.167333 4897 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bb7f0923-4b62-4114-8119-34c39e469323-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 21 14:24:38 crc kubenswrapper[4897]: I1121 14:24:38.167344 4897 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/bb7f0923-4b62-4114-8119-34c39e469323-service-ca\") on node \"crc\" DevicePath \"\"" Nov 21 14:24:38 crc kubenswrapper[4897]: I1121 14:24:38.167352 4897 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/bb7f0923-4b62-4114-8119-34c39e469323-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 21 14:24:38 crc kubenswrapper[4897]: I1121 14:24:38.167569 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb7f0923-4b62-4114-8119-34c39e469323-console-config" (OuterVolumeSpecName: "console-config") pod "bb7f0923-4b62-4114-8119-34c39e469323" (UID: "bb7f0923-4b62-4114-8119-34c39e469323"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:24:38 crc kubenswrapper[4897]: I1121 14:24:38.171400 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb7f0923-4b62-4114-8119-34c39e469323-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "bb7f0923-4b62-4114-8119-34c39e469323" (UID: "bb7f0923-4b62-4114-8119-34c39e469323"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:24:38 crc kubenswrapper[4897]: I1121 14:24:38.171469 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb7f0923-4b62-4114-8119-34c39e469323-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "bb7f0923-4b62-4114-8119-34c39e469323" (UID: "bb7f0923-4b62-4114-8119-34c39e469323"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:24:38 crc kubenswrapper[4897]: I1121 14:24:38.171478 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb7f0923-4b62-4114-8119-34c39e469323-kube-api-access-xktjm" (OuterVolumeSpecName: "kube-api-access-xktjm") pod "bb7f0923-4b62-4114-8119-34c39e469323" (UID: "bb7f0923-4b62-4114-8119-34c39e469323"). InnerVolumeSpecName "kube-api-access-xktjm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:24:38 crc kubenswrapper[4897]: I1121 14:24:38.268966 4897 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/bb7f0923-4b62-4114-8119-34c39e469323-console-config\") on node \"crc\" DevicePath \"\"" Nov 21 14:24:38 crc kubenswrapper[4897]: I1121 14:24:38.269008 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xktjm\" (UniqueName: \"kubernetes.io/projected/bb7f0923-4b62-4114-8119-34c39e469323-kube-api-access-xktjm\") on node \"crc\" DevicePath \"\"" Nov 21 14:24:38 crc kubenswrapper[4897]: I1121 14:24:38.269018 4897 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/bb7f0923-4b62-4114-8119-34c39e469323-console-oauth-config\") on node \"crc\" DevicePath \"\"" Nov 21 14:24:38 crc kubenswrapper[4897]: I1121 14:24:38.269026 4897 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/bb7f0923-4b62-4114-8119-34c39e469323-console-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 21 14:24:38 crc kubenswrapper[4897]: I1121 14:24:38.650271 4897 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-95c6d8978-l44nd_bb7f0923-4b62-4114-8119-34c39e469323/console/0.log" Nov 21 14:24:38 crc kubenswrapper[4897]: I1121 14:24:38.650337 4897 generic.go:334] "Generic (PLEG): container finished" podID="bb7f0923-4b62-4114-8119-34c39e469323" containerID="0f15d6842e5cb4a63dd02145ff446366682c58b091cbf480855eac870aec7321" exitCode=2 Nov 21 14:24:38 crc kubenswrapper[4897]: I1121 14:24:38.650403 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-95c6d8978-l44nd" Nov 21 14:24:38 crc kubenswrapper[4897]: I1121 14:24:38.650401 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-95c6d8978-l44nd" event={"ID":"bb7f0923-4b62-4114-8119-34c39e469323","Type":"ContainerDied","Data":"0f15d6842e5cb4a63dd02145ff446366682c58b091cbf480855eac870aec7321"} Nov 21 14:24:38 crc kubenswrapper[4897]: I1121 14:24:38.650479 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-95c6d8978-l44nd" event={"ID":"bb7f0923-4b62-4114-8119-34c39e469323","Type":"ContainerDied","Data":"a4c7a342595f49d237da49e5fa18b725a5e1b6dcd7146c34cb6d10ed016b44d6"} Nov 21 14:24:38 crc kubenswrapper[4897]: I1121 14:24:38.650497 4897 scope.go:117] "RemoveContainer" containerID="0f15d6842e5cb4a63dd02145ff446366682c58b091cbf480855eac870aec7321" Nov 21 14:24:38 crc kubenswrapper[4897]: I1121 14:24:38.678549 4897 scope.go:117] "RemoveContainer" containerID="0f15d6842e5cb4a63dd02145ff446366682c58b091cbf480855eac870aec7321" Nov 21 14:24:38 crc kubenswrapper[4897]: E1121 14:24:38.679932 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0f15d6842e5cb4a63dd02145ff446366682c58b091cbf480855eac870aec7321\": container with ID starting with 0f15d6842e5cb4a63dd02145ff446366682c58b091cbf480855eac870aec7321 not found: ID does not exist" containerID="0f15d6842e5cb4a63dd02145ff446366682c58b091cbf480855eac870aec7321" Nov 21 14:24:38 crc kubenswrapper[4897]: I1121 14:24:38.679968 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f15d6842e5cb4a63dd02145ff446366682c58b091cbf480855eac870aec7321"} err="failed to get container status \"0f15d6842e5cb4a63dd02145ff446366682c58b091cbf480855eac870aec7321\": rpc error: code = NotFound desc = could not find container \"0f15d6842e5cb4a63dd02145ff446366682c58b091cbf480855eac870aec7321\": container with ID starting with 0f15d6842e5cb4a63dd02145ff446366682c58b091cbf480855eac870aec7321 not found: ID does not exist" Nov 21 14:24:38 crc kubenswrapper[4897]: I1121 14:24:38.679988 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-95c6d8978-l44nd"] Nov 21 14:24:38 crc kubenswrapper[4897]: I1121 14:24:38.688852 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-95c6d8978-l44nd"] Nov 21 14:24:40 crc kubenswrapper[4897]: I1121 14:24:40.099866 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb7f0923-4b62-4114-8119-34c39e469323" path="/var/lib/kubelet/pods/bb7f0923-4b62-4114-8119-34c39e469323/volumes" Nov 21 14:24:40 crc kubenswrapper[4897]: I1121 14:24:40.467700 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62d72g"] Nov 21 14:24:40 crc kubenswrapper[4897]: E1121 14:24:40.467981 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb7f0923-4b62-4114-8119-34c39e469323" containerName="console" Nov 21 14:24:40 crc kubenswrapper[4897]: I1121 14:24:40.467995 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb7f0923-4b62-4114-8119-34c39e469323" containerName="console" Nov 21 14:24:40 crc kubenswrapper[4897]: I1121 14:24:40.468158 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb7f0923-4b62-4114-8119-34c39e469323" containerName="console" Nov 21 14:24:40 crc kubenswrapper[4897]: I1121 14:24:40.469094 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62d72g" Nov 21 14:24:40 crc kubenswrapper[4897]: I1121 14:24:40.471120 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 21 14:24:40 crc kubenswrapper[4897]: I1121 14:24:40.478308 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62d72g"] Nov 21 14:24:40 crc kubenswrapper[4897]: I1121 14:24:40.600739 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/79b038ef-a662-4fe7-9db9-d994370a5b8e-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62d72g\" (UID: \"79b038ef-a662-4fe7-9db9-d994370a5b8e\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62d72g" Nov 21 14:24:40 crc kubenswrapper[4897]: I1121 14:24:40.600821 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/79b038ef-a662-4fe7-9db9-d994370a5b8e-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62d72g\" (UID: \"79b038ef-a662-4fe7-9db9-d994370a5b8e\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62d72g" Nov 21 14:24:40 crc kubenswrapper[4897]: I1121 14:24:40.600873 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t89gz\" (UniqueName: \"kubernetes.io/projected/79b038ef-a662-4fe7-9db9-d994370a5b8e-kube-api-access-t89gz\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62d72g\" (UID: \"79b038ef-a662-4fe7-9db9-d994370a5b8e\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62d72g" Nov 21 14:24:40 crc kubenswrapper[4897]: I1121 14:24:40.702466 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/79b038ef-a662-4fe7-9db9-d994370a5b8e-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62d72g\" (UID: \"79b038ef-a662-4fe7-9db9-d994370a5b8e\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62d72g" Nov 21 14:24:40 crc kubenswrapper[4897]: I1121 14:24:40.702568 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t89gz\" (UniqueName: \"kubernetes.io/projected/79b038ef-a662-4fe7-9db9-d994370a5b8e-kube-api-access-t89gz\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62d72g\" (UID: \"79b038ef-a662-4fe7-9db9-d994370a5b8e\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62d72g" Nov 21 14:24:40 crc kubenswrapper[4897]: I1121 14:24:40.702650 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/79b038ef-a662-4fe7-9db9-d994370a5b8e-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62d72g\" (UID: \"79b038ef-a662-4fe7-9db9-d994370a5b8e\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62d72g" Nov 21 14:24:40 crc kubenswrapper[4897]: I1121 14:24:40.703036 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/79b038ef-a662-4fe7-9db9-d994370a5b8e-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62d72g\" (UID: \"79b038ef-a662-4fe7-9db9-d994370a5b8e\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62d72g" Nov 21 14:24:40 crc kubenswrapper[4897]: I1121 14:24:40.703069 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/79b038ef-a662-4fe7-9db9-d994370a5b8e-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62d72g\" (UID: \"79b038ef-a662-4fe7-9db9-d994370a5b8e\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62d72g" Nov 21 14:24:40 crc kubenswrapper[4897]: I1121 14:24:40.721222 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t89gz\" (UniqueName: \"kubernetes.io/projected/79b038ef-a662-4fe7-9db9-d994370a5b8e-kube-api-access-t89gz\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62d72g\" (UID: \"79b038ef-a662-4fe7-9db9-d994370a5b8e\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62d72g" Nov 21 14:24:40 crc kubenswrapper[4897]: I1121 14:24:40.791952 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62d72g" Nov 21 14:24:41 crc kubenswrapper[4897]: I1121 14:24:41.189865 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62d72g"] Nov 21 14:24:41 crc kubenswrapper[4897]: I1121 14:24:41.672616 4897 generic.go:334] "Generic (PLEG): container finished" podID="79b038ef-a662-4fe7-9db9-d994370a5b8e" containerID="359924ebf59ef1da32c67d567f452fd2d2873d1b04939b3a4fccad539e953c51" exitCode=0 Nov 21 14:24:41 crc kubenswrapper[4897]: I1121 14:24:41.672666 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62d72g" event={"ID":"79b038ef-a662-4fe7-9db9-d994370a5b8e","Type":"ContainerDied","Data":"359924ebf59ef1da32c67d567f452fd2d2873d1b04939b3a4fccad539e953c51"} Nov 21 14:24:41 crc kubenswrapper[4897]: I1121 14:24:41.672985 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62d72g" event={"ID":"79b038ef-a662-4fe7-9db9-d994370a5b8e","Type":"ContainerStarted","Data":"357ba421b98bf638ba0da1a1efc1508714f0257f0e1d5d84af27c4656996acca"} Nov 21 14:24:41 crc kubenswrapper[4897]: I1121 14:24:41.674232 4897 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 21 14:24:43 crc kubenswrapper[4897]: I1121 14:24:43.687060 4897 generic.go:334] "Generic (PLEG): container finished" podID="79b038ef-a662-4fe7-9db9-d994370a5b8e" containerID="8b5539bf7c7beb38b8a53773392abcaf6c44f69cab4f826e8f2f721fcd464040" exitCode=0 Nov 21 14:24:43 crc kubenswrapper[4897]: I1121 14:24:43.687184 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62d72g" event={"ID":"79b038ef-a662-4fe7-9db9-d994370a5b8e","Type":"ContainerDied","Data":"8b5539bf7c7beb38b8a53773392abcaf6c44f69cab4f826e8f2f721fcd464040"} Nov 21 14:24:44 crc kubenswrapper[4897]: I1121 14:24:44.698243 4897 generic.go:334] "Generic (PLEG): container finished" podID="79b038ef-a662-4fe7-9db9-d994370a5b8e" containerID="a494465a8fb444778e7e6924b8fff68db0bd36434b4ff3ed4719efca0be60a39" exitCode=0 Nov 21 14:24:44 crc kubenswrapper[4897]: I1121 14:24:44.698581 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62d72g" event={"ID":"79b038ef-a662-4fe7-9db9-d994370a5b8e","Type":"ContainerDied","Data":"a494465a8fb444778e7e6924b8fff68db0bd36434b4ff3ed4719efca0be60a39"} Nov 21 14:24:45 crc kubenswrapper[4897]: I1121 14:24:45.974377 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62d72g" Nov 21 14:24:46 crc kubenswrapper[4897]: I1121 14:24:46.091030 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/79b038ef-a662-4fe7-9db9-d994370a5b8e-util\") pod \"79b038ef-a662-4fe7-9db9-d994370a5b8e\" (UID: \"79b038ef-a662-4fe7-9db9-d994370a5b8e\") " Nov 21 14:24:46 crc kubenswrapper[4897]: I1121 14:24:46.091127 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t89gz\" (UniqueName: \"kubernetes.io/projected/79b038ef-a662-4fe7-9db9-d994370a5b8e-kube-api-access-t89gz\") pod \"79b038ef-a662-4fe7-9db9-d994370a5b8e\" (UID: \"79b038ef-a662-4fe7-9db9-d994370a5b8e\") " Nov 21 14:24:46 crc kubenswrapper[4897]: I1121 14:24:46.091211 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/79b038ef-a662-4fe7-9db9-d994370a5b8e-bundle\") pod \"79b038ef-a662-4fe7-9db9-d994370a5b8e\" (UID: \"79b038ef-a662-4fe7-9db9-d994370a5b8e\") " Nov 21 14:24:46 crc kubenswrapper[4897]: I1121 14:24:46.092706 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/79b038ef-a662-4fe7-9db9-d994370a5b8e-bundle" (OuterVolumeSpecName: "bundle") pod "79b038ef-a662-4fe7-9db9-d994370a5b8e" (UID: "79b038ef-a662-4fe7-9db9-d994370a5b8e"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:24:46 crc kubenswrapper[4897]: I1121 14:24:46.104280 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/79b038ef-a662-4fe7-9db9-d994370a5b8e-util" (OuterVolumeSpecName: "util") pod "79b038ef-a662-4fe7-9db9-d994370a5b8e" (UID: "79b038ef-a662-4fe7-9db9-d994370a5b8e"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:24:46 crc kubenswrapper[4897]: I1121 14:24:46.104872 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79b038ef-a662-4fe7-9db9-d994370a5b8e-kube-api-access-t89gz" (OuterVolumeSpecName: "kube-api-access-t89gz") pod "79b038ef-a662-4fe7-9db9-d994370a5b8e" (UID: "79b038ef-a662-4fe7-9db9-d994370a5b8e"). InnerVolumeSpecName "kube-api-access-t89gz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:24:46 crc kubenswrapper[4897]: I1121 14:24:46.193235 4897 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/79b038ef-a662-4fe7-9db9-d994370a5b8e-util\") on node \"crc\" DevicePath \"\"" Nov 21 14:24:46 crc kubenswrapper[4897]: I1121 14:24:46.193294 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t89gz\" (UniqueName: \"kubernetes.io/projected/79b038ef-a662-4fe7-9db9-d994370a5b8e-kube-api-access-t89gz\") on node \"crc\" DevicePath \"\"" Nov 21 14:24:46 crc kubenswrapper[4897]: I1121 14:24:46.193307 4897 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/79b038ef-a662-4fe7-9db9-d994370a5b8e-bundle\") on node \"crc\" DevicePath \"\"" Nov 21 14:24:46 crc kubenswrapper[4897]: I1121 14:24:46.716947 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62d72g" event={"ID":"79b038ef-a662-4fe7-9db9-d994370a5b8e","Type":"ContainerDied","Data":"357ba421b98bf638ba0da1a1efc1508714f0257f0e1d5d84af27c4656996acca"} Nov 21 14:24:46 crc kubenswrapper[4897]: I1121 14:24:46.717002 4897 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="357ba421b98bf638ba0da1a1efc1508714f0257f0e1d5d84af27c4656996acca" Nov 21 14:24:46 crc kubenswrapper[4897]: I1121 14:24:46.717041 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c62d72g" Nov 21 14:24:56 crc kubenswrapper[4897]: I1121 14:24:56.027487 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-5dbdbdc9f-xwr9r"] Nov 21 14:24:56 crc kubenswrapper[4897]: E1121 14:24:56.028258 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79b038ef-a662-4fe7-9db9-d994370a5b8e" containerName="util" Nov 21 14:24:56 crc kubenswrapper[4897]: I1121 14:24:56.028270 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="79b038ef-a662-4fe7-9db9-d994370a5b8e" containerName="util" Nov 21 14:24:56 crc kubenswrapper[4897]: E1121 14:24:56.028285 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79b038ef-a662-4fe7-9db9-d994370a5b8e" containerName="extract" Nov 21 14:24:56 crc kubenswrapper[4897]: I1121 14:24:56.028291 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="79b038ef-a662-4fe7-9db9-d994370a5b8e" containerName="extract" Nov 21 14:24:56 crc kubenswrapper[4897]: E1121 14:24:56.028303 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79b038ef-a662-4fe7-9db9-d994370a5b8e" containerName="pull" Nov 21 14:24:56 crc kubenswrapper[4897]: I1121 14:24:56.028310 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="79b038ef-a662-4fe7-9db9-d994370a5b8e" containerName="pull" Nov 21 14:24:56 crc kubenswrapper[4897]: I1121 14:24:56.028441 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="79b038ef-a662-4fe7-9db9-d994370a5b8e" containerName="extract" Nov 21 14:24:56 crc kubenswrapper[4897]: I1121 14:24:56.028960 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-5dbdbdc9f-xwr9r" Nov 21 14:24:56 crc kubenswrapper[4897]: I1121 14:24:56.031085 4897 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Nov 21 14:24:56 crc kubenswrapper[4897]: I1121 14:24:56.031308 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Nov 21 14:24:56 crc kubenswrapper[4897]: I1121 14:24:56.031583 4897 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-ntxgq" Nov 21 14:24:56 crc kubenswrapper[4897]: I1121 14:24:56.032383 4897 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Nov 21 14:24:56 crc kubenswrapper[4897]: I1121 14:24:56.032584 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Nov 21 14:24:56 crc kubenswrapper[4897]: I1121 14:24:56.046511 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-5dbdbdc9f-xwr9r"] Nov 21 14:24:56 crc kubenswrapper[4897]: I1121 14:24:56.156989 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c9620249-ddf2-49b0-bee0-5415988b4a1e-webhook-cert\") pod \"metallb-operator-controller-manager-5dbdbdc9f-xwr9r\" (UID: \"c9620249-ddf2-49b0-bee0-5415988b4a1e\") " pod="metallb-system/metallb-operator-controller-manager-5dbdbdc9f-xwr9r" Nov 21 14:24:56 crc kubenswrapper[4897]: I1121 14:24:56.157073 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c9620249-ddf2-49b0-bee0-5415988b4a1e-apiservice-cert\") pod \"metallb-operator-controller-manager-5dbdbdc9f-xwr9r\" (UID: \"c9620249-ddf2-49b0-bee0-5415988b4a1e\") " pod="metallb-system/metallb-operator-controller-manager-5dbdbdc9f-xwr9r" Nov 21 14:24:56 crc kubenswrapper[4897]: I1121 14:24:56.157182 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vns2x\" (UniqueName: \"kubernetes.io/projected/c9620249-ddf2-49b0-bee0-5415988b4a1e-kube-api-access-vns2x\") pod \"metallb-operator-controller-manager-5dbdbdc9f-xwr9r\" (UID: \"c9620249-ddf2-49b0-bee0-5415988b4a1e\") " pod="metallb-system/metallb-operator-controller-manager-5dbdbdc9f-xwr9r" Nov 21 14:24:56 crc kubenswrapper[4897]: I1121 14:24:56.258683 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vns2x\" (UniqueName: \"kubernetes.io/projected/c9620249-ddf2-49b0-bee0-5415988b4a1e-kube-api-access-vns2x\") pod \"metallb-operator-controller-manager-5dbdbdc9f-xwr9r\" (UID: \"c9620249-ddf2-49b0-bee0-5415988b4a1e\") " pod="metallb-system/metallb-operator-controller-manager-5dbdbdc9f-xwr9r" Nov 21 14:24:56 crc kubenswrapper[4897]: I1121 14:24:56.258840 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c9620249-ddf2-49b0-bee0-5415988b4a1e-webhook-cert\") pod \"metallb-operator-controller-manager-5dbdbdc9f-xwr9r\" (UID: \"c9620249-ddf2-49b0-bee0-5415988b4a1e\") " pod="metallb-system/metallb-operator-controller-manager-5dbdbdc9f-xwr9r" Nov 21 14:24:56 crc kubenswrapper[4897]: I1121 14:24:56.258919 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c9620249-ddf2-49b0-bee0-5415988b4a1e-apiservice-cert\") pod \"metallb-operator-controller-manager-5dbdbdc9f-xwr9r\" (UID: \"c9620249-ddf2-49b0-bee0-5415988b4a1e\") " pod="metallb-system/metallb-operator-controller-manager-5dbdbdc9f-xwr9r" Nov 21 14:24:56 crc kubenswrapper[4897]: I1121 14:24:56.264699 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c9620249-ddf2-49b0-bee0-5415988b4a1e-webhook-cert\") pod \"metallb-operator-controller-manager-5dbdbdc9f-xwr9r\" (UID: \"c9620249-ddf2-49b0-bee0-5415988b4a1e\") " pod="metallb-system/metallb-operator-controller-manager-5dbdbdc9f-xwr9r" Nov 21 14:24:56 crc kubenswrapper[4897]: I1121 14:24:56.264729 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c9620249-ddf2-49b0-bee0-5415988b4a1e-apiservice-cert\") pod \"metallb-operator-controller-manager-5dbdbdc9f-xwr9r\" (UID: \"c9620249-ddf2-49b0-bee0-5415988b4a1e\") " pod="metallb-system/metallb-operator-controller-manager-5dbdbdc9f-xwr9r" Nov 21 14:24:56 crc kubenswrapper[4897]: I1121 14:24:56.277388 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vns2x\" (UniqueName: \"kubernetes.io/projected/c9620249-ddf2-49b0-bee0-5415988b4a1e-kube-api-access-vns2x\") pod \"metallb-operator-controller-manager-5dbdbdc9f-xwr9r\" (UID: \"c9620249-ddf2-49b0-bee0-5415988b4a1e\") " pod="metallb-system/metallb-operator-controller-manager-5dbdbdc9f-xwr9r" Nov 21 14:24:56 crc kubenswrapper[4897]: I1121 14:24:56.348670 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-69588b6cf-fwgq4"] Nov 21 14:24:56 crc kubenswrapper[4897]: I1121 14:24:56.349560 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-69588b6cf-fwgq4" Nov 21 14:24:56 crc kubenswrapper[4897]: I1121 14:24:56.351307 4897 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Nov 21 14:24:56 crc kubenswrapper[4897]: I1121 14:24:56.353235 4897 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-2s8nj" Nov 21 14:24:56 crc kubenswrapper[4897]: I1121 14:24:56.353252 4897 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Nov 21 14:24:56 crc kubenswrapper[4897]: I1121 14:24:56.360944 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/76c9e30b-59a8-45fe-bf9d-5d7152710c73-apiservice-cert\") pod \"metallb-operator-webhook-server-69588b6cf-fwgq4\" (UID: \"76c9e30b-59a8-45fe-bf9d-5d7152710c73\") " pod="metallb-system/metallb-operator-webhook-server-69588b6cf-fwgq4" Nov 21 14:24:56 crc kubenswrapper[4897]: I1121 14:24:56.361000 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/76c9e30b-59a8-45fe-bf9d-5d7152710c73-webhook-cert\") pod \"metallb-operator-webhook-server-69588b6cf-fwgq4\" (UID: \"76c9e30b-59a8-45fe-bf9d-5d7152710c73\") " pod="metallb-system/metallb-operator-webhook-server-69588b6cf-fwgq4" Nov 21 14:24:56 crc kubenswrapper[4897]: I1121 14:24:56.361018 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-69588b6cf-fwgq4"] Nov 21 14:24:56 crc kubenswrapper[4897]: I1121 14:24:56.361072 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lcfbd\" (UniqueName: \"kubernetes.io/projected/76c9e30b-59a8-45fe-bf9d-5d7152710c73-kube-api-access-lcfbd\") pod \"metallb-operator-webhook-server-69588b6cf-fwgq4\" (UID: \"76c9e30b-59a8-45fe-bf9d-5d7152710c73\") " pod="metallb-system/metallb-operator-webhook-server-69588b6cf-fwgq4" Nov 21 14:24:56 crc kubenswrapper[4897]: I1121 14:24:56.383214 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-5dbdbdc9f-xwr9r" Nov 21 14:24:56 crc kubenswrapper[4897]: I1121 14:24:56.462485 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lcfbd\" (UniqueName: \"kubernetes.io/projected/76c9e30b-59a8-45fe-bf9d-5d7152710c73-kube-api-access-lcfbd\") pod \"metallb-operator-webhook-server-69588b6cf-fwgq4\" (UID: \"76c9e30b-59a8-45fe-bf9d-5d7152710c73\") " pod="metallb-system/metallb-operator-webhook-server-69588b6cf-fwgq4" Nov 21 14:24:56 crc kubenswrapper[4897]: I1121 14:24:56.462597 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/76c9e30b-59a8-45fe-bf9d-5d7152710c73-apiservice-cert\") pod \"metallb-operator-webhook-server-69588b6cf-fwgq4\" (UID: \"76c9e30b-59a8-45fe-bf9d-5d7152710c73\") " pod="metallb-system/metallb-operator-webhook-server-69588b6cf-fwgq4" Nov 21 14:24:56 crc kubenswrapper[4897]: I1121 14:24:56.462628 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/76c9e30b-59a8-45fe-bf9d-5d7152710c73-webhook-cert\") pod \"metallb-operator-webhook-server-69588b6cf-fwgq4\" (UID: \"76c9e30b-59a8-45fe-bf9d-5d7152710c73\") " pod="metallb-system/metallb-operator-webhook-server-69588b6cf-fwgq4" Nov 21 14:24:56 crc kubenswrapper[4897]: I1121 14:24:56.473220 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/76c9e30b-59a8-45fe-bf9d-5d7152710c73-webhook-cert\") pod \"metallb-operator-webhook-server-69588b6cf-fwgq4\" (UID: \"76c9e30b-59a8-45fe-bf9d-5d7152710c73\") " pod="metallb-system/metallb-operator-webhook-server-69588b6cf-fwgq4" Nov 21 14:24:56 crc kubenswrapper[4897]: I1121 14:24:56.473276 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/76c9e30b-59a8-45fe-bf9d-5d7152710c73-apiservice-cert\") pod \"metallb-operator-webhook-server-69588b6cf-fwgq4\" (UID: \"76c9e30b-59a8-45fe-bf9d-5d7152710c73\") " pod="metallb-system/metallb-operator-webhook-server-69588b6cf-fwgq4" Nov 21 14:24:56 crc kubenswrapper[4897]: I1121 14:24:56.480105 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lcfbd\" (UniqueName: \"kubernetes.io/projected/76c9e30b-59a8-45fe-bf9d-5d7152710c73-kube-api-access-lcfbd\") pod \"metallb-operator-webhook-server-69588b6cf-fwgq4\" (UID: \"76c9e30b-59a8-45fe-bf9d-5d7152710c73\") " pod="metallb-system/metallb-operator-webhook-server-69588b6cf-fwgq4" Nov 21 14:24:56 crc kubenswrapper[4897]: I1121 14:24:56.646188 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-5dbdbdc9f-xwr9r"] Nov 21 14:24:56 crc kubenswrapper[4897]: W1121 14:24:56.652041 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc9620249_ddf2_49b0_bee0_5415988b4a1e.slice/crio-6f95eb89c4fbf5cf3210721afca733e6c4a0183a0bf7ec883dd929e5130dcdbd WatchSource:0}: Error finding container 6f95eb89c4fbf5cf3210721afca733e6c4a0183a0bf7ec883dd929e5130dcdbd: Status 404 returned error can't find the container with id 6f95eb89c4fbf5cf3210721afca733e6c4a0183a0bf7ec883dd929e5130dcdbd Nov 21 14:24:56 crc kubenswrapper[4897]: I1121 14:24:56.667230 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-69588b6cf-fwgq4" Nov 21 14:24:56 crc kubenswrapper[4897]: I1121 14:24:56.789660 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-5dbdbdc9f-xwr9r" event={"ID":"c9620249-ddf2-49b0-bee0-5415988b4a1e","Type":"ContainerStarted","Data":"6f95eb89c4fbf5cf3210721afca733e6c4a0183a0bf7ec883dd929e5130dcdbd"} Nov 21 14:24:57 crc kubenswrapper[4897]: I1121 14:24:57.107105 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-69588b6cf-fwgq4"] Nov 21 14:24:57 crc kubenswrapper[4897]: W1121 14:24:57.115559 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod76c9e30b_59a8_45fe_bf9d_5d7152710c73.slice/crio-6af25de64419b44366565e245cc0a72386931b978b39589525113bce08b248cd WatchSource:0}: Error finding container 6af25de64419b44366565e245cc0a72386931b978b39589525113bce08b248cd: Status 404 returned error can't find the container with id 6af25de64419b44366565e245cc0a72386931b978b39589525113bce08b248cd Nov 21 14:24:57 crc kubenswrapper[4897]: I1121 14:24:57.798839 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-69588b6cf-fwgq4" event={"ID":"76c9e30b-59a8-45fe-bf9d-5d7152710c73","Type":"ContainerStarted","Data":"6af25de64419b44366565e245cc0a72386931b978b39589525113bce08b248cd"} Nov 21 14:25:04 crc kubenswrapper[4897]: I1121 14:25:04.371366 4897 patch_prober.go:28] interesting pod/machine-config-daemon-krv5b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 21 14:25:04 crc kubenswrapper[4897]: I1121 14:25:04.373880 4897 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 21 14:25:04 crc kubenswrapper[4897]: I1121 14:25:04.859466 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-5dbdbdc9f-xwr9r" event={"ID":"c9620249-ddf2-49b0-bee0-5415988b4a1e","Type":"ContainerStarted","Data":"778a0f7c86401dad450b70ebcbc96da18dd5d11c9c76599059ea644eceb24820"} Nov 21 14:25:04 crc kubenswrapper[4897]: I1121 14:25:04.859865 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-5dbdbdc9f-xwr9r" Nov 21 14:25:04 crc kubenswrapper[4897]: I1121 14:25:04.862222 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-69588b6cf-fwgq4" event={"ID":"76c9e30b-59a8-45fe-bf9d-5d7152710c73","Type":"ContainerStarted","Data":"7b03f46a0fcc4d4166bf0ed10cbab4d6bb6bec40bf3de4bb26e22ee3f02e05d0"} Nov 21 14:25:04 crc kubenswrapper[4897]: I1121 14:25:04.862354 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-69588b6cf-fwgq4" Nov 21 14:25:04 crc kubenswrapper[4897]: I1121 14:25:04.877319 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-5dbdbdc9f-xwr9r" podStartSLOduration=1.13896835 podStartE2EDuration="8.877302097s" podCreationTimestamp="2025-11-21 14:24:56 +0000 UTC" firstStartedPulling="2025-11-21 14:24:56.65367175 +0000 UTC m=+973.938265225" lastFinishedPulling="2025-11-21 14:25:04.392005497 +0000 UTC m=+981.676598972" observedRunningTime="2025-11-21 14:25:04.875404726 +0000 UTC m=+982.159998211" watchObservedRunningTime="2025-11-21 14:25:04.877302097 +0000 UTC m=+982.161895572" Nov 21 14:25:16 crc kubenswrapper[4897]: I1121 14:25:16.673760 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-69588b6cf-fwgq4" Nov 21 14:25:16 crc kubenswrapper[4897]: I1121 14:25:16.696077 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-69588b6cf-fwgq4" podStartSLOduration=13.388805979 podStartE2EDuration="20.696060376s" podCreationTimestamp="2025-11-21 14:24:56 +0000 UTC" firstStartedPulling="2025-11-21 14:24:57.118885195 +0000 UTC m=+974.403478670" lastFinishedPulling="2025-11-21 14:25:04.426139572 +0000 UTC m=+981.710733067" observedRunningTime="2025-11-21 14:25:04.905441131 +0000 UTC m=+982.190034606" watchObservedRunningTime="2025-11-21 14:25:16.696060376 +0000 UTC m=+993.980653851" Nov 21 14:25:34 crc kubenswrapper[4897]: I1121 14:25:34.371007 4897 patch_prober.go:28] interesting pod/machine-config-daemon-krv5b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 21 14:25:34 crc kubenswrapper[4897]: I1121 14:25:34.371856 4897 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 21 14:25:34 crc kubenswrapper[4897]: I1121 14:25:34.371930 4897 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" Nov 21 14:25:34 crc kubenswrapper[4897]: I1121 14:25:34.373098 4897 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6f8d8438a69d75aad7ccb460842fd43be11b0196e6779f692d047f7ed8f74e83"} pod="openshift-machine-config-operator/machine-config-daemon-krv5b" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 21 14:25:34 crc kubenswrapper[4897]: I1121 14:25:34.373238 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" containerName="machine-config-daemon" containerID="cri-o://6f8d8438a69d75aad7ccb460842fd43be11b0196e6779f692d047f7ed8f74e83" gracePeriod=600 Nov 21 14:25:35 crc kubenswrapper[4897]: I1121 14:25:35.110208 4897 generic.go:334] "Generic (PLEG): container finished" podID="e7670227-d280-4847-b882-754429f56b0f" containerID="6f8d8438a69d75aad7ccb460842fd43be11b0196e6779f692d047f7ed8f74e83" exitCode=0 Nov 21 14:25:35 crc kubenswrapper[4897]: I1121 14:25:35.110830 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" event={"ID":"e7670227-d280-4847-b882-754429f56b0f","Type":"ContainerDied","Data":"6f8d8438a69d75aad7ccb460842fd43be11b0196e6779f692d047f7ed8f74e83"} Nov 21 14:25:35 crc kubenswrapper[4897]: I1121 14:25:35.110858 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" event={"ID":"e7670227-d280-4847-b882-754429f56b0f","Type":"ContainerStarted","Data":"cdc84dcb39834c826868fb54c0e5d0aa88dbf0ae2b0ceea30db993577fc2bded"} Nov 21 14:25:35 crc kubenswrapper[4897]: I1121 14:25:35.110874 4897 scope.go:117] "RemoveContainer" containerID="cad6f00fa42dcd6129303195cd6590d896370bbb75a53d50c3950cca8fb66efa" Nov 21 14:25:36 crc kubenswrapper[4897]: I1121 14:25:36.385962 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-5dbdbdc9f-xwr9r" Nov 21 14:25:37 crc kubenswrapper[4897]: I1121 14:25:37.229444 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-x7jrs"] Nov 21 14:25:37 crc kubenswrapper[4897]: I1121 14:25:37.230648 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-6998585d5-x7jrs" Nov 21 14:25:37 crc kubenswrapper[4897]: I1121 14:25:37.232657 4897 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Nov 21 14:25:37 crc kubenswrapper[4897]: I1121 14:25:37.235552 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-2v89k"] Nov 21 14:25:37 crc kubenswrapper[4897]: I1121 14:25:37.240172 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-2v89k" Nov 21 14:25:37 crc kubenswrapper[4897]: I1121 14:25:37.243957 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Nov 21 14:25:37 crc kubenswrapper[4897]: I1121 14:25:37.245762 4897 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Nov 21 14:25:37 crc kubenswrapper[4897]: I1121 14:25:37.250698 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-x7jrs"] Nov 21 14:25:37 crc kubenswrapper[4897]: I1121 14:25:37.252741 4897 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-59w8p" Nov 21 14:25:37 crc kubenswrapper[4897]: I1121 14:25:37.309661 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-hqq7g"] Nov 21 14:25:37 crc kubenswrapper[4897]: I1121 14:25:37.310807 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-hqq7g" Nov 21 14:25:37 crc kubenswrapper[4897]: I1121 14:25:37.312882 4897 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Nov 21 14:25:37 crc kubenswrapper[4897]: I1121 14:25:37.312927 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Nov 21 14:25:37 crc kubenswrapper[4897]: I1121 14:25:37.313697 4897 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-6bsnh" Nov 21 14:25:37 crc kubenswrapper[4897]: I1121 14:25:37.314318 4897 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Nov 21 14:25:37 crc kubenswrapper[4897]: I1121 14:25:37.344295 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-6c7b4b5f48-6bhsc"] Nov 21 14:25:37 crc kubenswrapper[4897]: I1121 14:25:37.345825 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6c7b4b5f48-6bhsc" Nov 21 14:25:37 crc kubenswrapper[4897]: I1121 14:25:37.349174 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/72cb5830-c763-4dca-a08f-d5ea7abb07fe-reloader\") pod \"frr-k8s-2v89k\" (UID: \"72cb5830-c763-4dca-a08f-d5ea7abb07fe\") " pod="metallb-system/frr-k8s-2v89k" Nov 21 14:25:37 crc kubenswrapper[4897]: I1121 14:25:37.349282 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4c46f\" (UniqueName: \"kubernetes.io/projected/dda4b910-8a41-4076-b92d-030e277f9778-kube-api-access-4c46f\") pod \"frr-k8s-webhook-server-6998585d5-x7jrs\" (UID: \"dda4b910-8a41-4076-b92d-030e277f9778\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-x7jrs" Nov 21 14:25:37 crc kubenswrapper[4897]: I1121 14:25:37.349317 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/72cb5830-c763-4dca-a08f-d5ea7abb07fe-metrics\") pod \"frr-k8s-2v89k\" (UID: \"72cb5830-c763-4dca-a08f-d5ea7abb07fe\") " pod="metallb-system/frr-k8s-2v89k" Nov 21 14:25:37 crc kubenswrapper[4897]: I1121 14:25:37.349352 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hqtth\" (UniqueName: \"kubernetes.io/projected/72cb5830-c763-4dca-a08f-d5ea7abb07fe-kube-api-access-hqtth\") pod \"frr-k8s-2v89k\" (UID: \"72cb5830-c763-4dca-a08f-d5ea7abb07fe\") " pod="metallb-system/frr-k8s-2v89k" Nov 21 14:25:37 crc kubenswrapper[4897]: I1121 14:25:37.349391 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/72cb5830-c763-4dca-a08f-d5ea7abb07fe-metrics-certs\") pod \"frr-k8s-2v89k\" (UID: \"72cb5830-c763-4dca-a08f-d5ea7abb07fe\") " pod="metallb-system/frr-k8s-2v89k" Nov 21 14:25:37 crc kubenswrapper[4897]: I1121 14:25:37.349413 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/dda4b910-8a41-4076-b92d-030e277f9778-cert\") pod \"frr-k8s-webhook-server-6998585d5-x7jrs\" (UID: \"dda4b910-8a41-4076-b92d-030e277f9778\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-x7jrs" Nov 21 14:25:37 crc kubenswrapper[4897]: I1121 14:25:37.349435 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/72cb5830-c763-4dca-a08f-d5ea7abb07fe-frr-startup\") pod \"frr-k8s-2v89k\" (UID: \"72cb5830-c763-4dca-a08f-d5ea7abb07fe\") " pod="metallb-system/frr-k8s-2v89k" Nov 21 14:25:37 crc kubenswrapper[4897]: I1121 14:25:37.349466 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/72cb5830-c763-4dca-a08f-d5ea7abb07fe-frr-conf\") pod \"frr-k8s-2v89k\" (UID: \"72cb5830-c763-4dca-a08f-d5ea7abb07fe\") " pod="metallb-system/frr-k8s-2v89k" Nov 21 14:25:37 crc kubenswrapper[4897]: I1121 14:25:37.349541 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/72cb5830-c763-4dca-a08f-d5ea7abb07fe-frr-sockets\") pod \"frr-k8s-2v89k\" (UID: \"72cb5830-c763-4dca-a08f-d5ea7abb07fe\") " pod="metallb-system/frr-k8s-2v89k" Nov 21 14:25:37 crc kubenswrapper[4897]: I1121 14:25:37.350442 4897 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Nov 21 14:25:37 crc kubenswrapper[4897]: I1121 14:25:37.355114 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6c7b4b5f48-6bhsc"] Nov 21 14:25:37 crc kubenswrapper[4897]: I1121 14:25:37.450440 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4c46f\" (UniqueName: \"kubernetes.io/projected/dda4b910-8a41-4076-b92d-030e277f9778-kube-api-access-4c46f\") pod \"frr-k8s-webhook-server-6998585d5-x7jrs\" (UID: \"dda4b910-8a41-4076-b92d-030e277f9778\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-x7jrs" Nov 21 14:25:37 crc kubenswrapper[4897]: I1121 14:25:37.450798 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/72cb5830-c763-4dca-a08f-d5ea7abb07fe-metrics\") pod \"frr-k8s-2v89k\" (UID: \"72cb5830-c763-4dca-a08f-d5ea7abb07fe\") " pod="metallb-system/frr-k8s-2v89k" Nov 21 14:25:37 crc kubenswrapper[4897]: I1121 14:25:37.450830 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0c21212c-1adf-4bb1-8a7e-c1df5e98e0d8-metrics-certs\") pod \"controller-6c7b4b5f48-6bhsc\" (UID: \"0c21212c-1adf-4bb1-8a7e-c1df5e98e0d8\") " pod="metallb-system/controller-6c7b4b5f48-6bhsc" Nov 21 14:25:37 crc kubenswrapper[4897]: I1121 14:25:37.450856 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/89b1034c-5696-4975-a683-fdd9c163801f-metallb-excludel2\") pod \"speaker-hqq7g\" (UID: \"89b1034c-5696-4975-a683-fdd9c163801f\") " pod="metallb-system/speaker-hqq7g" Nov 21 14:25:37 crc kubenswrapper[4897]: I1121 14:25:37.450883 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hqtth\" (UniqueName: \"kubernetes.io/projected/72cb5830-c763-4dca-a08f-d5ea7abb07fe-kube-api-access-hqtth\") pod \"frr-k8s-2v89k\" (UID: \"72cb5830-c763-4dca-a08f-d5ea7abb07fe\") " pod="metallb-system/frr-k8s-2v89k" Nov 21 14:25:37 crc kubenswrapper[4897]: I1121 14:25:37.450909 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/72cb5830-c763-4dca-a08f-d5ea7abb07fe-metrics-certs\") pod \"frr-k8s-2v89k\" (UID: \"72cb5830-c763-4dca-a08f-d5ea7abb07fe\") " pod="metallb-system/frr-k8s-2v89k" Nov 21 14:25:37 crc kubenswrapper[4897]: I1121 14:25:37.450944 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/dda4b910-8a41-4076-b92d-030e277f9778-cert\") pod \"frr-k8s-webhook-server-6998585d5-x7jrs\" (UID: \"dda4b910-8a41-4076-b92d-030e277f9778\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-x7jrs" Nov 21 14:25:37 crc kubenswrapper[4897]: I1121 14:25:37.450964 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/72cb5830-c763-4dca-a08f-d5ea7abb07fe-frr-startup\") pod \"frr-k8s-2v89k\" (UID: \"72cb5830-c763-4dca-a08f-d5ea7abb07fe\") " pod="metallb-system/frr-k8s-2v89k" Nov 21 14:25:37 crc kubenswrapper[4897]: I1121 14:25:37.450983 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/89b1034c-5696-4975-a683-fdd9c163801f-metrics-certs\") pod \"speaker-hqq7g\" (UID: \"89b1034c-5696-4975-a683-fdd9c163801f\") " pod="metallb-system/speaker-hqq7g" Nov 21 14:25:37 crc kubenswrapper[4897]: I1121 14:25:37.451000 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0c21212c-1adf-4bb1-8a7e-c1df5e98e0d8-cert\") pod \"controller-6c7b4b5f48-6bhsc\" (UID: \"0c21212c-1adf-4bb1-8a7e-c1df5e98e0d8\") " pod="metallb-system/controller-6c7b4b5f48-6bhsc" Nov 21 14:25:37 crc kubenswrapper[4897]: I1121 14:25:37.451019 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/72cb5830-c763-4dca-a08f-d5ea7abb07fe-frr-conf\") pod \"frr-k8s-2v89k\" (UID: \"72cb5830-c763-4dca-a08f-d5ea7abb07fe\") " pod="metallb-system/frr-k8s-2v89k" Nov 21 14:25:37 crc kubenswrapper[4897]: I1121 14:25:37.451049 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/89b1034c-5696-4975-a683-fdd9c163801f-memberlist\") pod \"speaker-hqq7g\" (UID: \"89b1034c-5696-4975-a683-fdd9c163801f\") " pod="metallb-system/speaker-hqq7g" Nov 21 14:25:37 crc kubenswrapper[4897]: I1121 14:25:37.451082 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/72cb5830-c763-4dca-a08f-d5ea7abb07fe-frr-sockets\") pod \"frr-k8s-2v89k\" (UID: \"72cb5830-c763-4dca-a08f-d5ea7abb07fe\") " pod="metallb-system/frr-k8s-2v89k" Nov 21 14:25:37 crc kubenswrapper[4897]: I1121 14:25:37.451105 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rx2zv\" (UniqueName: \"kubernetes.io/projected/89b1034c-5696-4975-a683-fdd9c163801f-kube-api-access-rx2zv\") pod \"speaker-hqq7g\" (UID: \"89b1034c-5696-4975-a683-fdd9c163801f\") " pod="metallb-system/speaker-hqq7g" Nov 21 14:25:37 crc kubenswrapper[4897]: I1121 14:25:37.451125 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fkrhd\" (UniqueName: \"kubernetes.io/projected/0c21212c-1adf-4bb1-8a7e-c1df5e98e0d8-kube-api-access-fkrhd\") pod \"controller-6c7b4b5f48-6bhsc\" (UID: \"0c21212c-1adf-4bb1-8a7e-c1df5e98e0d8\") " pod="metallb-system/controller-6c7b4b5f48-6bhsc" Nov 21 14:25:37 crc kubenswrapper[4897]: I1121 14:25:37.451229 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/72cb5830-c763-4dca-a08f-d5ea7abb07fe-metrics\") pod \"frr-k8s-2v89k\" (UID: \"72cb5830-c763-4dca-a08f-d5ea7abb07fe\") " pod="metallb-system/frr-k8s-2v89k" Nov 21 14:25:37 crc kubenswrapper[4897]: I1121 14:25:37.451361 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/72cb5830-c763-4dca-a08f-d5ea7abb07fe-reloader\") pod \"frr-k8s-2v89k\" (UID: \"72cb5830-c763-4dca-a08f-d5ea7abb07fe\") " pod="metallb-system/frr-k8s-2v89k" Nov 21 14:25:37 crc kubenswrapper[4897]: I1121 14:25:37.451496 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/72cb5830-c763-4dca-a08f-d5ea7abb07fe-frr-conf\") pod \"frr-k8s-2v89k\" (UID: \"72cb5830-c763-4dca-a08f-d5ea7abb07fe\") " pod="metallb-system/frr-k8s-2v89k" Nov 21 14:25:37 crc kubenswrapper[4897]: I1121 14:25:37.451726 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/72cb5830-c763-4dca-a08f-d5ea7abb07fe-frr-sockets\") pod \"frr-k8s-2v89k\" (UID: \"72cb5830-c763-4dca-a08f-d5ea7abb07fe\") " pod="metallb-system/frr-k8s-2v89k" Nov 21 14:25:37 crc kubenswrapper[4897]: I1121 14:25:37.451751 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/72cb5830-c763-4dca-a08f-d5ea7abb07fe-reloader\") pod \"frr-k8s-2v89k\" (UID: \"72cb5830-c763-4dca-a08f-d5ea7abb07fe\") " pod="metallb-system/frr-k8s-2v89k" Nov 21 14:25:37 crc kubenswrapper[4897]: I1121 14:25:37.452261 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/72cb5830-c763-4dca-a08f-d5ea7abb07fe-frr-startup\") pod \"frr-k8s-2v89k\" (UID: \"72cb5830-c763-4dca-a08f-d5ea7abb07fe\") " pod="metallb-system/frr-k8s-2v89k" Nov 21 14:25:37 crc kubenswrapper[4897]: I1121 14:25:37.457104 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/dda4b910-8a41-4076-b92d-030e277f9778-cert\") pod \"frr-k8s-webhook-server-6998585d5-x7jrs\" (UID: \"dda4b910-8a41-4076-b92d-030e277f9778\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-x7jrs" Nov 21 14:25:37 crc kubenswrapper[4897]: I1121 14:25:37.470754 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/72cb5830-c763-4dca-a08f-d5ea7abb07fe-metrics-certs\") pod \"frr-k8s-2v89k\" (UID: \"72cb5830-c763-4dca-a08f-d5ea7abb07fe\") " pod="metallb-system/frr-k8s-2v89k" Nov 21 14:25:37 crc kubenswrapper[4897]: I1121 14:25:37.471881 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hqtth\" (UniqueName: \"kubernetes.io/projected/72cb5830-c763-4dca-a08f-d5ea7abb07fe-kube-api-access-hqtth\") pod \"frr-k8s-2v89k\" (UID: \"72cb5830-c763-4dca-a08f-d5ea7abb07fe\") " pod="metallb-system/frr-k8s-2v89k" Nov 21 14:25:37 crc kubenswrapper[4897]: I1121 14:25:37.472083 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4c46f\" (UniqueName: \"kubernetes.io/projected/dda4b910-8a41-4076-b92d-030e277f9778-kube-api-access-4c46f\") pod \"frr-k8s-webhook-server-6998585d5-x7jrs\" (UID: \"dda4b910-8a41-4076-b92d-030e277f9778\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-x7jrs" Nov 21 14:25:37 crc kubenswrapper[4897]: I1121 14:25:37.552891 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0c21212c-1adf-4bb1-8a7e-c1df5e98e0d8-metrics-certs\") pod \"controller-6c7b4b5f48-6bhsc\" (UID: \"0c21212c-1adf-4bb1-8a7e-c1df5e98e0d8\") " pod="metallb-system/controller-6c7b4b5f48-6bhsc" Nov 21 14:25:37 crc kubenswrapper[4897]: I1121 14:25:37.552946 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/89b1034c-5696-4975-a683-fdd9c163801f-metallb-excludel2\") pod \"speaker-hqq7g\" (UID: \"89b1034c-5696-4975-a683-fdd9c163801f\") " pod="metallb-system/speaker-hqq7g" Nov 21 14:25:37 crc kubenswrapper[4897]: I1121 14:25:37.552998 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/89b1034c-5696-4975-a683-fdd9c163801f-metrics-certs\") pod \"speaker-hqq7g\" (UID: \"89b1034c-5696-4975-a683-fdd9c163801f\") " pod="metallb-system/speaker-hqq7g" Nov 21 14:25:37 crc kubenswrapper[4897]: I1121 14:25:37.553021 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0c21212c-1adf-4bb1-8a7e-c1df5e98e0d8-cert\") pod \"controller-6c7b4b5f48-6bhsc\" (UID: \"0c21212c-1adf-4bb1-8a7e-c1df5e98e0d8\") " pod="metallb-system/controller-6c7b4b5f48-6bhsc" Nov 21 14:25:37 crc kubenswrapper[4897]: I1121 14:25:37.553095 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/89b1034c-5696-4975-a683-fdd9c163801f-memberlist\") pod \"speaker-hqq7g\" (UID: \"89b1034c-5696-4975-a683-fdd9c163801f\") " pod="metallb-system/speaker-hqq7g" Nov 21 14:25:37 crc kubenswrapper[4897]: E1121 14:25:37.553350 4897 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Nov 21 14:25:37 crc kubenswrapper[4897]: E1121 14:25:37.553452 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/89b1034c-5696-4975-a683-fdd9c163801f-memberlist podName:89b1034c-5696-4975-a683-fdd9c163801f nodeName:}" failed. No retries permitted until 2025-11-21 14:25:38.053430603 +0000 UTC m=+1015.338024078 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/89b1034c-5696-4975-a683-fdd9c163801f-memberlist") pod "speaker-hqq7g" (UID: "89b1034c-5696-4975-a683-fdd9c163801f") : secret "metallb-memberlist" not found Nov 21 14:25:37 crc kubenswrapper[4897]: I1121 14:25:37.553758 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rx2zv\" (UniqueName: \"kubernetes.io/projected/89b1034c-5696-4975-a683-fdd9c163801f-kube-api-access-rx2zv\") pod \"speaker-hqq7g\" (UID: \"89b1034c-5696-4975-a683-fdd9c163801f\") " pod="metallb-system/speaker-hqq7g" Nov 21 14:25:37 crc kubenswrapper[4897]: I1121 14:25:37.553895 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/89b1034c-5696-4975-a683-fdd9c163801f-metallb-excludel2\") pod \"speaker-hqq7g\" (UID: \"89b1034c-5696-4975-a683-fdd9c163801f\") " pod="metallb-system/speaker-hqq7g" Nov 21 14:25:37 crc kubenswrapper[4897]: I1121 14:25:37.554106 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fkrhd\" (UniqueName: \"kubernetes.io/projected/0c21212c-1adf-4bb1-8a7e-c1df5e98e0d8-kube-api-access-fkrhd\") pod \"controller-6c7b4b5f48-6bhsc\" (UID: \"0c21212c-1adf-4bb1-8a7e-c1df5e98e0d8\") " pod="metallb-system/controller-6c7b4b5f48-6bhsc" Nov 21 14:25:37 crc kubenswrapper[4897]: I1121 14:25:37.556013 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-6998585d5-x7jrs" Nov 21 14:25:37 crc kubenswrapper[4897]: I1121 14:25:37.556670 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/89b1034c-5696-4975-a683-fdd9c163801f-metrics-certs\") pod \"speaker-hqq7g\" (UID: \"89b1034c-5696-4975-a683-fdd9c163801f\") " pod="metallb-system/speaker-hqq7g" Nov 21 14:25:37 crc kubenswrapper[4897]: I1121 14:25:37.556847 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0c21212c-1adf-4bb1-8a7e-c1df5e98e0d8-cert\") pod \"controller-6c7b4b5f48-6bhsc\" (UID: \"0c21212c-1adf-4bb1-8a7e-c1df5e98e0d8\") " pod="metallb-system/controller-6c7b4b5f48-6bhsc" Nov 21 14:25:37 crc kubenswrapper[4897]: I1121 14:25:37.557330 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0c21212c-1adf-4bb1-8a7e-c1df5e98e0d8-metrics-certs\") pod \"controller-6c7b4b5f48-6bhsc\" (UID: \"0c21212c-1adf-4bb1-8a7e-c1df5e98e0d8\") " pod="metallb-system/controller-6c7b4b5f48-6bhsc" Nov 21 14:25:37 crc kubenswrapper[4897]: I1121 14:25:37.572090 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-2v89k" Nov 21 14:25:37 crc kubenswrapper[4897]: I1121 14:25:37.578401 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fkrhd\" (UniqueName: \"kubernetes.io/projected/0c21212c-1adf-4bb1-8a7e-c1df5e98e0d8-kube-api-access-fkrhd\") pod \"controller-6c7b4b5f48-6bhsc\" (UID: \"0c21212c-1adf-4bb1-8a7e-c1df5e98e0d8\") " pod="metallb-system/controller-6c7b4b5f48-6bhsc" Nov 21 14:25:37 crc kubenswrapper[4897]: I1121 14:25:37.583674 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rx2zv\" (UniqueName: \"kubernetes.io/projected/89b1034c-5696-4975-a683-fdd9c163801f-kube-api-access-rx2zv\") pod \"speaker-hqq7g\" (UID: \"89b1034c-5696-4975-a683-fdd9c163801f\") " pod="metallb-system/speaker-hqq7g" Nov 21 14:25:37 crc kubenswrapper[4897]: I1121 14:25:37.669828 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6c7b4b5f48-6bhsc" Nov 21 14:25:37 crc kubenswrapper[4897]: I1121 14:25:37.961611 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6c7b4b5f48-6bhsc"] Nov 21 14:25:37 crc kubenswrapper[4897]: W1121 14:25:37.965453 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0c21212c_1adf_4bb1_8a7e_c1df5e98e0d8.slice/crio-914c4fc6f4547738bd2a6d01f4a842b8554699c1b5b1637669bdd55f41511fe4 WatchSource:0}: Error finding container 914c4fc6f4547738bd2a6d01f4a842b8554699c1b5b1637669bdd55f41511fe4: Status 404 returned error can't find the container with id 914c4fc6f4547738bd2a6d01f4a842b8554699c1b5b1637669bdd55f41511fe4 Nov 21 14:25:38 crc kubenswrapper[4897]: I1121 14:25:38.035520 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-x7jrs"] Nov 21 14:25:38 crc kubenswrapper[4897]: I1121 14:25:38.068659 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/89b1034c-5696-4975-a683-fdd9c163801f-memberlist\") pod \"speaker-hqq7g\" (UID: \"89b1034c-5696-4975-a683-fdd9c163801f\") " pod="metallb-system/speaker-hqq7g" Nov 21 14:25:38 crc kubenswrapper[4897]: E1121 14:25:38.069002 4897 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Nov 21 14:25:38 crc kubenswrapper[4897]: E1121 14:25:38.069074 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/89b1034c-5696-4975-a683-fdd9c163801f-memberlist podName:89b1034c-5696-4975-a683-fdd9c163801f nodeName:}" failed. No retries permitted until 2025-11-21 14:25:39.069055965 +0000 UTC m=+1016.353649440 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/89b1034c-5696-4975-a683-fdd9c163801f-memberlist") pod "speaker-hqq7g" (UID: "89b1034c-5696-4975-a683-fdd9c163801f") : secret "metallb-memberlist" not found Nov 21 14:25:38 crc kubenswrapper[4897]: I1121 14:25:38.137007 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-2v89k" event={"ID":"72cb5830-c763-4dca-a08f-d5ea7abb07fe","Type":"ContainerStarted","Data":"931c799f0212713eea26ae1d83e77a622105acf65c2fc224f5b79b3fca98f6be"} Nov 21 14:25:38 crc kubenswrapper[4897]: I1121 14:25:38.139454 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-6bhsc" event={"ID":"0c21212c-1adf-4bb1-8a7e-c1df5e98e0d8","Type":"ContainerStarted","Data":"34426028a874f013208662e4e807afdf3d587ef0caab6dca67b0908efd0a9e57"} Nov 21 14:25:38 crc kubenswrapper[4897]: I1121 14:25:38.139886 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-6bhsc" event={"ID":"0c21212c-1adf-4bb1-8a7e-c1df5e98e0d8","Type":"ContainerStarted","Data":"914c4fc6f4547738bd2a6d01f4a842b8554699c1b5b1637669bdd55f41511fe4"} Nov 21 14:25:38 crc kubenswrapper[4897]: I1121 14:25:38.148573 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-6998585d5-x7jrs" event={"ID":"dda4b910-8a41-4076-b92d-030e277f9778","Type":"ContainerStarted","Data":"e30166b4eeb5c2f01232d28db41b50fcff1b13dc385d2208ad9da170f343347d"} Nov 21 14:25:39 crc kubenswrapper[4897]: I1121 14:25:39.090554 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/89b1034c-5696-4975-a683-fdd9c163801f-memberlist\") pod \"speaker-hqq7g\" (UID: \"89b1034c-5696-4975-a683-fdd9c163801f\") " pod="metallb-system/speaker-hqq7g" Nov 21 14:25:39 crc kubenswrapper[4897]: I1121 14:25:39.099799 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/89b1034c-5696-4975-a683-fdd9c163801f-memberlist\") pod \"speaker-hqq7g\" (UID: \"89b1034c-5696-4975-a683-fdd9c163801f\") " pod="metallb-system/speaker-hqq7g" Nov 21 14:25:39 crc kubenswrapper[4897]: I1121 14:25:39.132026 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-hqq7g" Nov 21 14:25:39 crc kubenswrapper[4897]: I1121 14:25:39.181567 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-6bhsc" event={"ID":"0c21212c-1adf-4bb1-8a7e-c1df5e98e0d8","Type":"ContainerStarted","Data":"18c9663082264710c0cdb72c17ef763e658474b3fa38d345bdaa2e100ed1c00f"} Nov 21 14:25:39 crc kubenswrapper[4897]: I1121 14:25:39.181763 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-6c7b4b5f48-6bhsc" Nov 21 14:25:39 crc kubenswrapper[4897]: W1121 14:25:39.199243 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod89b1034c_5696_4975_a683_fdd9c163801f.slice/crio-de4c6c3f2098052b120dad9e327fa9d825da97a13489a324cc08249aae5dba13 WatchSource:0}: Error finding container de4c6c3f2098052b120dad9e327fa9d825da97a13489a324cc08249aae5dba13: Status 404 returned error can't find the container with id de4c6c3f2098052b120dad9e327fa9d825da97a13489a324cc08249aae5dba13 Nov 21 14:25:39 crc kubenswrapper[4897]: I1121 14:25:39.211880 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-6c7b4b5f48-6bhsc" podStartSLOduration=2.211863286 podStartE2EDuration="2.211863286s" podCreationTimestamp="2025-11-21 14:25:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:25:39.204201278 +0000 UTC m=+1016.488794753" watchObservedRunningTime="2025-11-21 14:25:39.211863286 +0000 UTC m=+1016.496456761" Nov 21 14:25:40 crc kubenswrapper[4897]: I1121 14:25:40.190812 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-hqq7g" event={"ID":"89b1034c-5696-4975-a683-fdd9c163801f","Type":"ContainerStarted","Data":"ab6acf1f4dbbfe69c1c672255509d12dd0c402566ca82ccc8fb56b0408eb9ac7"} Nov 21 14:25:40 crc kubenswrapper[4897]: I1121 14:25:40.191152 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-hqq7g" event={"ID":"89b1034c-5696-4975-a683-fdd9c163801f","Type":"ContainerStarted","Data":"e9c9909bdb5a02e80fdb7f18a99b7362a032a4f591f5c0fcde3225a4db7fe9a8"} Nov 21 14:25:40 crc kubenswrapper[4897]: I1121 14:25:40.191164 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-hqq7g" event={"ID":"89b1034c-5696-4975-a683-fdd9c163801f","Type":"ContainerStarted","Data":"de4c6c3f2098052b120dad9e327fa9d825da97a13489a324cc08249aae5dba13"} Nov 21 14:25:40 crc kubenswrapper[4897]: I1121 14:25:40.191801 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-hqq7g" Nov 21 14:25:40 crc kubenswrapper[4897]: I1121 14:25:40.210573 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-hqq7g" podStartSLOduration=3.210553488 podStartE2EDuration="3.210553488s" podCreationTimestamp="2025-11-21 14:25:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:25:40.205327726 +0000 UTC m=+1017.489921201" watchObservedRunningTime="2025-11-21 14:25:40.210553488 +0000 UTC m=+1017.495146963" Nov 21 14:25:45 crc kubenswrapper[4897]: I1121 14:25:45.229468 4897 generic.go:334] "Generic (PLEG): container finished" podID="72cb5830-c763-4dca-a08f-d5ea7abb07fe" containerID="8f32b26f8c35cbd269fcc667673436f6230670e08776875b5ce7b2feb6d71f6f" exitCode=0 Nov 21 14:25:45 crc kubenswrapper[4897]: I1121 14:25:45.229588 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-2v89k" event={"ID":"72cb5830-c763-4dca-a08f-d5ea7abb07fe","Type":"ContainerDied","Data":"8f32b26f8c35cbd269fcc667673436f6230670e08776875b5ce7b2feb6d71f6f"} Nov 21 14:25:45 crc kubenswrapper[4897]: I1121 14:25:45.231631 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-6998585d5-x7jrs" event={"ID":"dda4b910-8a41-4076-b92d-030e277f9778","Type":"ContainerStarted","Data":"28a18d733b170369ad3e5453c9c41923d20085b63468058e06a22caf73b22a4b"} Nov 21 14:25:45 crc kubenswrapper[4897]: I1121 14:25:45.231783 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-6998585d5-x7jrs" Nov 21 14:25:45 crc kubenswrapper[4897]: I1121 14:25:45.279733 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-6998585d5-x7jrs" podStartSLOduration=1.550750078 podStartE2EDuration="8.279715223s" podCreationTimestamp="2025-11-21 14:25:37 +0000 UTC" firstStartedPulling="2025-11-21 14:25:38.044092439 +0000 UTC m=+1015.328685934" lastFinishedPulling="2025-11-21 14:25:44.773057604 +0000 UTC m=+1022.057651079" observedRunningTime="2025-11-21 14:25:45.273170816 +0000 UTC m=+1022.557764331" watchObservedRunningTime="2025-11-21 14:25:45.279715223 +0000 UTC m=+1022.564308698" Nov 21 14:25:46 crc kubenswrapper[4897]: I1121 14:25:46.250860 4897 generic.go:334] "Generic (PLEG): container finished" podID="72cb5830-c763-4dca-a08f-d5ea7abb07fe" containerID="83ae597841f8760ecca5424ea4979b7e80b76895c9c4025af59d24bb1955346d" exitCode=0 Nov 21 14:25:46 crc kubenswrapper[4897]: I1121 14:25:46.250936 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-2v89k" event={"ID":"72cb5830-c763-4dca-a08f-d5ea7abb07fe","Type":"ContainerDied","Data":"83ae597841f8760ecca5424ea4979b7e80b76895c9c4025af59d24bb1955346d"} Nov 21 14:25:47 crc kubenswrapper[4897]: I1121 14:25:47.264026 4897 generic.go:334] "Generic (PLEG): container finished" podID="72cb5830-c763-4dca-a08f-d5ea7abb07fe" containerID="f8dbf7d4ada497789dfb4f948db1b92a9d4594b1388833ed63db897ee68e7ddc" exitCode=0 Nov 21 14:25:47 crc kubenswrapper[4897]: I1121 14:25:47.264297 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-2v89k" event={"ID":"72cb5830-c763-4dca-a08f-d5ea7abb07fe","Type":"ContainerDied","Data":"f8dbf7d4ada497789dfb4f948db1b92a9d4594b1388833ed63db897ee68e7ddc"} Nov 21 14:25:48 crc kubenswrapper[4897]: I1121 14:25:48.282882 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-2v89k" event={"ID":"72cb5830-c763-4dca-a08f-d5ea7abb07fe","Type":"ContainerStarted","Data":"6ecbe3c949688a0cb2f4af36289f05654d0274aac3e90cceb3ba5cf732b35c99"} Nov 21 14:25:48 crc kubenswrapper[4897]: I1121 14:25:48.283520 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-2v89k" event={"ID":"72cb5830-c763-4dca-a08f-d5ea7abb07fe","Type":"ContainerStarted","Data":"2fed3765de3741c6265c258a46201ea95d0db399f54f4351b0c364a8e6f194d9"} Nov 21 14:25:48 crc kubenswrapper[4897]: I1121 14:25:48.283532 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-2v89k" event={"ID":"72cb5830-c763-4dca-a08f-d5ea7abb07fe","Type":"ContainerStarted","Data":"49935b438ad6476d3a201bc33871113171dad7b02035baf9e2de444c347e0634"} Nov 21 14:25:48 crc kubenswrapper[4897]: I1121 14:25:48.283541 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-2v89k" event={"ID":"72cb5830-c763-4dca-a08f-d5ea7abb07fe","Type":"ContainerStarted","Data":"a8aa593020db79b30d27bd38fe84db60a4cdfd5e3dfa03e46c2106ebc0867b1c"} Nov 21 14:25:48 crc kubenswrapper[4897]: I1121 14:25:48.283551 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-2v89k" event={"ID":"72cb5830-c763-4dca-a08f-d5ea7abb07fe","Type":"ContainerStarted","Data":"bfe3cab75eef93af989ce24c15e4509f39600ded54c888859066141b60a25247"} Nov 21 14:25:49 crc kubenswrapper[4897]: I1121 14:25:49.143211 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-hqq7g" Nov 21 14:25:49 crc kubenswrapper[4897]: I1121 14:25:49.294981 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-2v89k" event={"ID":"72cb5830-c763-4dca-a08f-d5ea7abb07fe","Type":"ContainerStarted","Data":"45d003bc14f699c60aa6f8bcdc13a5e972c8b8101db32240587a8e1ae4c5b60b"} Nov 21 14:25:49 crc kubenswrapper[4897]: I1121 14:25:49.309591 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-2v89k" Nov 21 14:25:49 crc kubenswrapper[4897]: I1121 14:25:49.339130 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-2v89k" podStartSLOduration=5.307602007 podStartE2EDuration="12.339112186s" podCreationTimestamp="2025-11-21 14:25:37 +0000 UTC" firstStartedPulling="2025-11-21 14:25:37.762576225 +0000 UTC m=+1015.047169700" lastFinishedPulling="2025-11-21 14:25:44.794086404 +0000 UTC m=+1022.078679879" observedRunningTime="2025-11-21 14:25:49.334256114 +0000 UTC m=+1026.618849609" watchObservedRunningTime="2025-11-21 14:25:49.339112186 +0000 UTC m=+1026.623705661" Nov 21 14:25:52 crc kubenswrapper[4897]: I1121 14:25:52.268384 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-xkfhh"] Nov 21 14:25:52 crc kubenswrapper[4897]: I1121 14:25:52.275107 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-xkfhh" Nov 21 14:25:52 crc kubenswrapper[4897]: I1121 14:25:52.276415 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-xkfhh"] Nov 21 14:25:52 crc kubenswrapper[4897]: I1121 14:25:52.280173 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Nov 21 14:25:52 crc kubenswrapper[4897]: I1121 14:25:52.280629 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Nov 21 14:25:52 crc kubenswrapper[4897]: I1121 14:25:52.280923 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-7jqh5" Nov 21 14:25:52 crc kubenswrapper[4897]: I1121 14:25:52.445342 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ssp2q\" (UniqueName: \"kubernetes.io/projected/62d19f58-73cb-4e00-8edf-742f01724e86-kube-api-access-ssp2q\") pod \"openstack-operator-index-xkfhh\" (UID: \"62d19f58-73cb-4e00-8edf-742f01724e86\") " pod="openstack-operators/openstack-operator-index-xkfhh" Nov 21 14:25:52 crc kubenswrapper[4897]: I1121 14:25:52.547579 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ssp2q\" (UniqueName: \"kubernetes.io/projected/62d19f58-73cb-4e00-8edf-742f01724e86-kube-api-access-ssp2q\") pod \"openstack-operator-index-xkfhh\" (UID: \"62d19f58-73cb-4e00-8edf-742f01724e86\") " pod="openstack-operators/openstack-operator-index-xkfhh" Nov 21 14:25:52 crc kubenswrapper[4897]: I1121 14:25:52.567709 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ssp2q\" (UniqueName: \"kubernetes.io/projected/62d19f58-73cb-4e00-8edf-742f01724e86-kube-api-access-ssp2q\") pod \"openstack-operator-index-xkfhh\" (UID: \"62d19f58-73cb-4e00-8edf-742f01724e86\") " pod="openstack-operators/openstack-operator-index-xkfhh" Nov 21 14:25:52 crc kubenswrapper[4897]: I1121 14:25:52.573386 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-2v89k" Nov 21 14:25:52 crc kubenswrapper[4897]: I1121 14:25:52.597628 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-xkfhh" Nov 21 14:25:52 crc kubenswrapper[4897]: I1121 14:25:52.611329 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-2v89k" Nov 21 14:25:53 crc kubenswrapper[4897]: I1121 14:25:53.080543 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-xkfhh"] Nov 21 14:25:53 crc kubenswrapper[4897]: W1121 14:25:53.088679 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod62d19f58_73cb_4e00_8edf_742f01724e86.slice/crio-7b0026e38ef5fc1b4b8435d43ae197d70a7ed288c0f57fb389614157335eb0cc WatchSource:0}: Error finding container 7b0026e38ef5fc1b4b8435d43ae197d70a7ed288c0f57fb389614157335eb0cc: Status 404 returned error can't find the container with id 7b0026e38ef5fc1b4b8435d43ae197d70a7ed288c0f57fb389614157335eb0cc Nov 21 14:25:53 crc kubenswrapper[4897]: I1121 14:25:53.325737 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-xkfhh" event={"ID":"62d19f58-73cb-4e00-8edf-742f01724e86","Type":"ContainerStarted","Data":"7b0026e38ef5fc1b4b8435d43ae197d70a7ed288c0f57fb389614157335eb0cc"} Nov 21 14:25:55 crc kubenswrapper[4897]: I1121 14:25:55.643556 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-xkfhh"] Nov 21 14:25:56 crc kubenswrapper[4897]: I1121 14:25:56.247005 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-5m2wq"] Nov 21 14:25:56 crc kubenswrapper[4897]: I1121 14:25:56.248116 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-5m2wq" Nov 21 14:25:56 crc kubenswrapper[4897]: I1121 14:25:56.262881 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-5m2wq"] Nov 21 14:25:56 crc kubenswrapper[4897]: I1121 14:25:56.423044 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s5d98\" (UniqueName: \"kubernetes.io/projected/e3e3e93c-5c28-4d6c-be4d-ccd252435642-kube-api-access-s5d98\") pod \"openstack-operator-index-5m2wq\" (UID: \"e3e3e93c-5c28-4d6c-be4d-ccd252435642\") " pod="openstack-operators/openstack-operator-index-5m2wq" Nov 21 14:25:56 crc kubenswrapper[4897]: I1121 14:25:56.524832 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s5d98\" (UniqueName: \"kubernetes.io/projected/e3e3e93c-5c28-4d6c-be4d-ccd252435642-kube-api-access-s5d98\") pod \"openstack-operator-index-5m2wq\" (UID: \"e3e3e93c-5c28-4d6c-be4d-ccd252435642\") " pod="openstack-operators/openstack-operator-index-5m2wq" Nov 21 14:25:56 crc kubenswrapper[4897]: I1121 14:25:56.543214 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s5d98\" (UniqueName: \"kubernetes.io/projected/e3e3e93c-5c28-4d6c-be4d-ccd252435642-kube-api-access-s5d98\") pod \"openstack-operator-index-5m2wq\" (UID: \"e3e3e93c-5c28-4d6c-be4d-ccd252435642\") " pod="openstack-operators/openstack-operator-index-5m2wq" Nov 21 14:25:56 crc kubenswrapper[4897]: I1121 14:25:56.572316 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-5m2wq" Nov 21 14:25:57 crc kubenswrapper[4897]: I1121 14:25:57.564754 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-6998585d5-x7jrs" Nov 21 14:25:57 crc kubenswrapper[4897]: I1121 14:25:57.575017 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-2v89k" Nov 21 14:25:57 crc kubenswrapper[4897]: I1121 14:25:57.680044 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-6c7b4b5f48-6bhsc" Nov 21 14:25:58 crc kubenswrapper[4897]: I1121 14:25:58.363556 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-xkfhh" event={"ID":"62d19f58-73cb-4e00-8edf-742f01724e86","Type":"ContainerStarted","Data":"286dc2cb151f5787a06650c1e6a4a059178ddc1b7d5d5986f25236de442028e8"} Nov 21 14:25:58 crc kubenswrapper[4897]: I1121 14:25:58.363844 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-xkfhh" podUID="62d19f58-73cb-4e00-8edf-742f01724e86" containerName="registry-server" containerID="cri-o://286dc2cb151f5787a06650c1e6a4a059178ddc1b7d5d5986f25236de442028e8" gracePeriod=2 Nov 21 14:25:58 crc kubenswrapper[4897]: I1121 14:25:58.379246 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-xkfhh" podStartSLOduration=1.327271325 podStartE2EDuration="6.379226234s" podCreationTimestamp="2025-11-21 14:25:52 +0000 UTC" firstStartedPulling="2025-11-21 14:25:53.090568007 +0000 UTC m=+1030.375161492" lastFinishedPulling="2025-11-21 14:25:58.142522926 +0000 UTC m=+1035.427116401" observedRunningTime="2025-11-21 14:25:58.378137725 +0000 UTC m=+1035.662731210" watchObservedRunningTime="2025-11-21 14:25:58.379226234 +0000 UTC m=+1035.663819709" Nov 21 14:25:58 crc kubenswrapper[4897]: I1121 14:25:58.506359 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-5m2wq"] Nov 21 14:25:58 crc kubenswrapper[4897]: W1121 14:25:58.522677 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode3e3e93c_5c28_4d6c_be4d_ccd252435642.slice/crio-42dbb217674479b9d1e8bd8cb9f8c2c7c8701e7a839fd04e47068a6070eb8595 WatchSource:0}: Error finding container 42dbb217674479b9d1e8bd8cb9f8c2c7c8701e7a839fd04e47068a6070eb8595: Status 404 returned error can't find the container with id 42dbb217674479b9d1e8bd8cb9f8c2c7c8701e7a839fd04e47068a6070eb8595 Nov 21 14:25:58 crc kubenswrapper[4897]: I1121 14:25:58.705287 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-xkfhh" Nov 21 14:25:58 crc kubenswrapper[4897]: I1121 14:25:58.861495 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ssp2q\" (UniqueName: \"kubernetes.io/projected/62d19f58-73cb-4e00-8edf-742f01724e86-kube-api-access-ssp2q\") pod \"62d19f58-73cb-4e00-8edf-742f01724e86\" (UID: \"62d19f58-73cb-4e00-8edf-742f01724e86\") " Nov 21 14:25:58 crc kubenswrapper[4897]: I1121 14:25:58.869405 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62d19f58-73cb-4e00-8edf-742f01724e86-kube-api-access-ssp2q" (OuterVolumeSpecName: "kube-api-access-ssp2q") pod "62d19f58-73cb-4e00-8edf-742f01724e86" (UID: "62d19f58-73cb-4e00-8edf-742f01724e86"). InnerVolumeSpecName "kube-api-access-ssp2q". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:25:58 crc kubenswrapper[4897]: I1121 14:25:58.963612 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ssp2q\" (UniqueName: \"kubernetes.io/projected/62d19f58-73cb-4e00-8edf-742f01724e86-kube-api-access-ssp2q\") on node \"crc\" DevicePath \"\"" Nov 21 14:25:59 crc kubenswrapper[4897]: I1121 14:25:59.372082 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-5m2wq" event={"ID":"e3e3e93c-5c28-4d6c-be4d-ccd252435642","Type":"ContainerStarted","Data":"5936e85d467f0b75ae4fe31d48e321aba05660ca36d00e1a88e90f898ac82d38"} Nov 21 14:25:59 crc kubenswrapper[4897]: I1121 14:25:59.372129 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-5m2wq" event={"ID":"e3e3e93c-5c28-4d6c-be4d-ccd252435642","Type":"ContainerStarted","Data":"42dbb217674479b9d1e8bd8cb9f8c2c7c8701e7a839fd04e47068a6070eb8595"} Nov 21 14:25:59 crc kubenswrapper[4897]: I1121 14:25:59.373496 4897 generic.go:334] "Generic (PLEG): container finished" podID="62d19f58-73cb-4e00-8edf-742f01724e86" containerID="286dc2cb151f5787a06650c1e6a4a059178ddc1b7d5d5986f25236de442028e8" exitCode=0 Nov 21 14:25:59 crc kubenswrapper[4897]: I1121 14:25:59.373586 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-xkfhh" Nov 21 14:25:59 crc kubenswrapper[4897]: I1121 14:25:59.373871 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-xkfhh" event={"ID":"62d19f58-73cb-4e00-8edf-742f01724e86","Type":"ContainerDied","Data":"286dc2cb151f5787a06650c1e6a4a059178ddc1b7d5d5986f25236de442028e8"} Nov 21 14:25:59 crc kubenswrapper[4897]: I1121 14:25:59.374074 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-xkfhh" event={"ID":"62d19f58-73cb-4e00-8edf-742f01724e86","Type":"ContainerDied","Data":"7b0026e38ef5fc1b4b8435d43ae197d70a7ed288c0f57fb389614157335eb0cc"} Nov 21 14:25:59 crc kubenswrapper[4897]: I1121 14:25:59.374175 4897 scope.go:117] "RemoveContainer" containerID="286dc2cb151f5787a06650c1e6a4a059178ddc1b7d5d5986f25236de442028e8" Nov 21 14:25:59 crc kubenswrapper[4897]: I1121 14:25:59.391316 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-5m2wq" podStartSLOduration=3.340637145 podStartE2EDuration="3.391286019s" podCreationTimestamp="2025-11-21 14:25:56 +0000 UTC" firstStartedPulling="2025-11-21 14:25:58.527164416 +0000 UTC m=+1035.811757891" lastFinishedPulling="2025-11-21 14:25:58.57781329 +0000 UTC m=+1035.862406765" observedRunningTime="2025-11-21 14:25:59.390088557 +0000 UTC m=+1036.674682042" watchObservedRunningTime="2025-11-21 14:25:59.391286019 +0000 UTC m=+1036.675879514" Nov 21 14:25:59 crc kubenswrapper[4897]: I1121 14:25:59.398116 4897 scope.go:117] "RemoveContainer" containerID="286dc2cb151f5787a06650c1e6a4a059178ddc1b7d5d5986f25236de442028e8" Nov 21 14:25:59 crc kubenswrapper[4897]: E1121 14:25:59.398985 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"286dc2cb151f5787a06650c1e6a4a059178ddc1b7d5d5986f25236de442028e8\": container with ID starting with 286dc2cb151f5787a06650c1e6a4a059178ddc1b7d5d5986f25236de442028e8 not found: ID does not exist" containerID="286dc2cb151f5787a06650c1e6a4a059178ddc1b7d5d5986f25236de442028e8" Nov 21 14:25:59 crc kubenswrapper[4897]: I1121 14:25:59.399128 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"286dc2cb151f5787a06650c1e6a4a059178ddc1b7d5d5986f25236de442028e8"} err="failed to get container status \"286dc2cb151f5787a06650c1e6a4a059178ddc1b7d5d5986f25236de442028e8\": rpc error: code = NotFound desc = could not find container \"286dc2cb151f5787a06650c1e6a4a059178ddc1b7d5d5986f25236de442028e8\": container with ID starting with 286dc2cb151f5787a06650c1e6a4a059178ddc1b7d5d5986f25236de442028e8 not found: ID does not exist" Nov 21 14:25:59 crc kubenswrapper[4897]: I1121 14:25:59.416428 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-xkfhh"] Nov 21 14:25:59 crc kubenswrapper[4897]: I1121 14:25:59.424739 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-xkfhh"] Nov 21 14:26:00 crc kubenswrapper[4897]: I1121 14:26:00.102031 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="62d19f58-73cb-4e00-8edf-742f01724e86" path="/var/lib/kubelet/pods/62d19f58-73cb-4e00-8edf-742f01724e86/volumes" Nov 21 14:26:06 crc kubenswrapper[4897]: I1121 14:26:06.572803 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-5m2wq" Nov 21 14:26:06 crc kubenswrapper[4897]: I1121 14:26:06.573185 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-5m2wq" Nov 21 14:26:06 crc kubenswrapper[4897]: I1121 14:26:06.609102 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-5m2wq" Nov 21 14:26:07 crc kubenswrapper[4897]: I1121 14:26:07.480909 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-5m2wq" Nov 21 14:26:14 crc kubenswrapper[4897]: I1121 14:26:14.102365 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ec5a3b0aec190a4ce66bffd4e725be5945d1a7630fe83e1312e4dba1febxxx9"] Nov 21 14:26:14 crc kubenswrapper[4897]: E1121 14:26:14.103193 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62d19f58-73cb-4e00-8edf-742f01724e86" containerName="registry-server" Nov 21 14:26:14 crc kubenswrapper[4897]: I1121 14:26:14.103208 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="62d19f58-73cb-4e00-8edf-742f01724e86" containerName="registry-server" Nov 21 14:26:14 crc kubenswrapper[4897]: I1121 14:26:14.103436 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="62d19f58-73cb-4e00-8edf-742f01724e86" containerName="registry-server" Nov 21 14:26:14 crc kubenswrapper[4897]: I1121 14:26:14.104999 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ec5a3b0aec190a4ce66bffd4e725be5945d1a7630fe83e1312e4dba1febxxx9" Nov 21 14:26:14 crc kubenswrapper[4897]: I1121 14:26:14.113475 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-b6k7w" Nov 21 14:26:14 crc kubenswrapper[4897]: I1121 14:26:14.115462 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ec5a3b0aec190a4ce66bffd4e725be5945d1a7630fe83e1312e4dba1febxxx9"] Nov 21 14:26:14 crc kubenswrapper[4897]: I1121 14:26:14.233873 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v5f9n\" (UniqueName: \"kubernetes.io/projected/25764720-62c7-44db-8112-225a13aeb7fa-kube-api-access-v5f9n\") pod \"ec5a3b0aec190a4ce66bffd4e725be5945d1a7630fe83e1312e4dba1febxxx9\" (UID: \"25764720-62c7-44db-8112-225a13aeb7fa\") " pod="openstack-operators/ec5a3b0aec190a4ce66bffd4e725be5945d1a7630fe83e1312e4dba1febxxx9" Nov 21 14:26:14 crc kubenswrapper[4897]: I1121 14:26:14.234461 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/25764720-62c7-44db-8112-225a13aeb7fa-util\") pod \"ec5a3b0aec190a4ce66bffd4e725be5945d1a7630fe83e1312e4dba1febxxx9\" (UID: \"25764720-62c7-44db-8112-225a13aeb7fa\") " pod="openstack-operators/ec5a3b0aec190a4ce66bffd4e725be5945d1a7630fe83e1312e4dba1febxxx9" Nov 21 14:26:14 crc kubenswrapper[4897]: I1121 14:26:14.234651 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/25764720-62c7-44db-8112-225a13aeb7fa-bundle\") pod \"ec5a3b0aec190a4ce66bffd4e725be5945d1a7630fe83e1312e4dba1febxxx9\" (UID: \"25764720-62c7-44db-8112-225a13aeb7fa\") " pod="openstack-operators/ec5a3b0aec190a4ce66bffd4e725be5945d1a7630fe83e1312e4dba1febxxx9" Nov 21 14:26:14 crc kubenswrapper[4897]: I1121 14:26:14.336032 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v5f9n\" (UniqueName: \"kubernetes.io/projected/25764720-62c7-44db-8112-225a13aeb7fa-kube-api-access-v5f9n\") pod \"ec5a3b0aec190a4ce66bffd4e725be5945d1a7630fe83e1312e4dba1febxxx9\" (UID: \"25764720-62c7-44db-8112-225a13aeb7fa\") " pod="openstack-operators/ec5a3b0aec190a4ce66bffd4e725be5945d1a7630fe83e1312e4dba1febxxx9" Nov 21 14:26:14 crc kubenswrapper[4897]: I1121 14:26:14.336118 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/25764720-62c7-44db-8112-225a13aeb7fa-util\") pod \"ec5a3b0aec190a4ce66bffd4e725be5945d1a7630fe83e1312e4dba1febxxx9\" (UID: \"25764720-62c7-44db-8112-225a13aeb7fa\") " pod="openstack-operators/ec5a3b0aec190a4ce66bffd4e725be5945d1a7630fe83e1312e4dba1febxxx9" Nov 21 14:26:14 crc kubenswrapper[4897]: I1121 14:26:14.336153 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/25764720-62c7-44db-8112-225a13aeb7fa-bundle\") pod \"ec5a3b0aec190a4ce66bffd4e725be5945d1a7630fe83e1312e4dba1febxxx9\" (UID: \"25764720-62c7-44db-8112-225a13aeb7fa\") " pod="openstack-operators/ec5a3b0aec190a4ce66bffd4e725be5945d1a7630fe83e1312e4dba1febxxx9" Nov 21 14:26:14 crc kubenswrapper[4897]: I1121 14:26:14.336650 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/25764720-62c7-44db-8112-225a13aeb7fa-bundle\") pod \"ec5a3b0aec190a4ce66bffd4e725be5945d1a7630fe83e1312e4dba1febxxx9\" (UID: \"25764720-62c7-44db-8112-225a13aeb7fa\") " pod="openstack-operators/ec5a3b0aec190a4ce66bffd4e725be5945d1a7630fe83e1312e4dba1febxxx9" Nov 21 14:26:14 crc kubenswrapper[4897]: I1121 14:26:14.337190 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/25764720-62c7-44db-8112-225a13aeb7fa-util\") pod \"ec5a3b0aec190a4ce66bffd4e725be5945d1a7630fe83e1312e4dba1febxxx9\" (UID: \"25764720-62c7-44db-8112-225a13aeb7fa\") " pod="openstack-operators/ec5a3b0aec190a4ce66bffd4e725be5945d1a7630fe83e1312e4dba1febxxx9" Nov 21 14:26:14 crc kubenswrapper[4897]: I1121 14:26:14.357841 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v5f9n\" (UniqueName: \"kubernetes.io/projected/25764720-62c7-44db-8112-225a13aeb7fa-kube-api-access-v5f9n\") pod \"ec5a3b0aec190a4ce66bffd4e725be5945d1a7630fe83e1312e4dba1febxxx9\" (UID: \"25764720-62c7-44db-8112-225a13aeb7fa\") " pod="openstack-operators/ec5a3b0aec190a4ce66bffd4e725be5945d1a7630fe83e1312e4dba1febxxx9" Nov 21 14:26:14 crc kubenswrapper[4897]: I1121 14:26:14.430534 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ec5a3b0aec190a4ce66bffd4e725be5945d1a7630fe83e1312e4dba1febxxx9" Nov 21 14:26:14 crc kubenswrapper[4897]: I1121 14:26:14.920622 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ec5a3b0aec190a4ce66bffd4e725be5945d1a7630fe83e1312e4dba1febxxx9"] Nov 21 14:26:15 crc kubenswrapper[4897]: I1121 14:26:15.512808 4897 generic.go:334] "Generic (PLEG): container finished" podID="25764720-62c7-44db-8112-225a13aeb7fa" containerID="f663d868c16c3fc7f940ae63997dfd1c2c77deb25b99c880085dea48737d5856" exitCode=0 Nov 21 14:26:15 crc kubenswrapper[4897]: I1121 14:26:15.512863 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ec5a3b0aec190a4ce66bffd4e725be5945d1a7630fe83e1312e4dba1febxxx9" event={"ID":"25764720-62c7-44db-8112-225a13aeb7fa","Type":"ContainerDied","Data":"f663d868c16c3fc7f940ae63997dfd1c2c77deb25b99c880085dea48737d5856"} Nov 21 14:26:15 crc kubenswrapper[4897]: I1121 14:26:15.513104 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ec5a3b0aec190a4ce66bffd4e725be5945d1a7630fe83e1312e4dba1febxxx9" event={"ID":"25764720-62c7-44db-8112-225a13aeb7fa","Type":"ContainerStarted","Data":"8019e1957da0428cd5e0491d7161f46f570780ca39c85e69b9869897d6473c7b"} Nov 21 14:26:16 crc kubenswrapper[4897]: I1121 14:26:16.521794 4897 generic.go:334] "Generic (PLEG): container finished" podID="25764720-62c7-44db-8112-225a13aeb7fa" containerID="413bd9030189a8935d8f7bb8392c6579e26811198e0c35a6a41d2ec250ca4c3d" exitCode=0 Nov 21 14:26:16 crc kubenswrapper[4897]: I1121 14:26:16.521866 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ec5a3b0aec190a4ce66bffd4e725be5945d1a7630fe83e1312e4dba1febxxx9" event={"ID":"25764720-62c7-44db-8112-225a13aeb7fa","Type":"ContainerDied","Data":"413bd9030189a8935d8f7bb8392c6579e26811198e0c35a6a41d2ec250ca4c3d"} Nov 21 14:26:17 crc kubenswrapper[4897]: I1121 14:26:17.537665 4897 generic.go:334] "Generic (PLEG): container finished" podID="25764720-62c7-44db-8112-225a13aeb7fa" containerID="21b0191a9a3ab8d4a8aef2acc862dad23b518b0a7237f318dc815e6bcadf0b2e" exitCode=0 Nov 21 14:26:17 crc kubenswrapper[4897]: I1121 14:26:17.537751 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ec5a3b0aec190a4ce66bffd4e725be5945d1a7630fe83e1312e4dba1febxxx9" event={"ID":"25764720-62c7-44db-8112-225a13aeb7fa","Type":"ContainerDied","Data":"21b0191a9a3ab8d4a8aef2acc862dad23b518b0a7237f318dc815e6bcadf0b2e"} Nov 21 14:26:18 crc kubenswrapper[4897]: I1121 14:26:18.896471 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ec5a3b0aec190a4ce66bffd4e725be5945d1a7630fe83e1312e4dba1febxxx9" Nov 21 14:26:19 crc kubenswrapper[4897]: I1121 14:26:19.048240 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/25764720-62c7-44db-8112-225a13aeb7fa-util\") pod \"25764720-62c7-44db-8112-225a13aeb7fa\" (UID: \"25764720-62c7-44db-8112-225a13aeb7fa\") " Nov 21 14:26:19 crc kubenswrapper[4897]: I1121 14:26:19.048385 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v5f9n\" (UniqueName: \"kubernetes.io/projected/25764720-62c7-44db-8112-225a13aeb7fa-kube-api-access-v5f9n\") pod \"25764720-62c7-44db-8112-225a13aeb7fa\" (UID: \"25764720-62c7-44db-8112-225a13aeb7fa\") " Nov 21 14:26:19 crc kubenswrapper[4897]: I1121 14:26:19.048547 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/25764720-62c7-44db-8112-225a13aeb7fa-bundle\") pod \"25764720-62c7-44db-8112-225a13aeb7fa\" (UID: \"25764720-62c7-44db-8112-225a13aeb7fa\") " Nov 21 14:26:19 crc kubenswrapper[4897]: I1121 14:26:19.049073 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/25764720-62c7-44db-8112-225a13aeb7fa-bundle" (OuterVolumeSpecName: "bundle") pod "25764720-62c7-44db-8112-225a13aeb7fa" (UID: "25764720-62c7-44db-8112-225a13aeb7fa"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:26:19 crc kubenswrapper[4897]: I1121 14:26:19.054146 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25764720-62c7-44db-8112-225a13aeb7fa-kube-api-access-v5f9n" (OuterVolumeSpecName: "kube-api-access-v5f9n") pod "25764720-62c7-44db-8112-225a13aeb7fa" (UID: "25764720-62c7-44db-8112-225a13aeb7fa"). InnerVolumeSpecName "kube-api-access-v5f9n". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:26:19 crc kubenswrapper[4897]: I1121 14:26:19.064203 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/25764720-62c7-44db-8112-225a13aeb7fa-util" (OuterVolumeSpecName: "util") pod "25764720-62c7-44db-8112-225a13aeb7fa" (UID: "25764720-62c7-44db-8112-225a13aeb7fa"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:26:19 crc kubenswrapper[4897]: I1121 14:26:19.150684 4897 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/25764720-62c7-44db-8112-225a13aeb7fa-bundle\") on node \"crc\" DevicePath \"\"" Nov 21 14:26:19 crc kubenswrapper[4897]: I1121 14:26:19.150710 4897 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/25764720-62c7-44db-8112-225a13aeb7fa-util\") on node \"crc\" DevicePath \"\"" Nov 21 14:26:19 crc kubenswrapper[4897]: I1121 14:26:19.150721 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v5f9n\" (UniqueName: \"kubernetes.io/projected/25764720-62c7-44db-8112-225a13aeb7fa-kube-api-access-v5f9n\") on node \"crc\" DevicePath \"\"" Nov 21 14:26:19 crc kubenswrapper[4897]: I1121 14:26:19.562658 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ec5a3b0aec190a4ce66bffd4e725be5945d1a7630fe83e1312e4dba1febxxx9" event={"ID":"25764720-62c7-44db-8112-225a13aeb7fa","Type":"ContainerDied","Data":"8019e1957da0428cd5e0491d7161f46f570780ca39c85e69b9869897d6473c7b"} Nov 21 14:26:19 crc kubenswrapper[4897]: I1121 14:26:19.562704 4897 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8019e1957da0428cd5e0491d7161f46f570780ca39c85e69b9869897d6473c7b" Nov 21 14:26:19 crc kubenswrapper[4897]: I1121 14:26:19.562775 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ec5a3b0aec190a4ce66bffd4e725be5945d1a7630fe83e1312e4dba1febxxx9" Nov 21 14:26:30 crc kubenswrapper[4897]: I1121 14:26:30.155766 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-59d9f856bd-24lrb"] Nov 21 14:26:30 crc kubenswrapper[4897]: E1121 14:26:30.156467 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25764720-62c7-44db-8112-225a13aeb7fa" containerName="extract" Nov 21 14:26:30 crc kubenswrapper[4897]: I1121 14:26:30.156477 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="25764720-62c7-44db-8112-225a13aeb7fa" containerName="extract" Nov 21 14:26:30 crc kubenswrapper[4897]: E1121 14:26:30.156497 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25764720-62c7-44db-8112-225a13aeb7fa" containerName="util" Nov 21 14:26:30 crc kubenswrapper[4897]: I1121 14:26:30.156507 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="25764720-62c7-44db-8112-225a13aeb7fa" containerName="util" Nov 21 14:26:30 crc kubenswrapper[4897]: E1121 14:26:30.156597 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25764720-62c7-44db-8112-225a13aeb7fa" containerName="pull" Nov 21 14:26:30 crc kubenswrapper[4897]: I1121 14:26:30.156603 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="25764720-62c7-44db-8112-225a13aeb7fa" containerName="pull" Nov 21 14:26:30 crc kubenswrapper[4897]: I1121 14:26:30.156773 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="25764720-62c7-44db-8112-225a13aeb7fa" containerName="extract" Nov 21 14:26:30 crc kubenswrapper[4897]: I1121 14:26:30.157547 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-59d9f856bd-24lrb" Nov 21 14:26:30 crc kubenswrapper[4897]: I1121 14:26:30.160755 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-6sknr" Nov 21 14:26:30 crc kubenswrapper[4897]: I1121 14:26:30.232208 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7shkf\" (UniqueName: \"kubernetes.io/projected/b3fd9236-baa7-41fe-a03b-35eb59ecb132-kube-api-access-7shkf\") pod \"openstack-operator-controller-operator-59d9f856bd-24lrb\" (UID: \"b3fd9236-baa7-41fe-a03b-35eb59ecb132\") " pod="openstack-operators/openstack-operator-controller-operator-59d9f856bd-24lrb" Nov 21 14:26:30 crc kubenswrapper[4897]: I1121 14:26:30.271533 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-59d9f856bd-24lrb"] Nov 21 14:26:30 crc kubenswrapper[4897]: I1121 14:26:30.334396 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7shkf\" (UniqueName: \"kubernetes.io/projected/b3fd9236-baa7-41fe-a03b-35eb59ecb132-kube-api-access-7shkf\") pod \"openstack-operator-controller-operator-59d9f856bd-24lrb\" (UID: \"b3fd9236-baa7-41fe-a03b-35eb59ecb132\") " pod="openstack-operators/openstack-operator-controller-operator-59d9f856bd-24lrb" Nov 21 14:26:30 crc kubenswrapper[4897]: I1121 14:26:30.354315 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7shkf\" (UniqueName: \"kubernetes.io/projected/b3fd9236-baa7-41fe-a03b-35eb59ecb132-kube-api-access-7shkf\") pod \"openstack-operator-controller-operator-59d9f856bd-24lrb\" (UID: \"b3fd9236-baa7-41fe-a03b-35eb59ecb132\") " pod="openstack-operators/openstack-operator-controller-operator-59d9f856bd-24lrb" Nov 21 14:26:30 crc kubenswrapper[4897]: I1121 14:26:30.475548 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-59d9f856bd-24lrb" Nov 21 14:26:30 crc kubenswrapper[4897]: I1121 14:26:30.930558 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-59d9f856bd-24lrb"] Nov 21 14:26:31 crc kubenswrapper[4897]: I1121 14:26:31.659231 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-59d9f856bd-24lrb" event={"ID":"b3fd9236-baa7-41fe-a03b-35eb59ecb132","Type":"ContainerStarted","Data":"f05d65259afe6ceedcd7c8c565af4802a8a9a07ff58d7b61018fda956d240b1a"} Nov 21 14:26:36 crc kubenswrapper[4897]: I1121 14:26:36.697073 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-59d9f856bd-24lrb" event={"ID":"b3fd9236-baa7-41fe-a03b-35eb59ecb132","Type":"ContainerStarted","Data":"ff51f4c51d4f8f4008a44b72f72f869e635ff33c86925392f38bda20de72adaf"} Nov 21 14:26:38 crc kubenswrapper[4897]: I1121 14:26:38.718729 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-59d9f856bd-24lrb" event={"ID":"b3fd9236-baa7-41fe-a03b-35eb59ecb132","Type":"ContainerStarted","Data":"0266dcde31a190d5db90f3c767cdffe5317ff5795d92b168a6f8a0bae6ef8914"} Nov 21 14:26:38 crc kubenswrapper[4897]: I1121 14:26:38.719292 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-59d9f856bd-24lrb" Nov 21 14:26:38 crc kubenswrapper[4897]: I1121 14:26:38.751403 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-59d9f856bd-24lrb" podStartSLOduration=1.652763923 podStartE2EDuration="8.751379022s" podCreationTimestamp="2025-11-21 14:26:30 +0000 UTC" firstStartedPulling="2025-11-21 14:26:30.950998833 +0000 UTC m=+1068.235592348" lastFinishedPulling="2025-11-21 14:26:38.049613972 +0000 UTC m=+1075.334207447" observedRunningTime="2025-11-21 14:26:38.741782203 +0000 UTC m=+1076.026375688" watchObservedRunningTime="2025-11-21 14:26:38.751379022 +0000 UTC m=+1076.035972497" Nov 21 14:26:40 crc kubenswrapper[4897]: I1121 14:26:40.478610 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-59d9f856bd-24lrb" Nov 21 14:27:12 crc kubenswrapper[4897]: I1121 14:27:12.398059 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-75fb479bcc-cr58g"] Nov 21 14:27:12 crc kubenswrapper[4897]: I1121 14:27:12.400062 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-cr58g" Nov 21 14:27:12 crc kubenswrapper[4897]: I1121 14:27:12.402390 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-7tjkr" Nov 21 14:27:12 crc kubenswrapper[4897]: I1121 14:27:12.409915 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-75fb479bcc-cr58g"] Nov 21 14:27:12 crc kubenswrapper[4897]: I1121 14:27:12.417401 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6498cbf48f-kq4ss"] Nov 21 14:27:12 crc kubenswrapper[4897]: I1121 14:27:12.419191 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-kq4ss" Nov 21 14:27:12 crc kubenswrapper[4897]: I1121 14:27:12.420857 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-nzh55" Nov 21 14:27:12 crc kubenswrapper[4897]: I1121 14:27:12.439591 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6498cbf48f-kq4ss"] Nov 21 14:27:12 crc kubenswrapper[4897]: I1121 14:27:12.478708 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-767ccfd65f-csgnr"] Nov 21 14:27:12 crc kubenswrapper[4897]: I1121 14:27:12.483905 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-csgnr" Nov 21 14:27:12 crc kubenswrapper[4897]: I1121 14:27:12.499253 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-7969689c84-xlbhh"] Nov 21 14:27:12 crc kubenswrapper[4897]: I1121 14:27:12.502185 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-rdhdw" Nov 21 14:27:12 crc kubenswrapper[4897]: I1121 14:27:12.541943 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-7969689c84-xlbhh" Nov 21 14:27:12 crc kubenswrapper[4897]: I1121 14:27:12.542741 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85qxx\" (UniqueName: \"kubernetes.io/projected/e3891c15-c0b9-4965-84fe-05262fe71853-kube-api-access-85qxx\") pod \"barbican-operator-controller-manager-75fb479bcc-cr58g\" (UID: \"e3891c15-c0b9-4965-84fe-05262fe71853\") " pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-cr58g" Nov 21 14:27:12 crc kubenswrapper[4897]: I1121 14:27:12.544248 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-767ccfd65f-csgnr"] Nov 21 14:27:12 crc kubenswrapper[4897]: I1121 14:27:12.547797 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-ckwj5" Nov 21 14:27:12 crc kubenswrapper[4897]: I1121 14:27:12.562623 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-7969689c84-xlbhh"] Nov 21 14:27:12 crc kubenswrapper[4897]: I1121 14:27:12.590903 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-56f54d6746-nrp4j"] Nov 21 14:27:12 crc kubenswrapper[4897]: I1121 14:27:12.592570 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-56f54d6746-nrp4j" Nov 21 14:27:12 crc kubenswrapper[4897]: I1121 14:27:12.598571 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-rwbmv" Nov 21 14:27:12 crc kubenswrapper[4897]: I1121 14:27:12.617648 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-598f69df5d-l2njs"] Nov 21 14:27:12 crc kubenswrapper[4897]: I1121 14:27:12.619114 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-l2njs" Nov 21 14:27:12 crc kubenswrapper[4897]: I1121 14:27:12.629993 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-nsw87" Nov 21 14:27:12 crc kubenswrapper[4897]: I1121 14:27:12.642842 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-56f54d6746-nrp4j"] Nov 21 14:27:12 crc kubenswrapper[4897]: I1121 14:27:12.644192 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-85qxx\" (UniqueName: \"kubernetes.io/projected/e3891c15-c0b9-4965-84fe-05262fe71853-kube-api-access-85qxx\") pod \"barbican-operator-controller-manager-75fb479bcc-cr58g\" (UID: \"e3891c15-c0b9-4965-84fe-05262fe71853\") " pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-cr58g" Nov 21 14:27:12 crc kubenswrapper[4897]: I1121 14:27:12.644233 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5hrv8\" (UniqueName: \"kubernetes.io/projected/39c184ba-7360-4d31-99f5-239711efdf91-kube-api-access-5hrv8\") pod \"cinder-operator-controller-manager-6498cbf48f-kq4ss\" (UID: \"39c184ba-7360-4d31-99f5-239711efdf91\") " pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-kq4ss" Nov 21 14:27:12 crc kubenswrapper[4897]: I1121 14:27:12.644255 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cjcbl\" (UniqueName: \"kubernetes.io/projected/717d3738-bb33-4a91-ad36-7cf849ae783f-kube-api-access-cjcbl\") pod \"designate-operator-controller-manager-767ccfd65f-csgnr\" (UID: \"717d3738-bb33-4a91-ad36-7cf849ae783f\") " pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-csgnr" Nov 21 14:27:12 crc kubenswrapper[4897]: I1121 14:27:12.644314 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c9wcx\" (UniqueName: \"kubernetes.io/projected/55f66c84-22f9-4840-83b5-624cce4d5543-kube-api-access-c9wcx\") pod \"glance-operator-controller-manager-7969689c84-xlbhh\" (UID: \"55f66c84-22f9-4840-83b5-624cce4d5543\") " pod="openstack-operators/glance-operator-controller-manager-7969689c84-xlbhh" Nov 21 14:27:12 crc kubenswrapper[4897]: I1121 14:27:12.654583 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-598f69df5d-l2njs"] Nov 21 14:27:12 crc kubenswrapper[4897]: I1121 14:27:12.659198 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-6dd8864d7c-pt5vd"] Nov 21 14:27:12 crc kubenswrapper[4897]: I1121 14:27:12.660760 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-pt5vd" Nov 21 14:27:12 crc kubenswrapper[4897]: I1121 14:27:12.664479 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Nov 21 14:27:12 crc kubenswrapper[4897]: I1121 14:27:12.664955 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-6dd8864d7c-pt5vd"] Nov 21 14:27:12 crc kubenswrapper[4897]: I1121 14:27:12.664995 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-tgw9w" Nov 21 14:27:12 crc kubenswrapper[4897]: I1121 14:27:12.673228 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-99b499f4-phs97"] Nov 21 14:27:12 crc kubenswrapper[4897]: I1121 14:27:12.674671 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-phs97" Nov 21 14:27:12 crc kubenswrapper[4897]: I1121 14:27:12.677054 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-j2j8r" Nov 21 14:27:12 crc kubenswrapper[4897]: I1121 14:27:12.692370 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-85qxx\" (UniqueName: \"kubernetes.io/projected/e3891c15-c0b9-4965-84fe-05262fe71853-kube-api-access-85qxx\") pod \"barbican-operator-controller-manager-75fb479bcc-cr58g\" (UID: \"e3891c15-c0b9-4965-84fe-05262fe71853\") " pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-cr58g" Nov 21 14:27:12 crc kubenswrapper[4897]: I1121 14:27:12.692436 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7454b96578-gzzl8"] Nov 21 14:27:12 crc kubenswrapper[4897]: I1121 14:27:12.693785 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-gzzl8" Nov 21 14:27:12 crc kubenswrapper[4897]: I1121 14:27:12.695972 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-k22gw" Nov 21 14:27:12 crc kubenswrapper[4897]: I1121 14:27:12.706739 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-99b499f4-phs97"] Nov 21 14:27:12 crc kubenswrapper[4897]: I1121 14:27:12.715164 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7454b96578-gzzl8"] Nov 21 14:27:12 crc kubenswrapper[4897]: I1121 14:27:12.719993 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-58f887965d-dhc5w"] Nov 21 14:27:12 crc kubenswrapper[4897]: I1121 14:27:12.722136 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-58f887965d-dhc5w" Nov 21 14:27:12 crc kubenswrapper[4897]: I1121 14:27:12.722153 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-cr58g" Nov 21 14:27:12 crc kubenswrapper[4897]: I1121 14:27:12.724615 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-8vw8c" Nov 21 14:27:12 crc kubenswrapper[4897]: I1121 14:27:12.739730 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-58f887965d-dhc5w"] Nov 21 14:27:12 crc kubenswrapper[4897]: I1121 14:27:12.745981 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5hrv8\" (UniqueName: \"kubernetes.io/projected/39c184ba-7360-4d31-99f5-239711efdf91-kube-api-access-5hrv8\") pod \"cinder-operator-controller-manager-6498cbf48f-kq4ss\" (UID: \"39c184ba-7360-4d31-99f5-239711efdf91\") " pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-kq4ss" Nov 21 14:27:12 crc kubenswrapper[4897]: I1121 14:27:12.746030 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cjcbl\" (UniqueName: \"kubernetes.io/projected/717d3738-bb33-4a91-ad36-7cf849ae783f-kube-api-access-cjcbl\") pod \"designate-operator-controller-manager-767ccfd65f-csgnr\" (UID: \"717d3738-bb33-4a91-ad36-7cf849ae783f\") " pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-csgnr" Nov 21 14:27:12 crc kubenswrapper[4897]: I1121 14:27:12.746117 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c9wcx\" (UniqueName: \"kubernetes.io/projected/55f66c84-22f9-4840-83b5-624cce4d5543-kube-api-access-c9wcx\") pod \"glance-operator-controller-manager-7969689c84-xlbhh\" (UID: \"55f66c84-22f9-4840-83b5-624cce4d5543\") " pod="openstack-operators/glance-operator-controller-manager-7969689c84-xlbhh" Nov 21 14:27:12 crc kubenswrapper[4897]: I1121 14:27:12.746256 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hl682\" (UniqueName: \"kubernetes.io/projected/8aee4e2c-4b47-41ec-9953-7e73d0bd6524-kube-api-access-hl682\") pod \"horizon-operator-controller-manager-598f69df5d-l2njs\" (UID: \"8aee4e2c-4b47-41ec-9953-7e73d0bd6524\") " pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-l2njs" Nov 21 14:27:12 crc kubenswrapper[4897]: I1121 14:27:12.746287 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qn5fn\" (UniqueName: \"kubernetes.io/projected/65770e5f-95ee-4501-993d-8d5a015da258-kube-api-access-qn5fn\") pod \"heat-operator-controller-manager-56f54d6746-nrp4j\" (UID: \"65770e5f-95ee-4501-993d-8d5a015da258\") " pod="openstack-operators/heat-operator-controller-manager-56f54d6746-nrp4j" Nov 21 14:27:12 crc kubenswrapper[4897]: I1121 14:27:12.782416 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5hrv8\" (UniqueName: \"kubernetes.io/projected/39c184ba-7360-4d31-99f5-239711efdf91-kube-api-access-5hrv8\") pod \"cinder-operator-controller-manager-6498cbf48f-kq4ss\" (UID: \"39c184ba-7360-4d31-99f5-239711efdf91\") " pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-kq4ss" Nov 21 14:27:12 crc kubenswrapper[4897]: I1121 14:27:12.795239 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c9wcx\" (UniqueName: \"kubernetes.io/projected/55f66c84-22f9-4840-83b5-624cce4d5543-kube-api-access-c9wcx\") pod \"glance-operator-controller-manager-7969689c84-xlbhh\" (UID: \"55f66c84-22f9-4840-83b5-624cce4d5543\") " pod="openstack-operators/glance-operator-controller-manager-7969689c84-xlbhh" Nov 21 14:27:12 crc kubenswrapper[4897]: I1121 14:27:12.795492 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cjcbl\" (UniqueName: \"kubernetes.io/projected/717d3738-bb33-4a91-ad36-7cf849ae783f-kube-api-access-cjcbl\") pod \"designate-operator-controller-manager-767ccfd65f-csgnr\" (UID: \"717d3738-bb33-4a91-ad36-7cf849ae783f\") " pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-csgnr" Nov 21 14:27:12 crc kubenswrapper[4897]: I1121 14:27:12.815873 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-78bd47f458-2grhg"] Nov 21 14:27:12 crc kubenswrapper[4897]: I1121 14:27:12.817622 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-2grhg" Nov 21 14:27:12 crc kubenswrapper[4897]: I1121 14:27:12.822854 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-7ksnf" Nov 21 14:27:12 crc kubenswrapper[4897]: I1121 14:27:12.834582 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-54b5986bb8-2dnk2"] Nov 21 14:27:12 crc kubenswrapper[4897]: I1121 14:27:12.836036 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-2dnk2" Nov 21 14:27:12 crc kubenswrapper[4897]: I1121 14:27:12.845491 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-7hmxg" Nov 21 14:27:12 crc kubenswrapper[4897]: I1121 14:27:12.847706 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kqtsp\" (UniqueName: \"kubernetes.io/projected/282cd160-d81e-4e20-866f-dee909546aa2-kube-api-access-kqtsp\") pod \"manila-operator-controller-manager-58f887965d-dhc5w\" (UID: \"282cd160-d81e-4e20-866f-dee909546aa2\") " pod="openstack-operators/manila-operator-controller-manager-58f887965d-dhc5w" Nov 21 14:27:12 crc kubenswrapper[4897]: I1121 14:27:12.847755 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j4xfp\" (UniqueName: \"kubernetes.io/projected/a7d8ae48-9dd4-40d0-bab5-26e3cff24b20-kube-api-access-j4xfp\") pod \"infra-operator-controller-manager-6dd8864d7c-pt5vd\" (UID: \"a7d8ae48-9dd4-40d0-bab5-26e3cff24b20\") " pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-pt5vd" Nov 21 14:27:12 crc kubenswrapper[4897]: I1121 14:27:12.847785 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-88c4f\" (UniqueName: \"kubernetes.io/projected/d1fc2027-5a14-4de3-8c0e-c4ecee66c40b-kube-api-access-88c4f\") pod \"ironic-operator-controller-manager-99b499f4-phs97\" (UID: \"d1fc2027-5a14-4de3-8c0e-c4ecee66c40b\") " pod="openstack-operators/ironic-operator-controller-manager-99b499f4-phs97" Nov 21 14:27:12 crc kubenswrapper[4897]: I1121 14:27:12.847815 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a7d8ae48-9dd4-40d0-bab5-26e3cff24b20-cert\") pod \"infra-operator-controller-manager-6dd8864d7c-pt5vd\" (UID: \"a7d8ae48-9dd4-40d0-bab5-26e3cff24b20\") " pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-pt5vd" Nov 21 14:27:12 crc kubenswrapper[4897]: I1121 14:27:12.847874 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pb4rl\" (UniqueName: \"kubernetes.io/projected/d4c39700-beb8-4272-9f54-1c501332af0d-kube-api-access-pb4rl\") pod \"keystone-operator-controller-manager-7454b96578-gzzl8\" (UID: \"d4c39700-beb8-4272-9f54-1c501332af0d\") " pod="openstack-operators/keystone-operator-controller-manager-7454b96578-gzzl8" Nov 21 14:27:12 crc kubenswrapper[4897]: I1121 14:27:12.847961 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hl682\" (UniqueName: \"kubernetes.io/projected/8aee4e2c-4b47-41ec-9953-7e73d0bd6524-kube-api-access-hl682\") pod \"horizon-operator-controller-manager-598f69df5d-l2njs\" (UID: \"8aee4e2c-4b47-41ec-9953-7e73d0bd6524\") " pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-l2njs" Nov 21 14:27:12 crc kubenswrapper[4897]: I1121 14:27:12.847991 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qn5fn\" (UniqueName: \"kubernetes.io/projected/65770e5f-95ee-4501-993d-8d5a015da258-kube-api-access-qn5fn\") pod \"heat-operator-controller-manager-56f54d6746-nrp4j\" (UID: \"65770e5f-95ee-4501-993d-8d5a015da258\") " pod="openstack-operators/heat-operator-controller-manager-56f54d6746-nrp4j" Nov 21 14:27:12 crc kubenswrapper[4897]: I1121 14:27:12.853286 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-csgnr" Nov 21 14:27:12 crc kubenswrapper[4897]: I1121 14:27:12.863924 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-7969689c84-xlbhh" Nov 21 14:27:12 crc kubenswrapper[4897]: I1121 14:27:12.865459 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-78bd47f458-2grhg"] Nov 21 14:27:12 crc kubenswrapper[4897]: I1121 14:27:12.868153 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qn5fn\" (UniqueName: \"kubernetes.io/projected/65770e5f-95ee-4501-993d-8d5a015da258-kube-api-access-qn5fn\") pod \"heat-operator-controller-manager-56f54d6746-nrp4j\" (UID: \"65770e5f-95ee-4501-993d-8d5a015da258\") " pod="openstack-operators/heat-operator-controller-manager-56f54d6746-nrp4j" Nov 21 14:27:12 crc kubenswrapper[4897]: I1121 14:27:12.885860 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-54b5986bb8-2dnk2"] Nov 21 14:27:12 crc kubenswrapper[4897]: I1121 14:27:12.894200 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hl682\" (UniqueName: \"kubernetes.io/projected/8aee4e2c-4b47-41ec-9953-7e73d0bd6524-kube-api-access-hl682\") pod \"horizon-operator-controller-manager-598f69df5d-l2njs\" (UID: \"8aee4e2c-4b47-41ec-9953-7e73d0bd6524\") " pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-l2njs" Nov 21 14:27:12 crc kubenswrapper[4897]: I1121 14:27:12.901165 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-cfbb9c588-bh4rl"] Nov 21 14:27:12 crc kubenswrapper[4897]: I1121 14:27:12.902802 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-bh4rl" Nov 21 14:27:12 crc kubenswrapper[4897]: I1121 14:27:12.923872 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-2sqhj"] Nov 21 14:27:12 crc kubenswrapper[4897]: I1121 14:27:12.924584 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-fvvtd" Nov 21 14:27:12 crc kubenswrapper[4897]: I1121 14:27:12.926208 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-2sqhj" Nov 21 14:27:12 crc kubenswrapper[4897]: I1121 14:27:12.927446 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-56f54d6746-nrp4j" Nov 21 14:27:12 crc kubenswrapper[4897]: I1121 14:27:12.931083 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-7gqkp" Nov 21 14:27:12 crc kubenswrapper[4897]: I1121 14:27:12.950252 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pb4rl\" (UniqueName: \"kubernetes.io/projected/d4c39700-beb8-4272-9f54-1c501332af0d-kube-api-access-pb4rl\") pod \"keystone-operator-controller-manager-7454b96578-gzzl8\" (UID: \"d4c39700-beb8-4272-9f54-1c501332af0d\") " pod="openstack-operators/keystone-operator-controller-manager-7454b96578-gzzl8" Nov 21 14:27:12 crc kubenswrapper[4897]: I1121 14:27:12.950341 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xccf8\" (UniqueName: \"kubernetes.io/projected/3b3d91dd-852c-4b65-a2c9-dbc02e79daaa-kube-api-access-xccf8\") pod \"mariadb-operator-controller-manager-54b5986bb8-2dnk2\" (UID: \"3b3d91dd-852c-4b65-a2c9-dbc02e79daaa\") " pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-2dnk2" Nov 21 14:27:12 crc kubenswrapper[4897]: I1121 14:27:12.950421 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xv2s8\" (UniqueName: \"kubernetes.io/projected/77f4c339-9ff1-4943-922d-88d81a73a1f1-kube-api-access-xv2s8\") pod \"neutron-operator-controller-manager-78bd47f458-2grhg\" (UID: \"77f4c339-9ff1-4943-922d-88d81a73a1f1\") " pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-2grhg" Nov 21 14:27:12 crc kubenswrapper[4897]: I1121 14:27:12.950478 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kqtsp\" (UniqueName: \"kubernetes.io/projected/282cd160-d81e-4e20-866f-dee909546aa2-kube-api-access-kqtsp\") pod \"manila-operator-controller-manager-58f887965d-dhc5w\" (UID: \"282cd160-d81e-4e20-866f-dee909546aa2\") " pod="openstack-operators/manila-operator-controller-manager-58f887965d-dhc5w" Nov 21 14:27:12 crc kubenswrapper[4897]: I1121 14:27:12.950564 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j4xfp\" (UniqueName: \"kubernetes.io/projected/a7d8ae48-9dd4-40d0-bab5-26e3cff24b20-kube-api-access-j4xfp\") pod \"infra-operator-controller-manager-6dd8864d7c-pt5vd\" (UID: \"a7d8ae48-9dd4-40d0-bab5-26e3cff24b20\") " pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-pt5vd" Nov 21 14:27:12 crc kubenswrapper[4897]: I1121 14:27:12.950592 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-88c4f\" (UniqueName: \"kubernetes.io/projected/d1fc2027-5a14-4de3-8c0e-c4ecee66c40b-kube-api-access-88c4f\") pod \"ironic-operator-controller-manager-99b499f4-phs97\" (UID: \"d1fc2027-5a14-4de3-8c0e-c4ecee66c40b\") " pod="openstack-operators/ironic-operator-controller-manager-99b499f4-phs97" Nov 21 14:27:12 crc kubenswrapper[4897]: I1121 14:27:12.950638 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a7d8ae48-9dd4-40d0-bab5-26e3cff24b20-cert\") pod \"infra-operator-controller-manager-6dd8864d7c-pt5vd\" (UID: \"a7d8ae48-9dd4-40d0-bab5-26e3cff24b20\") " pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-pt5vd" Nov 21 14:27:12 crc kubenswrapper[4897]: I1121 14:27:12.956352 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a7d8ae48-9dd4-40d0-bab5-26e3cff24b20-cert\") pod \"infra-operator-controller-manager-6dd8864d7c-pt5vd\" (UID: \"a7d8ae48-9dd4-40d0-bab5-26e3cff24b20\") " pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-pt5vd" Nov 21 14:27:12 crc kubenswrapper[4897]: I1121 14:27:12.960478 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-cfbb9c588-bh4rl"] Nov 21 14:27:12 crc kubenswrapper[4897]: I1121 14:27:12.963271 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-l2njs" Nov 21 14:27:12 crc kubenswrapper[4897]: I1121 14:27:12.974398 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-2sqhj"] Nov 21 14:27:12 crc kubenswrapper[4897]: I1121 14:27:12.977140 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kqtsp\" (UniqueName: \"kubernetes.io/projected/282cd160-d81e-4e20-866f-dee909546aa2-kube-api-access-kqtsp\") pod \"manila-operator-controller-manager-58f887965d-dhc5w\" (UID: \"282cd160-d81e-4e20-866f-dee909546aa2\") " pod="openstack-operators/manila-operator-controller-manager-58f887965d-dhc5w" Nov 21 14:27:12 crc kubenswrapper[4897]: I1121 14:27:12.988318 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pb4rl\" (UniqueName: \"kubernetes.io/projected/d4c39700-beb8-4272-9f54-1c501332af0d-kube-api-access-pb4rl\") pod \"keystone-operator-controller-manager-7454b96578-gzzl8\" (UID: \"d4c39700-beb8-4272-9f54-1c501332af0d\") " pod="openstack-operators/keystone-operator-controller-manager-7454b96578-gzzl8" Nov 21 14:27:12 crc kubenswrapper[4897]: I1121 14:27:12.993610 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j4xfp\" (UniqueName: \"kubernetes.io/projected/a7d8ae48-9dd4-40d0-bab5-26e3cff24b20-kube-api-access-j4xfp\") pod \"infra-operator-controller-manager-6dd8864d7c-pt5vd\" (UID: \"a7d8ae48-9dd4-40d0-bab5-26e3cff24b20\") " pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-pt5vd" Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.026239 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-lfbrg"] Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.027535 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-lfbrg" Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.029804 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-88c4f\" (UniqueName: \"kubernetes.io/projected/d1fc2027-5a14-4de3-8c0e-c4ecee66c40b-kube-api-access-88c4f\") pod \"ironic-operator-controller-manager-99b499f4-phs97\" (UID: \"d1fc2027-5a14-4de3-8c0e-c4ecee66c40b\") " pod="openstack-operators/ironic-operator-controller-manager-99b499f4-phs97" Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.039986 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-gzzl8" Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.046457 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-f8rx4" Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.047462 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.047603 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-kq4ss" Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.052672 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jnh5z\" (UniqueName: \"kubernetes.io/projected/499cc09f-fafe-47c2-87ac-bc00d79e949e-kube-api-access-jnh5z\") pod \"nova-operator-controller-manager-cfbb9c588-bh4rl\" (UID: \"499cc09f-fafe-47c2-87ac-bc00d79e949e\") " pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-bh4rl" Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.052734 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2rjbr\" (UniqueName: \"kubernetes.io/projected/81f3265d-f571-47a2-9ad8-b7cdf952d850-kube-api-access-2rjbr\") pod \"octavia-operator-controller-manager-54cfbf4c7d-2sqhj\" (UID: \"81f3265d-f571-47a2-9ad8-b7cdf952d850\") " pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-2sqhj" Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.052794 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xccf8\" (UniqueName: \"kubernetes.io/projected/3b3d91dd-852c-4b65-a2c9-dbc02e79daaa-kube-api-access-xccf8\") pod \"mariadb-operator-controller-manager-54b5986bb8-2dnk2\" (UID: \"3b3d91dd-852c-4b65-a2c9-dbc02e79daaa\") " pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-2dnk2" Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.052908 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xv2s8\" (UniqueName: \"kubernetes.io/projected/77f4c339-9ff1-4943-922d-88d81a73a1f1-kube-api-access-xv2s8\") pod \"neutron-operator-controller-manager-78bd47f458-2grhg\" (UID: \"77f4c339-9ff1-4943-922d-88d81a73a1f1\") " pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-2grhg" Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.084212 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-54fc5f65b7-glmx7"] Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.091398 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-glmx7" Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.092750 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xccf8\" (UniqueName: \"kubernetes.io/projected/3b3d91dd-852c-4b65-a2c9-dbc02e79daaa-kube-api-access-xccf8\") pod \"mariadb-operator-controller-manager-54b5986bb8-2dnk2\" (UID: \"3b3d91dd-852c-4b65-a2c9-dbc02e79daaa\") " pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-2dnk2" Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.096991 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-7w78j" Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.104875 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xv2s8\" (UniqueName: \"kubernetes.io/projected/77f4c339-9ff1-4943-922d-88d81a73a1f1-kube-api-access-xv2s8\") pod \"neutron-operator-controller-manager-78bd47f458-2grhg\" (UID: \"77f4c339-9ff1-4943-922d-88d81a73a1f1\") " pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-2grhg" Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.148267 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-58f887965d-dhc5w" Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.158662 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2rjbr\" (UniqueName: \"kubernetes.io/projected/81f3265d-f571-47a2-9ad8-b7cdf952d850-kube-api-access-2rjbr\") pod \"octavia-operator-controller-manager-54cfbf4c7d-2sqhj\" (UID: \"81f3265d-f571-47a2-9ad8-b7cdf952d850\") " pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-2sqhj" Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.158811 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jnh5z\" (UniqueName: \"kubernetes.io/projected/499cc09f-fafe-47c2-87ac-bc00d79e949e-kube-api-access-jnh5z\") pod \"nova-operator-controller-manager-cfbb9c588-bh4rl\" (UID: \"499cc09f-fafe-47c2-87ac-bc00d79e949e\") " pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-bh4rl" Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.158837 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7952fa01-945e-4491-87aa-af0835a97532-cert\") pod \"openstack-baremetal-operator-controller-manager-8c7444f48-lfbrg\" (UID: \"7952fa01-945e-4491-87aa-af0835a97532\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-lfbrg" Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.158856 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-htc9d\" (UniqueName: \"kubernetes.io/projected/7952fa01-945e-4491-87aa-af0835a97532-kube-api-access-htc9d\") pod \"openstack-baremetal-operator-controller-manager-8c7444f48-lfbrg\" (UID: \"7952fa01-945e-4491-87aa-af0835a97532\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-lfbrg" Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.159294 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-5b797b8dff-smlb7"] Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.160600 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-smlb7" Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.171170 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-269bm" Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.186624 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-54fc5f65b7-glmx7"] Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.197725 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2rjbr\" (UniqueName: \"kubernetes.io/projected/81f3265d-f571-47a2-9ad8-b7cdf952d850-kube-api-access-2rjbr\") pod \"octavia-operator-controller-manager-54cfbf4c7d-2sqhj\" (UID: \"81f3265d-f571-47a2-9ad8-b7cdf952d850\") " pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-2sqhj" Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.207080 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-2grhg" Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.213186 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jnh5z\" (UniqueName: \"kubernetes.io/projected/499cc09f-fafe-47c2-87ac-bc00d79e949e-kube-api-access-jnh5z\") pod \"nova-operator-controller-manager-cfbb9c588-bh4rl\" (UID: \"499cc09f-fafe-47c2-87ac-bc00d79e949e\") " pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-bh4rl" Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.217938 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-2dnk2" Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.227905 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-5b797b8dff-smlb7"] Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.234902 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-bh4rl" Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.235247 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-lfbrg"] Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.260402 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-d656998f4-p8zts"] Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.261944 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-d656998f4-p8zts" Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.263078 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-2sqhj" Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.266756 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-bw9rn" Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.285348 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gvkmg\" (UniqueName: \"kubernetes.io/projected/4b3dcab3-52d5-4f1d-bde6-2ae9803a0a8f-kube-api-access-gvkmg\") pod \"placement-operator-controller-manager-5b797b8dff-smlb7\" (UID: \"4b3dcab3-52d5-4f1d-bde6-2ae9803a0a8f\") " pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-smlb7" Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.285517 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7952fa01-945e-4491-87aa-af0835a97532-cert\") pod \"openstack-baremetal-operator-controller-manager-8c7444f48-lfbrg\" (UID: \"7952fa01-945e-4491-87aa-af0835a97532\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-lfbrg" Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.285537 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-htc9d\" (UniqueName: \"kubernetes.io/projected/7952fa01-945e-4491-87aa-af0835a97532-kube-api-access-htc9d\") pod \"openstack-baremetal-operator-controller-manager-8c7444f48-lfbrg\" (UID: \"7952fa01-945e-4491-87aa-af0835a97532\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-lfbrg" Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.285657 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jpwp5\" (UniqueName: \"kubernetes.io/projected/09df9b2c-a507-4a35-b646-a2dc45e1eb71-kube-api-access-jpwp5\") pod \"ovn-operator-controller-manager-54fc5f65b7-glmx7\" (UID: \"09df9b2c-a507-4a35-b646-a2dc45e1eb71\") " pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-glmx7" Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.286600 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-pt5vd" Nov 21 14:27:13 crc kubenswrapper[4897]: E1121 14:27:13.287133 4897 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 21 14:27:13 crc kubenswrapper[4897]: E1121 14:27:13.287194 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7952fa01-945e-4491-87aa-af0835a97532-cert podName:7952fa01-945e-4491-87aa-af0835a97532 nodeName:}" failed. No retries permitted until 2025-11-21 14:27:13.787177112 +0000 UTC m=+1111.071770587 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/7952fa01-945e-4491-87aa-af0835a97532-cert") pod "openstack-baremetal-operator-controller-manager-8c7444f48-lfbrg" (UID: "7952fa01-945e-4491-87aa-af0835a97532") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.313196 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-htc9d\" (UniqueName: \"kubernetes.io/projected/7952fa01-945e-4491-87aa-af0835a97532-kube-api-access-htc9d\") pod \"openstack-baremetal-operator-controller-manager-8c7444f48-lfbrg\" (UID: \"7952fa01-945e-4491-87aa-af0835a97532\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-lfbrg" Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.332582 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-77797865fc-9n8nt"] Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.334016 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-77797865fc-9n8nt" Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.336333 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-phs97" Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.339936 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-99jdk" Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.375831 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-d656998f4-p8zts"] Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.392720 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lj29c\" (UniqueName: \"kubernetes.io/projected/1557d794-d052-44ad-b908-4766d1abe234-kube-api-access-lj29c\") pod \"telemetry-operator-controller-manager-77797865fc-9n8nt\" (UID: \"1557d794-d052-44ad-b908-4766d1abe234\") " pod="openstack-operators/telemetry-operator-controller-manager-77797865fc-9n8nt" Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.392807 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jpwp5\" (UniqueName: \"kubernetes.io/projected/09df9b2c-a507-4a35-b646-a2dc45e1eb71-kube-api-access-jpwp5\") pod \"ovn-operator-controller-manager-54fc5f65b7-glmx7\" (UID: \"09df9b2c-a507-4a35-b646-a2dc45e1eb71\") " pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-glmx7" Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.392867 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-52dqw\" (UniqueName: \"kubernetes.io/projected/a031ee6b-00fb-4795-9e6d-6e21068c405e-kube-api-access-52dqw\") pod \"swift-operator-controller-manager-d656998f4-p8zts\" (UID: \"a031ee6b-00fb-4795-9e6d-6e21068c405e\") " pod="openstack-operators/swift-operator-controller-manager-d656998f4-p8zts" Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.392896 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gvkmg\" (UniqueName: \"kubernetes.io/projected/4b3dcab3-52d5-4f1d-bde6-2ae9803a0a8f-kube-api-access-gvkmg\") pod \"placement-operator-controller-manager-5b797b8dff-smlb7\" (UID: \"4b3dcab3-52d5-4f1d-bde6-2ae9803a0a8f\") " pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-smlb7" Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.452205 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gvkmg\" (UniqueName: \"kubernetes.io/projected/4b3dcab3-52d5-4f1d-bde6-2ae9803a0a8f-kube-api-access-gvkmg\") pod \"placement-operator-controller-manager-5b797b8dff-smlb7\" (UID: \"4b3dcab3-52d5-4f1d-bde6-2ae9803a0a8f\") " pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-smlb7" Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.458465 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jpwp5\" (UniqueName: \"kubernetes.io/projected/09df9b2c-a507-4a35-b646-a2dc45e1eb71-kube-api-access-jpwp5\") pod \"ovn-operator-controller-manager-54fc5f65b7-glmx7\" (UID: \"09df9b2c-a507-4a35-b646-a2dc45e1eb71\") " pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-glmx7" Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.492361 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-smlb7" Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.506216 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-52dqw\" (UniqueName: \"kubernetes.io/projected/a031ee6b-00fb-4795-9e6d-6e21068c405e-kube-api-access-52dqw\") pod \"swift-operator-controller-manager-d656998f4-p8zts\" (UID: \"a031ee6b-00fb-4795-9e6d-6e21068c405e\") " pod="openstack-operators/swift-operator-controller-manager-d656998f4-p8zts" Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.506688 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lj29c\" (UniqueName: \"kubernetes.io/projected/1557d794-d052-44ad-b908-4766d1abe234-kube-api-access-lj29c\") pod \"telemetry-operator-controller-manager-77797865fc-9n8nt\" (UID: \"1557d794-d052-44ad-b908-4766d1abe234\") " pod="openstack-operators/telemetry-operator-controller-manager-77797865fc-9n8nt" Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.522389 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-77797865fc-9n8nt"] Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.539525 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lj29c\" (UniqueName: \"kubernetes.io/projected/1557d794-d052-44ad-b908-4766d1abe234-kube-api-access-lj29c\") pod \"telemetry-operator-controller-manager-77797865fc-9n8nt\" (UID: \"1557d794-d052-44ad-b908-4766d1abe234\") " pod="openstack-operators/telemetry-operator-controller-manager-77797865fc-9n8nt" Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.542339 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-52dqw\" (UniqueName: \"kubernetes.io/projected/a031ee6b-00fb-4795-9e6d-6e21068c405e-kube-api-access-52dqw\") pod \"swift-operator-controller-manager-d656998f4-p8zts\" (UID: \"a031ee6b-00fb-4795-9e6d-6e21068c405e\") " pod="openstack-operators/swift-operator-controller-manager-d656998f4-p8zts" Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.573279 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-b4c496f69-9fkd7"] Nov 21 14:27:13 crc kubenswrapper[4897]: W1121 14:27:13.589656 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode3891c15_c0b9_4965_84fe_05262fe71853.slice/crio-f8a5353c95aac3ac71cb54285cfc09d8cad6d35807c980e8899f5f89e72e1674 WatchSource:0}: Error finding container f8a5353c95aac3ac71cb54285cfc09d8cad6d35807c980e8899f5f89e72e1674: Status 404 returned error can't find the container with id f8a5353c95aac3ac71cb54285cfc09d8cad6d35807c980e8899f5f89e72e1674 Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.590074 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-d656998f4-p8zts" Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.591527 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-b4c496f69-9fkd7" Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.598992 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-nkkdq" Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.608241 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m95jk\" (UniqueName: \"kubernetes.io/projected/e9edea52-e8de-4e6b-b672-fa71b01d053c-kube-api-access-m95jk\") pod \"test-operator-controller-manager-b4c496f69-9fkd7\" (UID: \"e9edea52-e8de-4e6b-b672-fa71b01d053c\") " pod="openstack-operators/test-operator-controller-manager-b4c496f69-9fkd7" Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.614582 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-b4c496f69-9fkd7"] Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.685155 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-8c6448b9f-jrlfk"] Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.685884 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-77797865fc-9n8nt" Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.686636 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-jrlfk" Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.690844 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-m5vsd" Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.714390 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m95jk\" (UniqueName: \"kubernetes.io/projected/e9edea52-e8de-4e6b-b672-fa71b01d053c-kube-api-access-m95jk\") pod \"test-operator-controller-manager-b4c496f69-9fkd7\" (UID: \"e9edea52-e8de-4e6b-b672-fa71b01d053c\") " pod="openstack-operators/test-operator-controller-manager-b4c496f69-9fkd7" Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.715150 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5m5mp\" (UniqueName: \"kubernetes.io/projected/40748f20-acda-4ced-ac8b-2e741dae0c66-kube-api-access-5m5mp\") pod \"watcher-operator-controller-manager-8c6448b9f-jrlfk\" (UID: \"40748f20-acda-4ced-ac8b-2e741dae0c66\") " pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-jrlfk" Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.716751 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-8c6448b9f-jrlfk"] Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.721362 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-glmx7" Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.752608 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m95jk\" (UniqueName: \"kubernetes.io/projected/e9edea52-e8de-4e6b-b672-fa71b01d053c-kube-api-access-m95jk\") pod \"test-operator-controller-manager-b4c496f69-9fkd7\" (UID: \"e9edea52-e8de-4e6b-b672-fa71b01d053c\") " pod="openstack-operators/test-operator-controller-manager-b4c496f69-9fkd7" Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.757738 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-6c5c9b94-7hk8h"] Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.759772 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-6c5c9b94-7hk8h" Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.763286 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.765597 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-wd2qg" Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.783845 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-2422v"] Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.786898 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-2422v" Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.793437 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-5lhkr" Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.797797 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-6c5c9b94-7hk8h"] Nov 21 14:27:13 crc kubenswrapper[4897]: W1121 14:27:13.801213 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod717d3738_bb33_4a91_ad36_7cf849ae783f.slice/crio-306021a4e3bd1bdc2e6a74dc4acfeb896b34d48a0000d0c677b9a8162d6bce44 WatchSource:0}: Error finding container 306021a4e3bd1bdc2e6a74dc4acfeb896b34d48a0000d0c677b9a8162d6bce44: Status 404 returned error can't find the container with id 306021a4e3bd1bdc2e6a74dc4acfeb896b34d48a0000d0c677b9a8162d6bce44 Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.803681 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-2422v"] Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.818943 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qp2vs\" (UniqueName: \"kubernetes.io/projected/ef9b1721-eadf-4f29-a175-008292188ef2-kube-api-access-qp2vs\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-2422v\" (UID: \"ef9b1721-eadf-4f29-a175-008292188ef2\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-2422v" Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.819007 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6b47cf09-64d3-4ae5-a7d7-d0125af9d456-cert\") pod \"openstack-operator-controller-manager-6c5c9b94-7hk8h\" (UID: \"6b47cf09-64d3-4ae5-a7d7-d0125af9d456\") " pod="openstack-operators/openstack-operator-controller-manager-6c5c9b94-7hk8h" Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.819171 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7952fa01-945e-4491-87aa-af0835a97532-cert\") pod \"openstack-baremetal-operator-controller-manager-8c7444f48-lfbrg\" (UID: \"7952fa01-945e-4491-87aa-af0835a97532\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-lfbrg" Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.819412 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5m5mp\" (UniqueName: \"kubernetes.io/projected/40748f20-acda-4ced-ac8b-2e741dae0c66-kube-api-access-5m5mp\") pod \"watcher-operator-controller-manager-8c6448b9f-jrlfk\" (UID: \"40748f20-acda-4ced-ac8b-2e741dae0c66\") " pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-jrlfk" Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.819469 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5dgzf\" (UniqueName: \"kubernetes.io/projected/6b47cf09-64d3-4ae5-a7d7-d0125af9d456-kube-api-access-5dgzf\") pod \"openstack-operator-controller-manager-6c5c9b94-7hk8h\" (UID: \"6b47cf09-64d3-4ae5-a7d7-d0125af9d456\") " pod="openstack-operators/openstack-operator-controller-manager-6c5c9b94-7hk8h" Nov 21 14:27:13 crc kubenswrapper[4897]: E1121 14:27:13.819430 4897 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 21 14:27:13 crc kubenswrapper[4897]: E1121 14:27:13.819792 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7952fa01-945e-4491-87aa-af0835a97532-cert podName:7952fa01-945e-4491-87aa-af0835a97532 nodeName:}" failed. No retries permitted until 2025-11-21 14:27:14.819772867 +0000 UTC m=+1112.104366342 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/7952fa01-945e-4491-87aa-af0835a97532-cert") pod "openstack-baremetal-operator-controller-manager-8c7444f48-lfbrg" (UID: "7952fa01-945e-4491-87aa-af0835a97532") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.842688 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-75fb479bcc-cr58g"] Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.843647 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5m5mp\" (UniqueName: \"kubernetes.io/projected/40748f20-acda-4ced-ac8b-2e741dae0c66-kube-api-access-5m5mp\") pod \"watcher-operator-controller-manager-8c6448b9f-jrlfk\" (UID: \"40748f20-acda-4ced-ac8b-2e741dae0c66\") " pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-jrlfk" Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.866687 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-767ccfd65f-csgnr"] Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.877035 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-7969689c84-xlbhh"] Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.922451 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5dgzf\" (UniqueName: \"kubernetes.io/projected/6b47cf09-64d3-4ae5-a7d7-d0125af9d456-kube-api-access-5dgzf\") pod \"openstack-operator-controller-manager-6c5c9b94-7hk8h\" (UID: \"6b47cf09-64d3-4ae5-a7d7-d0125af9d456\") " pod="openstack-operators/openstack-operator-controller-manager-6c5c9b94-7hk8h" Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.922653 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qp2vs\" (UniqueName: \"kubernetes.io/projected/ef9b1721-eadf-4f29-a175-008292188ef2-kube-api-access-qp2vs\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-2422v\" (UID: \"ef9b1721-eadf-4f29-a175-008292188ef2\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-2422v" Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.922786 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6b47cf09-64d3-4ae5-a7d7-d0125af9d456-cert\") pod \"openstack-operator-controller-manager-6c5c9b94-7hk8h\" (UID: \"6b47cf09-64d3-4ae5-a7d7-d0125af9d456\") " pod="openstack-operators/openstack-operator-controller-manager-6c5c9b94-7hk8h" Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.940282 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5dgzf\" (UniqueName: \"kubernetes.io/projected/6b47cf09-64d3-4ae5-a7d7-d0125af9d456-kube-api-access-5dgzf\") pod \"openstack-operator-controller-manager-6c5c9b94-7hk8h\" (UID: \"6b47cf09-64d3-4ae5-a7d7-d0125af9d456\") " pod="openstack-operators/openstack-operator-controller-manager-6c5c9b94-7hk8h" Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.940909 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6b47cf09-64d3-4ae5-a7d7-d0125af9d456-cert\") pod \"openstack-operator-controller-manager-6c5c9b94-7hk8h\" (UID: \"6b47cf09-64d3-4ae5-a7d7-d0125af9d456\") " pod="openstack-operators/openstack-operator-controller-manager-6c5c9b94-7hk8h" Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.945240 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qp2vs\" (UniqueName: \"kubernetes.io/projected/ef9b1721-eadf-4f29-a175-008292188ef2-kube-api-access-qp2vs\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-2422v\" (UID: \"ef9b1721-eadf-4f29-a175-008292188ef2\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-2422v" Nov 21 14:27:13 crc kubenswrapper[4897]: I1121 14:27:13.966143 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-b4c496f69-9fkd7" Nov 21 14:27:14 crc kubenswrapper[4897]: I1121 14:27:14.043585 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-cr58g" event={"ID":"e3891c15-c0b9-4965-84fe-05262fe71853","Type":"ContainerStarted","Data":"f8a5353c95aac3ac71cb54285cfc09d8cad6d35807c980e8899f5f89e72e1674"} Nov 21 14:27:14 crc kubenswrapper[4897]: I1121 14:27:14.046948 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-7969689c84-xlbhh" event={"ID":"55f66c84-22f9-4840-83b5-624cce4d5543","Type":"ContainerStarted","Data":"67d012094b39500871a225c59cac5f85891b2aa86607787790f1f5779e2c2672"} Nov 21 14:27:14 crc kubenswrapper[4897]: I1121 14:27:14.048357 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-csgnr" event={"ID":"717d3738-bb33-4a91-ad36-7cf849ae783f","Type":"ContainerStarted","Data":"306021a4e3bd1bdc2e6a74dc4acfeb896b34d48a0000d0c677b9a8162d6bce44"} Nov 21 14:27:14 crc kubenswrapper[4897]: I1121 14:27:14.126235 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-jrlfk" Nov 21 14:27:14 crc kubenswrapper[4897]: I1121 14:27:14.159561 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-6c5c9b94-7hk8h" Nov 21 14:27:14 crc kubenswrapper[4897]: I1121 14:27:14.208355 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-2422v" Nov 21 14:27:14 crc kubenswrapper[4897]: I1121 14:27:14.554439 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-56f54d6746-nrp4j"] Nov 21 14:27:14 crc kubenswrapper[4897]: I1121 14:27:14.565872 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-598f69df5d-l2njs"] Nov 21 14:27:14 crc kubenswrapper[4897]: W1121 14:27:14.566309 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8aee4e2c_4b47_41ec_9953_7e73d0bd6524.slice/crio-34b487106459705607943969c07c370e629c5029392ea7c50d5f79305a506332 WatchSource:0}: Error finding container 34b487106459705607943969c07c370e629c5029392ea7c50d5f79305a506332: Status 404 returned error can't find the container with id 34b487106459705607943969c07c370e629c5029392ea7c50d5f79305a506332 Nov 21 14:27:14 crc kubenswrapper[4897]: W1121 14:27:14.569827 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod65770e5f_95ee_4501_993d_8d5a015da258.slice/crio-00246a6704812da5b9648c385b39f5aefda5234069f1d603b09f32352d100529 WatchSource:0}: Error finding container 00246a6704812da5b9648c385b39f5aefda5234069f1d603b09f32352d100529: Status 404 returned error can't find the container with id 00246a6704812da5b9648c385b39f5aefda5234069f1d603b09f32352d100529 Nov 21 14:27:14 crc kubenswrapper[4897]: I1121 14:27:14.853613 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7952fa01-945e-4491-87aa-af0835a97532-cert\") pod \"openstack-baremetal-operator-controller-manager-8c7444f48-lfbrg\" (UID: \"7952fa01-945e-4491-87aa-af0835a97532\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-lfbrg" Nov 21 14:27:14 crc kubenswrapper[4897]: I1121 14:27:14.869581 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7952fa01-945e-4491-87aa-af0835a97532-cert\") pod \"openstack-baremetal-operator-controller-manager-8c7444f48-lfbrg\" (UID: \"7952fa01-945e-4491-87aa-af0835a97532\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-lfbrg" Nov 21 14:27:14 crc kubenswrapper[4897]: I1121 14:27:14.893215 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-lfbrg" Nov 21 14:27:15 crc kubenswrapper[4897]: I1121 14:27:15.065167 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-56f54d6746-nrp4j" event={"ID":"65770e5f-95ee-4501-993d-8d5a015da258","Type":"ContainerStarted","Data":"00246a6704812da5b9648c385b39f5aefda5234069f1d603b09f32352d100529"} Nov 21 14:27:15 crc kubenswrapper[4897]: I1121 14:27:15.069847 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-l2njs" event={"ID":"8aee4e2c-4b47-41ec-9953-7e73d0bd6524","Type":"ContainerStarted","Data":"34b487106459705607943969c07c370e629c5029392ea7c50d5f79305a506332"} Nov 21 14:27:15 crc kubenswrapper[4897]: I1121 14:27:15.184529 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-54b5986bb8-2dnk2"] Nov 21 14:27:15 crc kubenswrapper[4897]: I1121 14:27:15.193653 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-6dd8864d7c-pt5vd"] Nov 21 14:27:15 crc kubenswrapper[4897]: I1121 14:27:15.199962 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-2sqhj"] Nov 21 14:27:15 crc kubenswrapper[4897]: I1121 14:27:15.210090 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-78bd47f458-2grhg"] Nov 21 14:27:15 crc kubenswrapper[4897]: I1121 14:27:15.220609 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7454b96578-gzzl8"] Nov 21 14:27:15 crc kubenswrapper[4897]: I1121 14:27:15.223138 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6498cbf48f-kq4ss"] Nov 21 14:27:15 crc kubenswrapper[4897]: I1121 14:27:15.232138 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-5b797b8dff-smlb7"] Nov 21 14:27:15 crc kubenswrapper[4897]: I1121 14:27:15.241662 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-99b499f4-phs97"] Nov 21 14:27:15 crc kubenswrapper[4897]: I1121 14:27:15.250055 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-58f887965d-dhc5w"] Nov 21 14:27:15 crc kubenswrapper[4897]: I1121 14:27:15.257463 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-d656998f4-p8zts"] Nov 21 14:27:15 crc kubenswrapper[4897]: W1121 14:27:15.285730 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod282cd160_d81e_4e20_866f_dee909546aa2.slice/crio-b02970d2f1ec3dd81b61533f04f74684ae2a7959e6902856d8670943505b1b9b WatchSource:0}: Error finding container b02970d2f1ec3dd81b61533f04f74684ae2a7959e6902856d8670943505b1b9b: Status 404 returned error can't find the container with id b02970d2f1ec3dd81b61533f04f74684ae2a7959e6902856d8670943505b1b9b Nov 21 14:27:15 crc kubenswrapper[4897]: I1121 14:27:15.672939 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-54fc5f65b7-glmx7"] Nov 21 14:27:15 crc kubenswrapper[4897]: I1121 14:27:15.697144 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-cfbb9c588-bh4rl"] Nov 21 14:27:15 crc kubenswrapper[4897]: I1121 14:27:15.704497 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-b4c496f69-9fkd7"] Nov 21 14:27:15 crc kubenswrapper[4897]: I1121 14:27:15.726920 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-77797865fc-9n8nt"] Nov 21 14:27:15 crc kubenswrapper[4897]: I1121 14:27:15.739683 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-2422v"] Nov 21 14:27:15 crc kubenswrapper[4897]: W1121 14:27:15.743353 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1557d794_d052_44ad_b908_4766d1abe234.slice/crio-8d2a69c03cfa05e482903187952cd03b7f159c5f7908165b3b5c38bc41d97a8d WatchSource:0}: Error finding container 8d2a69c03cfa05e482903187952cd03b7f159c5f7908165b3b5c38bc41d97a8d: Status 404 returned error can't find the container with id 8d2a69c03cfa05e482903187952cd03b7f159c5f7908165b3b5c38bc41d97a8d Nov 21 14:27:15 crc kubenswrapper[4897]: E1121 14:27:15.745938 4897 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:82207e753574d4be246f86c4b074500d66cf20214aa80f0a8525cf3287a35e6d,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-m95jk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-b4c496f69-9fkd7_openstack-operators(e9edea52-e8de-4e6b-b672-fa71b01d053c): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 21 14:27:15 crc kubenswrapper[4897]: E1121 14:27:15.750050 4897 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:38.102.83.70:5001/openstack-k8s-operators/telemetry-operator:6280f54c4d86e239852669b9aa334e584f1fe080,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-lj29c,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-77797865fc-9n8nt_openstack-operators(1557d794-d052-44ad-b908-4766d1abe234): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 21 14:27:15 crc kubenswrapper[4897]: W1121 14:27:15.770363 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6b47cf09_64d3_4ae5_a7d7_d0125af9d456.slice/crio-6c2a8030b519f806f59cb32687a1e53ddb94234235a068eaa1fd678844f01332 WatchSource:0}: Error finding container 6c2a8030b519f806f59cb32687a1e53ddb94234235a068eaa1fd678844f01332: Status 404 returned error can't find the container with id 6c2a8030b519f806f59cb32687a1e53ddb94234235a068eaa1fd678844f01332 Nov 21 14:27:15 crc kubenswrapper[4897]: I1121 14:27:15.770412 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-8c6448b9f-jrlfk"] Nov 21 14:27:15 crc kubenswrapper[4897]: I1121 14:27:15.777910 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-6c5c9b94-7hk8h"] Nov 21 14:27:15 crc kubenswrapper[4897]: E1121 14:27:15.795966 4897 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:4838402d41d42c56613d43dc5041aae475a2b18e6172491d6c4d4a78a580697f,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5m5mp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-8c6448b9f-jrlfk_openstack-operators(40748f20-acda-4ced-ac8b-2e741dae0c66): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 21 14:27:15 crc kubenswrapper[4897]: W1121 14:27:15.803127 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef9b1721_eadf_4f29_a175_008292188ef2.slice/crio-06fd56fb80a2d8314994b5f6e529c94f2d92ac65706dd1bd4eb1291adaa67497 WatchSource:0}: Error finding container 06fd56fb80a2d8314994b5f6e529c94f2d92ac65706dd1bd4eb1291adaa67497: Status 404 returned error can't find the container with id 06fd56fb80a2d8314994b5f6e529c94f2d92ac65706dd1bd4eb1291adaa67497 Nov 21 14:27:15 crc kubenswrapper[4897]: E1121 14:27:15.829048 4897 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-qp2vs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-5f97d8c699-2422v_openstack-operators(ef9b1721-eadf-4f29-a175-008292188ef2): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 21 14:27:15 crc kubenswrapper[4897]: E1121 14:27:15.830831 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-2422v" podUID="ef9b1721-eadf-4f29-a175-008292188ef2" Nov 21 14:27:15 crc kubenswrapper[4897]: I1121 14:27:15.853299 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-lfbrg"] Nov 21 14:27:16 crc kubenswrapper[4897]: I1121 14:27:16.081371 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-b4c496f69-9fkd7" event={"ID":"e9edea52-e8de-4e6b-b672-fa71b01d053c","Type":"ContainerStarted","Data":"8748c7297e4b3dfc89b61c8056aee60314175fceefd54b0fde1ddefeeea3db14"} Nov 21 14:27:16 crc kubenswrapper[4897]: I1121 14:27:16.083148 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-2sqhj" event={"ID":"81f3265d-f571-47a2-9ad8-b7cdf952d850","Type":"ContainerStarted","Data":"4ef8b86814a25054c0e055491bcdf80bd94a082caecf1336a79d0e0045b12e82"} Nov 21 14:27:16 crc kubenswrapper[4897]: I1121 14:27:16.084926 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-2dnk2" event={"ID":"3b3d91dd-852c-4b65-a2c9-dbc02e79daaa","Type":"ContainerStarted","Data":"58105753a99ef93a414d951a3dd70152f294b140ce25fa2bdd1b16692bdcedd9"} Nov 21 14:27:16 crc kubenswrapper[4897]: I1121 14:27:16.086312 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-d656998f4-p8zts" event={"ID":"a031ee6b-00fb-4795-9e6d-6e21068c405e","Type":"ContainerStarted","Data":"30cddfe4fdd4e84cb28bbcf7b116d2b63dadb62a79175f06f8f8a467f6d31679"} Nov 21 14:27:16 crc kubenswrapper[4897]: I1121 14:27:16.087731 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-smlb7" event={"ID":"4b3dcab3-52d5-4f1d-bde6-2ae9803a0a8f","Type":"ContainerStarted","Data":"c7fc3d98d96bf109cb8371dce2f4f8cfc77ece99b4c2a4ba46b63a86cbf3dc3a"} Nov 21 14:27:16 crc kubenswrapper[4897]: I1121 14:27:16.104362 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-gzzl8" event={"ID":"d4c39700-beb8-4272-9f54-1c501332af0d","Type":"ContainerStarted","Data":"def9829a84b2f63f169fabb765cbb49ef79bacc3048c5c9ecfa9d889ac184f0e"} Nov 21 14:27:16 crc kubenswrapper[4897]: I1121 14:27:16.104414 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-58f887965d-dhc5w" event={"ID":"282cd160-d81e-4e20-866f-dee909546aa2","Type":"ContainerStarted","Data":"b02970d2f1ec3dd81b61533f04f74684ae2a7959e6902856d8670943505b1b9b"} Nov 21 14:27:16 crc kubenswrapper[4897]: I1121 14:27:16.104430 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-jrlfk" event={"ID":"40748f20-acda-4ced-ac8b-2e741dae0c66","Type":"ContainerStarted","Data":"792b4effe96d2e8e30f57b78605d582848ca7ad1add33f306e4504d5ee2e0718"} Nov 21 14:27:16 crc kubenswrapper[4897]: I1121 14:27:16.104442 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-bh4rl" event={"ID":"499cc09f-fafe-47c2-87ac-bc00d79e949e","Type":"ContainerStarted","Data":"bb44c050a4f081677270e7a2dc28907907eef804994d087c2576e8732a945a03"} Nov 21 14:27:16 crc kubenswrapper[4897]: I1121 14:27:16.104455 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-6c5c9b94-7hk8h" event={"ID":"6b47cf09-64d3-4ae5-a7d7-d0125af9d456","Type":"ContainerStarted","Data":"6c2a8030b519f806f59cb32687a1e53ddb94234235a068eaa1fd678844f01332"} Nov 21 14:27:16 crc kubenswrapper[4897]: I1121 14:27:16.104469 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-77797865fc-9n8nt" event={"ID":"1557d794-d052-44ad-b908-4766d1abe234","Type":"ContainerStarted","Data":"8d2a69c03cfa05e482903187952cd03b7f159c5f7908165b3b5c38bc41d97a8d"} Nov 21 14:27:16 crc kubenswrapper[4897]: I1121 14:27:16.106103 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-pt5vd" event={"ID":"a7d8ae48-9dd4-40d0-bab5-26e3cff24b20","Type":"ContainerStarted","Data":"13665b88c0a53d09954efcc5e9ed06fa6943174a45c19d0b3d3ed5caf209c124"} Nov 21 14:27:16 crc kubenswrapper[4897]: I1121 14:27:16.107104 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-lfbrg" event={"ID":"7952fa01-945e-4491-87aa-af0835a97532","Type":"ContainerStarted","Data":"16714fb8ef5e9b745ab86d9b3aa54b995dc0b10554a40ffa3e4b7213d6f554ae"} Nov 21 14:27:16 crc kubenswrapper[4897]: I1121 14:27:16.108882 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-2422v" event={"ID":"ef9b1721-eadf-4f29-a175-008292188ef2","Type":"ContainerStarted","Data":"06fd56fb80a2d8314994b5f6e529c94f2d92ac65706dd1bd4eb1291adaa67497"} Nov 21 14:27:16 crc kubenswrapper[4897]: E1121 14:27:16.112756 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-2422v" podUID="ef9b1721-eadf-4f29-a175-008292188ef2" Nov 21 14:27:16 crc kubenswrapper[4897]: I1121 14:27:16.113801 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-kq4ss" event={"ID":"39c184ba-7360-4d31-99f5-239711efdf91","Type":"ContainerStarted","Data":"931a874a268b9496a1d9ac8a74a2ddfabe51af1acc641a9ef5238ddc81074897"} Nov 21 14:27:16 crc kubenswrapper[4897]: I1121 14:27:16.120099 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-phs97" event={"ID":"d1fc2027-5a14-4de3-8c0e-c4ecee66c40b","Type":"ContainerStarted","Data":"f09cb85d38e5430b90ef063958f9f9ef932f1830e7bb10f2dadfb9e8957a7024"} Nov 21 14:27:16 crc kubenswrapper[4897]: I1121 14:27:16.122483 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-2grhg" event={"ID":"77f4c339-9ff1-4943-922d-88d81a73a1f1","Type":"ContainerStarted","Data":"96596b468d09398eaf2677e227235ba871eb4750691717fd6658f92bd615fca8"} Nov 21 14:27:16 crc kubenswrapper[4897]: I1121 14:27:16.125198 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-glmx7" event={"ID":"09df9b2c-a507-4a35-b646-a2dc45e1eb71","Type":"ContainerStarted","Data":"a7c82cca80c53623bc33d96ff1d1c286a7c2620640009e9e265c252a486f6902"} Nov 21 14:27:16 crc kubenswrapper[4897]: E1121 14:27:16.241676 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/test-operator-controller-manager-b4c496f69-9fkd7" podUID="e9edea52-e8de-4e6b-b672-fa71b01d053c" Nov 21 14:27:16 crc kubenswrapper[4897]: E1121 14:27:16.248435 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-jrlfk" podUID="40748f20-acda-4ced-ac8b-2e741dae0c66" Nov 21 14:27:16 crc kubenswrapper[4897]: E1121 14:27:16.249211 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/telemetry-operator-controller-manager-77797865fc-9n8nt" podUID="1557d794-d052-44ad-b908-4766d1abe234" Nov 21 14:27:17 crc kubenswrapper[4897]: I1121 14:27:17.154784 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-6c5c9b94-7hk8h" event={"ID":"6b47cf09-64d3-4ae5-a7d7-d0125af9d456","Type":"ContainerStarted","Data":"4646992da50871dff6d0b205ac2b0589d18f392718ff65a876d02ccc36a8bbac"} Nov 21 14:27:17 crc kubenswrapper[4897]: I1121 14:27:17.160041 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-77797865fc-9n8nt" event={"ID":"1557d794-d052-44ad-b908-4766d1abe234","Type":"ContainerStarted","Data":"2303636529f7f3fd46f04a04fa5a8a899ecc4a9bc8ece06d6a32480d7a2e0e7d"} Nov 21 14:27:17 crc kubenswrapper[4897]: E1121 14:27:17.166571 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.70:5001/openstack-k8s-operators/telemetry-operator:6280f54c4d86e239852669b9aa334e584f1fe080\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-77797865fc-9n8nt" podUID="1557d794-d052-44ad-b908-4766d1abe234" Nov 21 14:27:17 crc kubenswrapper[4897]: I1121 14:27:17.167954 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-jrlfk" event={"ID":"40748f20-acda-4ced-ac8b-2e741dae0c66","Type":"ContainerStarted","Data":"7f6345ba3687286e5353ab020dffdd7a04401b6609df36f9ebf469e1a9d5d27b"} Nov 21 14:27:17 crc kubenswrapper[4897]: E1121 14:27:17.170351 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:4838402d41d42c56613d43dc5041aae475a2b18e6172491d6c4d4a78a580697f\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-jrlfk" podUID="40748f20-acda-4ced-ac8b-2e741dae0c66" Nov 21 14:27:17 crc kubenswrapper[4897]: I1121 14:27:17.174312 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-b4c496f69-9fkd7" event={"ID":"e9edea52-e8de-4e6b-b672-fa71b01d053c","Type":"ContainerStarted","Data":"416703e540fd23f161515098492a6e0344fe5d4b3b55bc7226325af19c725ec4"} Nov 21 14:27:17 crc kubenswrapper[4897]: E1121 14:27:17.175320 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-2422v" podUID="ef9b1721-eadf-4f29-a175-008292188ef2" Nov 21 14:27:17 crc kubenswrapper[4897]: E1121 14:27:17.181871 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:82207e753574d4be246f86c4b074500d66cf20214aa80f0a8525cf3287a35e6d\\\"\"" pod="openstack-operators/test-operator-controller-manager-b4c496f69-9fkd7" podUID="e9edea52-e8de-4e6b-b672-fa71b01d053c" Nov 21 14:27:18 crc kubenswrapper[4897]: E1121 14:27:18.184551 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:4838402d41d42c56613d43dc5041aae475a2b18e6172491d6c4d4a78a580697f\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-jrlfk" podUID="40748f20-acda-4ced-ac8b-2e741dae0c66" Nov 21 14:27:18 crc kubenswrapper[4897]: E1121 14:27:18.184922 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.70:5001/openstack-k8s-operators/telemetry-operator:6280f54c4d86e239852669b9aa334e584f1fe080\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-77797865fc-9n8nt" podUID="1557d794-d052-44ad-b908-4766d1abe234" Nov 21 14:27:18 crc kubenswrapper[4897]: E1121 14:27:18.185773 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:82207e753574d4be246f86c4b074500d66cf20214aa80f0a8525cf3287a35e6d\\\"\"" pod="openstack-operators/test-operator-controller-manager-b4c496f69-9fkd7" podUID="e9edea52-e8de-4e6b-b672-fa71b01d053c" Nov 21 14:27:34 crc kubenswrapper[4897]: I1121 14:27:34.371380 4897 patch_prober.go:28] interesting pod/machine-config-daemon-krv5b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 21 14:27:34 crc kubenswrapper[4897]: I1121 14:27:34.372164 4897 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 21 14:27:40 crc kubenswrapper[4897]: E1121 14:27:40.135361 4897 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/mariadb-operator@sha256:7b90521b9e9cb4eb43c2f1c3bf85dbd068d684315f4f705b07708dd078df9d04" Nov 21 14:27:40 crc kubenswrapper[4897]: E1121 14:27:40.135892 4897 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/mariadb-operator@sha256:7b90521b9e9cb4eb43c2f1c3bf85dbd068d684315f4f705b07708dd078df9d04,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-xccf8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-54b5986bb8-2dnk2_openstack-operators(3b3d91dd-852c-4b65-a2c9-dbc02e79daaa): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 21 14:27:42 crc kubenswrapper[4897]: E1121 14:27:42.646711 4897 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ovn-operator@sha256:5d49d4594c66eda7b151746cc6e1d3c67c0129b4503eeb043a64ae8ec2da6a1b" Nov 21 14:27:42 crc kubenswrapper[4897]: E1121 14:27:42.647911 4897 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:5d49d4594c66eda7b151746cc6e1d3c67c0129b4503eeb043a64ae8ec2da6a1b,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-jpwp5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-54fc5f65b7-glmx7_openstack-operators(09df9b2c-a507-4a35-b646-a2dc45e1eb71): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 21 14:27:43 crc kubenswrapper[4897]: E1121 14:27:43.712045 4897 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/infra-operator@sha256:86df58f744c1d23233cc98f6ea17c8d6da637c50003d0fc8c100045594aa9894" Nov 21 14:27:43 crc kubenswrapper[4897]: E1121 14:27:43.712250 4897 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/infra-operator@sha256:86df58f744c1d23233cc98f6ea17c8d6da637c50003d0fc8c100045594aa9894,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{600 -3} {} 600m DecimalSI},memory: {{2147483648 0} {} 2Gi BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{536870912 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-j4xfp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod infra-operator-controller-manager-6dd8864d7c-pt5vd_openstack-operators(a7d8ae48-9dd4-40d0-bab5-26e3cff24b20): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 21 14:27:44 crc kubenswrapper[4897]: E1121 14:27:44.567294 4897 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/keystone-operator@sha256:3ef72bbd7cce89ff54d850ff44ca6d7b2360834a502da3d561aeb6fd3d9af50a" Nov 21 14:27:44 crc kubenswrapper[4897]: E1121 14:27:44.567839 4897 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:3ef72bbd7cce89ff54d850ff44ca6d7b2360834a502da3d561aeb6fd3d9af50a,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-pb4rl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-7454b96578-gzzl8_openstack-operators(d4c39700-beb8-4272-9f54-1c501332af0d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 21 14:27:50 crc kubenswrapper[4897]: E1121 14:27:50.077928 4897 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/neutron-operator@sha256:207578cb433471cc1a79c21a808c8a15489d1d3c9fa77e29f3f697c33917fec6" Nov 21 14:27:50 crc kubenswrapper[4897]: E1121 14:27:50.078675 4897 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:207578cb433471cc1a79c21a808c8a15489d1d3c9fa77e29f3f697c33917fec6,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-xv2s8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-78bd47f458-2grhg_openstack-operators(77f4c339-9ff1-4943-922d-88d81a73a1f1): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 21 14:27:52 crc kubenswrapper[4897]: E1121 14:27:52.203640 4897 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ironic-operator@sha256:b582189b55fddc180a6d468c9dba7078009a693db37b4093d4ba0c99ec675377" Nov 21 14:27:52 crc kubenswrapper[4897]: E1121 14:27:52.205123 4897 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ironic-operator@sha256:b582189b55fddc180a6d468c9dba7078009a693db37b4093d4ba0c99ec675377,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-88c4f,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ironic-operator-controller-manager-99b499f4-phs97_openstack-operators(d1fc2027-5a14-4de3-8c0e-c4ecee66c40b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 21 14:27:52 crc kubenswrapper[4897]: E1121 14:27:52.653560 4897 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/placement-operator@sha256:4094e7fc11a33e8e2b6768a053cafaf5b122446d23f9113d43d520cb64e9776c" Nov 21 14:27:52 crc kubenswrapper[4897]: E1121 14:27:52.653725 4897 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:4094e7fc11a33e8e2b6768a053cafaf5b122446d23f9113d43d520cb64e9776c,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-gvkmg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-5b797b8dff-smlb7_openstack-operators(4b3dcab3-52d5-4f1d-bde6-2ae9803a0a8f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 21 14:27:53 crc kubenswrapper[4897]: E1121 14:27:53.482740 4897 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/nova-operator@sha256:c053e34316044f14929e16e4f0d97f9f1b24cb68b5e22b925ca74c66aaaed0a7" Nov 21 14:27:53 crc kubenswrapper[4897]: E1121 14:27:53.482928 4897 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:c053e34316044f14929e16e4f0d97f9f1b24cb68b5e22b925ca74c66aaaed0a7,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-jnh5z,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-cfbb9c588-bh4rl_openstack-operators(499cc09f-fafe-47c2-87ac-bc00d79e949e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 21 14:28:04 crc kubenswrapper[4897]: I1121 14:28:04.371373 4897 patch_prober.go:28] interesting pod/machine-config-daemon-krv5b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 21 14:28:04 crc kubenswrapper[4897]: I1121 14:28:04.372129 4897 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 21 14:28:06 crc kubenswrapper[4897]: E1121 14:28:06.222608 4897 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/watcher-operator@sha256:4838402d41d42c56613d43dc5041aae475a2b18e6172491d6c4d4a78a580697f" Nov 21 14:28:06 crc kubenswrapper[4897]: E1121 14:28:06.222847 4897 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:4838402d41d42c56613d43dc5041aae475a2b18e6172491d6c4d4a78a580697f,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5m5mp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-8c6448b9f-jrlfk_openstack-operators(40748f20-acda-4ced-ac8b-2e741dae0c66): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 21 14:28:06 crc kubenswrapper[4897]: E1121 14:28:06.234617 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-jrlfk" podUID="40748f20-acda-4ced-ac8b-2e741dae0c66" Nov 21 14:28:07 crc kubenswrapper[4897]: E1121 14:28:07.191046 4897 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2" Nov 21 14:28:07 crc kubenswrapper[4897]: E1121 14:28:07.191645 4897 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-qp2vs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-5f97d8c699-2422v_openstack-operators(ef9b1721-eadf-4f29-a175-008292188ef2): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 21 14:28:07 crc kubenswrapper[4897]: E1121 14:28:07.192993 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-2422v" podUID="ef9b1721-eadf-4f29-a175-008292188ef2" Nov 21 14:28:07 crc kubenswrapper[4897]: I1121 14:28:07.604836 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-6c5c9b94-7hk8h" event={"ID":"6b47cf09-64d3-4ae5-a7d7-d0125af9d456","Type":"ContainerStarted","Data":"b8b44b612e9f6d9d18e2635b51091197bb9340b91e5635ae8ac38ec1dbd6dc38"} Nov 21 14:28:07 crc kubenswrapper[4897]: I1121 14:28:07.605305 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-6c5c9b94-7hk8h" Nov 21 14:28:07 crc kubenswrapper[4897]: I1121 14:28:07.613454 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-6c5c9b94-7hk8h" Nov 21 14:28:07 crc kubenswrapper[4897]: I1121 14:28:07.641919 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-6c5c9b94-7hk8h" podStartSLOduration=54.641901424 podStartE2EDuration="54.641901424s" podCreationTimestamp="2025-11-21 14:27:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:28:07.638989876 +0000 UTC m=+1164.923583371" watchObservedRunningTime="2025-11-21 14:28:07.641901424 +0000 UTC m=+1164.926494909" Nov 21 14:28:07 crc kubenswrapper[4897]: E1121 14:28:07.654404 4897 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.70:5001/openstack-k8s-operators/telemetry-operator:6280f54c4d86e239852669b9aa334e584f1fe080" Nov 21 14:28:07 crc kubenswrapper[4897]: E1121 14:28:07.654474 4897 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.70:5001/openstack-k8s-operators/telemetry-operator:6280f54c4d86e239852669b9aa334e584f1fe080" Nov 21 14:28:07 crc kubenswrapper[4897]: E1121 14:28:07.654640 4897 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:38.102.83.70:5001/openstack-k8s-operators/telemetry-operator:6280f54c4d86e239852669b9aa334e584f1fe080,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-lj29c,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-77797865fc-9n8nt_openstack-operators(1557d794-d052-44ad-b908-4766d1abe234): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 21 14:28:07 crc kubenswrapper[4897]: E1121 14:28:07.655945 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/telemetry-operator-controller-manager-77797865fc-9n8nt" podUID="1557d794-d052-44ad-b908-4766d1abe234" Nov 21 14:28:08 crc kubenswrapper[4897]: E1121 14:28:08.228134 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-glmx7" podUID="09df9b2c-a507-4a35-b646-a2dc45e1eb71" Nov 21 14:28:08 crc kubenswrapper[4897]: E1121 14:28:08.270247 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-2dnk2" podUID="3b3d91dd-852c-4b65-a2c9-dbc02e79daaa" Nov 21 14:28:08 crc kubenswrapper[4897]: E1121 14:28:08.307215 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-pt5vd" podUID="a7d8ae48-9dd4-40d0-bab5-26e3cff24b20" Nov 21 14:28:08 crc kubenswrapper[4897]: E1121 14:28:08.323024 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-smlb7" podUID="4b3dcab3-52d5-4f1d-bde6-2ae9803a0a8f" Nov 21 14:28:08 crc kubenswrapper[4897]: I1121 14:28:08.627060 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-2sqhj" event={"ID":"81f3265d-f571-47a2-9ad8-b7cdf952d850","Type":"ContainerStarted","Data":"d88d110ca80eda4b4dd0f6c3e483800dbddcda39a908dff37a5b481372679a4f"} Nov 21 14:28:08 crc kubenswrapper[4897]: I1121 14:28:08.643055 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-56f54d6746-nrp4j" event={"ID":"65770e5f-95ee-4501-993d-8d5a015da258","Type":"ContainerStarted","Data":"a8163cca7daf290a4f1050dd23d0da37b2a1ee34b7d0aa37bcb89323a5e67589"} Nov 21 14:28:08 crc kubenswrapper[4897]: I1121 14:28:08.682892 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-pt5vd" event={"ID":"a7d8ae48-9dd4-40d0-bab5-26e3cff24b20","Type":"ContainerStarted","Data":"78efac0630331fa68a4d4cf40e5f2e14d68b9f85102e18be8ebd0aadd674d80c"} Nov 21 14:28:08 crc kubenswrapper[4897]: I1121 14:28:08.684937 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-csgnr" event={"ID":"717d3738-bb33-4a91-ad36-7cf849ae783f","Type":"ContainerStarted","Data":"32f4d47a6b0c4d07992b877e9ace721e9413aa1dea320dafaeb16c57a9cbcbc3"} Nov 21 14:28:08 crc kubenswrapper[4897]: I1121 14:28:08.689665 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-d656998f4-p8zts" event={"ID":"a031ee6b-00fb-4795-9e6d-6e21068c405e","Type":"ContainerStarted","Data":"e23f1c5ee174cc28c7117bfabb6887f190b643cbfa917d8bde77cf3852a4ba25"} Nov 21 14:28:08 crc kubenswrapper[4897]: I1121 14:28:08.695347 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-phs97" event={"ID":"d1fc2027-5a14-4de3-8c0e-c4ecee66c40b","Type":"ContainerStarted","Data":"8d3f3208609dc00a75c0dcbdb7c466629b4746a652ab48064e53ac3562f2b9b4"} Nov 21 14:28:08 crc kubenswrapper[4897]: I1121 14:28:08.700194 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-7969689c84-xlbhh" event={"ID":"55f66c84-22f9-4840-83b5-624cce4d5543","Type":"ContainerStarted","Data":"ad22df1cd2effc5c31b30087eb18d7f4cb69c090320f805642492268574e48e5"} Nov 21 14:28:08 crc kubenswrapper[4897]: I1121 14:28:08.701222 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-smlb7" event={"ID":"4b3dcab3-52d5-4f1d-bde6-2ae9803a0a8f","Type":"ContainerStarted","Data":"423edc2849f9b629419bc375c7f58478ac40c8b1c739736a9fc5ca1beecfad1b"} Nov 21 14:28:08 crc kubenswrapper[4897]: I1121 14:28:08.711806 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-gzzl8" event={"ID":"d4c39700-beb8-4272-9f54-1c501332af0d","Type":"ContainerStarted","Data":"d07bbcd50e7a524d66e59265c050b686ac9afc4c525719fcd7cf1c81a4775046"} Nov 21 14:28:08 crc kubenswrapper[4897]: I1121 14:28:08.728653 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-b4c496f69-9fkd7" event={"ID":"e9edea52-e8de-4e6b-b672-fa71b01d053c","Type":"ContainerStarted","Data":"080f98299de33c24b1bfc0575f1ca12bf98805f49862e1378c0c3ce542acd0a1"} Nov 21 14:28:08 crc kubenswrapper[4897]: I1121 14:28:08.739684 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-bh4rl" event={"ID":"499cc09f-fafe-47c2-87ac-bc00d79e949e","Type":"ContainerStarted","Data":"4e90a7cbd95357ed4a6c79e10e1029a37fca4eeb744cf208e4a162a2b2330bb9"} Nov 21 14:28:08 crc kubenswrapper[4897]: I1121 14:28:08.754036 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-cr58g" event={"ID":"e3891c15-c0b9-4965-84fe-05262fe71853","Type":"ContainerStarted","Data":"bfad6e7e30558b4b9168691b2c02f24ca0dd7113f62b038631af89383ac9082a"} Nov 21 14:28:08 crc kubenswrapper[4897]: I1121 14:28:08.774772 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-2grhg" event={"ID":"77f4c339-9ff1-4943-922d-88d81a73a1f1","Type":"ContainerStarted","Data":"ec490d0d498a6ea17827e3f36c649dca0ec2ed6673885cfe90a09b12187aef10"} Nov 21 14:28:08 crc kubenswrapper[4897]: I1121 14:28:08.794280 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-glmx7" event={"ID":"09df9b2c-a507-4a35-b646-a2dc45e1eb71","Type":"ContainerStarted","Data":"6ad493216a88c81dcf2ec555a41a1909cab899ceb21d078dbfc74004f68123ab"} Nov 21 14:28:08 crc kubenswrapper[4897]: I1121 14:28:08.821771 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-2dnk2" event={"ID":"3b3d91dd-852c-4b65-a2c9-dbc02e79daaa","Type":"ContainerStarted","Data":"6b1b1b7b5f8a0a27b3e3a77650340be6e1d4f0e3f574188bbb3c50968279527c"} Nov 21 14:28:08 crc kubenswrapper[4897]: I1121 14:28:08.861001 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-kq4ss" event={"ID":"39c184ba-7360-4d31-99f5-239711efdf91","Type":"ContainerStarted","Data":"8c7693d0561e03130840d460188294250a805b1e24a29b0149f09f6a8a8874e9"} Nov 21 14:28:09 crc kubenswrapper[4897]: I1121 14:28:09.873448 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-l2njs" event={"ID":"8aee4e2c-4b47-41ec-9953-7e73d0bd6524","Type":"ContainerStarted","Data":"07db59ac2f33d11a88aa0dce10c66605f36c0a53bd22971dbb7797c8c4c3ce78"} Nov 21 14:28:09 crc kubenswrapper[4897]: I1121 14:28:09.874847 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-lfbrg" event={"ID":"7952fa01-945e-4491-87aa-af0835a97532","Type":"ContainerStarted","Data":"9b47aedf6dc0094ffc7a80975844db4bcafc96f6c24521f04c77da0ce1d49058"} Nov 21 14:28:10 crc kubenswrapper[4897]: E1121 14:28:10.400659 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-phs97" podUID="d1fc2027-5a14-4de3-8c0e-c4ecee66c40b" Nov 21 14:28:10 crc kubenswrapper[4897]: E1121 14:28:10.400757 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-gzzl8" podUID="d4c39700-beb8-4272-9f54-1c501332af0d" Nov 21 14:28:10 crc kubenswrapper[4897]: E1121 14:28:10.400830 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-2grhg" podUID="77f4c339-9ff1-4943-922d-88d81a73a1f1" Nov 21 14:28:10 crc kubenswrapper[4897]: E1121 14:28:10.400924 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-bh4rl" podUID="499cc09f-fafe-47c2-87ac-bc00d79e949e" Nov 21 14:28:10 crc kubenswrapper[4897]: I1121 14:28:10.885232 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-58f887965d-dhc5w" event={"ID":"282cd160-d81e-4e20-866f-dee909546aa2","Type":"ContainerStarted","Data":"985ee33e24c459413e090ee8cdf231145498ed8d0a8abe59c720a1de3864bd20"} Nov 21 14:28:10 crc kubenswrapper[4897]: I1121 14:28:10.887177 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-cr58g" event={"ID":"e3891c15-c0b9-4965-84fe-05262fe71853","Type":"ContainerStarted","Data":"d62203462dfac088910bbbc2555f8360bf637e9f797f80949dfdced314fa03ca"} Nov 21 14:28:10 crc kubenswrapper[4897]: I1121 14:28:10.888302 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-cr58g" Nov 21 14:28:10 crc kubenswrapper[4897]: I1121 14:28:10.890304 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-2sqhj" event={"ID":"81f3265d-f571-47a2-9ad8-b7cdf952d850","Type":"ContainerStarted","Data":"25773b8bb6f3fa1ac76e0aab3bcece9780bf1bd3001e245736a338763c234151"} Nov 21 14:28:10 crc kubenswrapper[4897]: I1121 14:28:10.890606 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-2sqhj" Nov 21 14:28:10 crc kubenswrapper[4897]: I1121 14:28:10.895932 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-7969689c84-xlbhh" event={"ID":"55f66c84-22f9-4840-83b5-624cce4d5543","Type":"ContainerStarted","Data":"5a4e23afc0bc22b69ed4d3e7386b8fb33437904a670fd3355c55b16d73da06ce"} Nov 21 14:28:10 crc kubenswrapper[4897]: I1121 14:28:10.896164 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-7969689c84-xlbhh" Nov 21 14:28:10 crc kubenswrapper[4897]: I1121 14:28:10.898467 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-56f54d6746-nrp4j" event={"ID":"65770e5f-95ee-4501-993d-8d5a015da258","Type":"ContainerStarted","Data":"abb1d908e07e5b8e9ba9ec8186cf2d845d3a1b0c7e08ba7b079c425907afefa4"} Nov 21 14:28:10 crc kubenswrapper[4897]: I1121 14:28:10.926423 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-cr58g" podStartSLOduration=44.878705666 podStartE2EDuration="58.926400015s" podCreationTimestamp="2025-11-21 14:27:12 +0000 UTC" firstStartedPulling="2025-11-21 14:27:13.644718909 +0000 UTC m=+1110.929312384" lastFinishedPulling="2025-11-21 14:27:27.692413258 +0000 UTC m=+1124.977006733" observedRunningTime="2025-11-21 14:28:10.908242149 +0000 UTC m=+1168.192835634" watchObservedRunningTime="2025-11-21 14:28:10.926400015 +0000 UTC m=+1168.210993500" Nov 21 14:28:10 crc kubenswrapper[4897]: I1121 14:28:10.936649 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-b4c496f69-9fkd7" podStartSLOduration=7.21456585 podStartE2EDuration="58.936629089s" podCreationTimestamp="2025-11-21 14:27:12 +0000 UTC" firstStartedPulling="2025-11-21 14:27:15.745713891 +0000 UTC m=+1113.030307366" lastFinishedPulling="2025-11-21 14:28:07.4677771 +0000 UTC m=+1164.752370605" observedRunningTime="2025-11-21 14:28:10.927249258 +0000 UTC m=+1168.211842733" watchObservedRunningTime="2025-11-21 14:28:10.936629089 +0000 UTC m=+1168.221222564" Nov 21 14:28:10 crc kubenswrapper[4897]: I1121 14:28:10.983981 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-2sqhj" podStartSLOduration=12.434419678 podStartE2EDuration="58.983962207s" podCreationTimestamp="2025-11-21 14:27:12 +0000 UTC" firstStartedPulling="2025-11-21 14:27:15.256385645 +0000 UTC m=+1112.540979120" lastFinishedPulling="2025-11-21 14:28:01.805928164 +0000 UTC m=+1159.090521649" observedRunningTime="2025-11-21 14:28:10.982482937 +0000 UTC m=+1168.267076412" watchObservedRunningTime="2025-11-21 14:28:10.983962207 +0000 UTC m=+1168.268555672" Nov 21 14:28:11 crc kubenswrapper[4897]: I1121 14:28:11.014794 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-56f54d6746-nrp4j" podStartSLOduration=37.634250103 podStartE2EDuration="59.014774122s" podCreationTimestamp="2025-11-21 14:27:12 +0000 UTC" firstStartedPulling="2025-11-21 14:27:14.57480412 +0000 UTC m=+1111.859397585" lastFinishedPulling="2025-11-21 14:27:35.955328109 +0000 UTC m=+1133.239921604" observedRunningTime="2025-11-21 14:28:11.009803549 +0000 UTC m=+1168.294397034" watchObservedRunningTime="2025-11-21 14:28:11.014774122 +0000 UTC m=+1168.299367597" Nov 21 14:28:11 crc kubenswrapper[4897]: I1121 14:28:11.026127 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-7969689c84-xlbhh" podStartSLOduration=45.655754458 podStartE2EDuration="59.026104835s" podCreationTimestamp="2025-11-21 14:27:12 +0000 UTC" firstStartedPulling="2025-11-21 14:27:13.910474477 +0000 UTC m=+1111.195067952" lastFinishedPulling="2025-11-21 14:27:27.280824854 +0000 UTC m=+1124.565418329" observedRunningTime="2025-11-21 14:28:11.026032413 +0000 UTC m=+1168.310625898" watchObservedRunningTime="2025-11-21 14:28:11.026104835 +0000 UTC m=+1168.310698310" Nov 21 14:28:11 crc kubenswrapper[4897]: I1121 14:28:11.927073 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-lfbrg" event={"ID":"7952fa01-945e-4491-87aa-af0835a97532","Type":"ContainerStarted","Data":"f88bb0f267e6734df4535ce095888252ef7d67e369db34e4322e447f97fcacdb"} Nov 21 14:28:11 crc kubenswrapper[4897]: I1121 14:28:11.927518 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-lfbrg" Nov 21 14:28:11 crc kubenswrapper[4897]: I1121 14:28:11.942068 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-csgnr" event={"ID":"717d3738-bb33-4a91-ad36-7cf849ae783f","Type":"ContainerStarted","Data":"9425d613e0505879aaae9468962d27d8548351444440307f050dd4483e841ad4"} Nov 21 14:28:11 crc kubenswrapper[4897]: I1121 14:28:11.942204 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-csgnr" Nov 21 14:28:11 crc kubenswrapper[4897]: I1121 14:28:11.945369 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-58f887965d-dhc5w" event={"ID":"282cd160-d81e-4e20-866f-dee909546aa2","Type":"ContainerStarted","Data":"049f3402a21151702492ccf82275e030ea8499fe0e90af0b61c222c60382e992"} Nov 21 14:28:11 crc kubenswrapper[4897]: I1121 14:28:11.946427 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-58f887965d-dhc5w" Nov 21 14:28:11 crc kubenswrapper[4897]: I1121 14:28:11.953335 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-kq4ss" event={"ID":"39c184ba-7360-4d31-99f5-239711efdf91","Type":"ContainerStarted","Data":"b7f7d2960b75b5b55556644f127388820e96cd7b4d0de104cdd96c7cb2848852"} Nov 21 14:28:11 crc kubenswrapper[4897]: I1121 14:28:11.954208 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-kq4ss" Nov 21 14:28:11 crc kubenswrapper[4897]: I1121 14:28:11.955878 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-smlb7" event={"ID":"4b3dcab3-52d5-4f1d-bde6-2ae9803a0a8f","Type":"ContainerStarted","Data":"fa804cbed21bde4141ca644f8c1f422736079ec65f388f226ad02eb390f592c2"} Nov 21 14:28:11 crc kubenswrapper[4897]: I1121 14:28:11.956411 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-smlb7" Nov 21 14:28:11 crc kubenswrapper[4897]: I1121 14:28:11.957722 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-pt5vd" event={"ID":"a7d8ae48-9dd4-40d0-bab5-26e3cff24b20","Type":"ContainerStarted","Data":"c02cddd7e502ed29cd7c65cc16d7ee1f621eb3da29023bab91739df2168278fb"} Nov 21 14:28:11 crc kubenswrapper[4897]: I1121 14:28:11.958211 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-pt5vd" Nov 21 14:28:11 crc kubenswrapper[4897]: I1121 14:28:11.959895 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-d656998f4-p8zts" event={"ID":"a031ee6b-00fb-4795-9e6d-6e21068c405e","Type":"ContainerStarted","Data":"6854d569932c45972e7898305b4c8516a8708e091f25072d6da0ce934fdeb244"} Nov 21 14:28:11 crc kubenswrapper[4897]: I1121 14:28:11.965972 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-d656998f4-p8zts" Nov 21 14:28:11 crc kubenswrapper[4897]: I1121 14:28:11.976819 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-glmx7" event={"ID":"09df9b2c-a507-4a35-b646-a2dc45e1eb71","Type":"ContainerStarted","Data":"f41c4a197bf9ac46929b4c2cd7f2bdf68c34484331753473c19be1c1535342e6"} Nov 21 14:28:11 crc kubenswrapper[4897]: I1121 14:28:11.977070 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-glmx7" Nov 21 14:28:11 crc kubenswrapper[4897]: I1121 14:28:11.996037 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-lfbrg" podStartSLOduration=21.870367536 podStartE2EDuration="59.996020214s" podCreationTimestamp="2025-11-21 14:27:12 +0000 UTC" firstStartedPulling="2025-11-21 14:27:15.8886915 +0000 UTC m=+1113.173284975" lastFinishedPulling="2025-11-21 14:27:54.014344178 +0000 UTC m=+1151.298937653" observedRunningTime="2025-11-21 14:28:11.959134766 +0000 UTC m=+1169.243728241" watchObservedRunningTime="2025-11-21 14:28:11.996020214 +0000 UTC m=+1169.280613729" Nov 21 14:28:12 crc kubenswrapper[4897]: I1121 14:28:12.025808 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-l2njs" event={"ID":"8aee4e2c-4b47-41ec-9953-7e73d0bd6524","Type":"ContainerStarted","Data":"4674ac51800b8bf35c8c878bfe8b0ec0a669c404338a68aff6ef2bce28172400"} Nov 21 14:28:12 crc kubenswrapper[4897]: I1121 14:28:12.026153 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-l2njs" Nov 21 14:28:12 crc kubenswrapper[4897]: I1121 14:28:12.042310 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-smlb7" podStartSLOduration=3.960686709 podStartE2EDuration="1m0.042293763s" podCreationTimestamp="2025-11-21 14:27:12 +0000 UTC" firstStartedPulling="2025-11-21 14:27:15.278865107 +0000 UTC m=+1112.563458582" lastFinishedPulling="2025-11-21 14:28:11.360472161 +0000 UTC m=+1168.645065636" observedRunningTime="2025-11-21 14:28:12.027639901 +0000 UTC m=+1169.312233386" watchObservedRunningTime="2025-11-21 14:28:12.042293763 +0000 UTC m=+1169.326887238" Nov 21 14:28:12 crc kubenswrapper[4897]: I1121 14:28:12.048698 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-2dnk2" event={"ID":"3b3d91dd-852c-4b65-a2c9-dbc02e79daaa","Type":"ContainerStarted","Data":"99e67eedd3765db1802dba734895b060a4e1cac4de20d47b40910365c7f6948b"} Nov 21 14:28:12 crc kubenswrapper[4897]: I1121 14:28:12.050011 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-56f54d6746-nrp4j" Nov 21 14:28:12 crc kubenswrapper[4897]: I1121 14:28:12.050049 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-2dnk2" Nov 21 14:28:12 crc kubenswrapper[4897]: I1121 14:28:12.082991 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-csgnr" podStartSLOduration=19.956626027 podStartE2EDuration="1m0.082973672s" podCreationTimestamp="2025-11-21 14:27:12 +0000 UTC" firstStartedPulling="2025-11-21 14:27:13.884986613 +0000 UTC m=+1111.169580088" lastFinishedPulling="2025-11-21 14:27:54.011334258 +0000 UTC m=+1151.295927733" observedRunningTime="2025-11-21 14:28:12.079911891 +0000 UTC m=+1169.364505366" watchObservedRunningTime="2025-11-21 14:28:12.082973672 +0000 UTC m=+1169.367567147" Nov 21 14:28:12 crc kubenswrapper[4897]: I1121 14:28:12.091172 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-kq4ss" podStartSLOduration=22.691806128 podStartE2EDuration="1m0.091155132s" podCreationTimestamp="2025-11-21 14:27:12 +0000 UTC" firstStartedPulling="2025-11-21 14:27:15.235346352 +0000 UTC m=+1112.519939827" lastFinishedPulling="2025-11-21 14:27:52.634695356 +0000 UTC m=+1149.919288831" observedRunningTime="2025-11-21 14:28:12.059937666 +0000 UTC m=+1169.344531141" watchObservedRunningTime="2025-11-21 14:28:12.091155132 +0000 UTC m=+1169.375748597" Nov 21 14:28:12 crc kubenswrapper[4897]: I1121 14:28:12.122349 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-pt5vd" podStartSLOduration=3.991110342 podStartE2EDuration="1m0.122331376s" podCreationTimestamp="2025-11-21 14:27:12 +0000 UTC" firstStartedPulling="2025-11-21 14:27:15.229078823 +0000 UTC m=+1112.513672298" lastFinishedPulling="2025-11-21 14:28:11.360299847 +0000 UTC m=+1168.644893332" observedRunningTime="2025-11-21 14:28:12.110911831 +0000 UTC m=+1169.395505306" watchObservedRunningTime="2025-11-21 14:28:12.122331376 +0000 UTC m=+1169.406924851" Nov 21 14:28:12 crc kubenswrapper[4897]: I1121 14:28:12.131026 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-58f887965d-dhc5w" podStartSLOduration=13.615779558 podStartE2EDuration="1m0.131011359s" podCreationTimestamp="2025-11-21 14:27:12 +0000 UTC" firstStartedPulling="2025-11-21 14:27:15.296124949 +0000 UTC m=+1112.580718424" lastFinishedPulling="2025-11-21 14:28:01.81135671 +0000 UTC m=+1159.095950225" observedRunningTime="2025-11-21 14:28:12.128166073 +0000 UTC m=+1169.412759548" watchObservedRunningTime="2025-11-21 14:28:12.131011359 +0000 UTC m=+1169.415604834" Nov 21 14:28:12 crc kubenswrapper[4897]: I1121 14:28:12.157799 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-glmx7" podStartSLOduration=4.540632931 podStartE2EDuration="1m0.157779285s" podCreationTimestamp="2025-11-21 14:27:12 +0000 UTC" firstStartedPulling="2025-11-21 14:27:15.744411586 +0000 UTC m=+1113.029005061" lastFinishedPulling="2025-11-21 14:28:11.36155794 +0000 UTC m=+1168.646151415" observedRunningTime="2025-11-21 14:28:12.153574973 +0000 UTC m=+1169.438168448" watchObservedRunningTime="2025-11-21 14:28:12.157779285 +0000 UTC m=+1169.442372761" Nov 21 14:28:12 crc kubenswrapper[4897]: I1121 14:28:12.176016 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-2dnk2" podStartSLOduration=3.918134669 podStartE2EDuration="1m0.175997454s" podCreationTimestamp="2025-11-21 14:27:12 +0000 UTC" firstStartedPulling="2025-11-21 14:27:15.229519036 +0000 UTC m=+1112.514112511" lastFinishedPulling="2025-11-21 14:28:11.487381821 +0000 UTC m=+1168.771975296" observedRunningTime="2025-11-21 14:28:12.169976233 +0000 UTC m=+1169.454569718" watchObservedRunningTime="2025-11-21 14:28:12.175997454 +0000 UTC m=+1169.460590929" Nov 21 14:28:12 crc kubenswrapper[4897]: I1121 14:28:12.196320 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-d656998f4-p8zts" podStartSLOduration=21.463347585 podStartE2EDuration="1m0.196304138s" podCreationTimestamp="2025-11-21 14:27:12 +0000 UTC" firstStartedPulling="2025-11-21 14:27:15.280314556 +0000 UTC m=+1112.564908031" lastFinishedPulling="2025-11-21 14:27:54.013271099 +0000 UTC m=+1151.297864584" observedRunningTime="2025-11-21 14:28:12.189820664 +0000 UTC m=+1169.474414139" watchObservedRunningTime="2025-11-21 14:28:12.196304138 +0000 UTC m=+1169.480897613" Nov 21 14:28:12 crc kubenswrapper[4897]: I1121 14:28:12.206698 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-l2njs" podStartSLOduration=31.260117118 podStartE2EDuration="1m0.206684365s" podCreationTimestamp="2025-11-21 14:27:12 +0000 UTC" firstStartedPulling="2025-11-21 14:27:14.568951353 +0000 UTC m=+1111.853544828" lastFinishedPulling="2025-11-21 14:27:43.5155186 +0000 UTC m=+1140.800112075" observedRunningTime="2025-11-21 14:28:12.205387621 +0000 UTC m=+1169.489981096" watchObservedRunningTime="2025-11-21 14:28:12.206684365 +0000 UTC m=+1169.491277830" Nov 21 14:28:13 crc kubenswrapper[4897]: I1121 14:28:13.050478 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-kq4ss" Nov 21 14:28:13 crc kubenswrapper[4897]: I1121 14:28:13.055108 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-phs97" event={"ID":"d1fc2027-5a14-4de3-8c0e-c4ecee66c40b","Type":"ContainerStarted","Data":"87b6d38a8ef6bbbaca2b3a6a40ac289ac284f1b66f2abb10acd0c77fb0a1ce43"} Nov 21 14:28:13 crc kubenswrapper[4897]: I1121 14:28:13.055454 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-phs97" Nov 21 14:28:13 crc kubenswrapper[4897]: I1121 14:28:13.056952 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-2grhg" event={"ID":"77f4c339-9ff1-4943-922d-88d81a73a1f1","Type":"ContainerStarted","Data":"11c03c24f46148fc7577285dd40c070dc859872c96f7b182eeedf65cd039aa81"} Nov 21 14:28:13 crc kubenswrapper[4897]: I1121 14:28:13.057349 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-2grhg" Nov 21 14:28:13 crc kubenswrapper[4897]: I1121 14:28:13.059033 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-gzzl8" event={"ID":"d4c39700-beb8-4272-9f54-1c501332af0d","Type":"ContainerStarted","Data":"ca39ffd54f7935c301ebb724d71485761a08b543ce0eba8ae13f0c5e6f3ff300"} Nov 21 14:28:13 crc kubenswrapper[4897]: I1121 14:28:13.059348 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-gzzl8" Nov 21 14:28:13 crc kubenswrapper[4897]: I1121 14:28:13.061393 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-bh4rl" event={"ID":"499cc09f-fafe-47c2-87ac-bc00d79e949e","Type":"ContainerStarted","Data":"260c0d16962adc346bd845a461717a0f8fc205998b770f083de22c1fc680be0f"} Nov 21 14:28:13 crc kubenswrapper[4897]: I1121 14:28:13.065088 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-56f54d6746-nrp4j" Nov 21 14:28:13 crc kubenswrapper[4897]: I1121 14:28:13.066232 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-d656998f4-p8zts" Nov 21 14:28:13 crc kubenswrapper[4897]: I1121 14:28:13.066303 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-csgnr" Nov 21 14:28:13 crc kubenswrapper[4897]: I1121 14:28:13.066345 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-cr58g" Nov 21 14:28:13 crc kubenswrapper[4897]: I1121 14:28:13.068513 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-lfbrg" Nov 21 14:28:13 crc kubenswrapper[4897]: I1121 14:28:13.132405 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-bh4rl" podStartSLOduration=4.908125706 podStartE2EDuration="1m1.13238915s" podCreationTimestamp="2025-11-21 14:27:12 +0000 UTC" firstStartedPulling="2025-11-21 14:27:15.74528392 +0000 UTC m=+1113.029877395" lastFinishedPulling="2025-11-21 14:28:11.969547364 +0000 UTC m=+1169.254140839" observedRunningTime="2025-11-21 14:28:13.130311124 +0000 UTC m=+1170.414904609" watchObservedRunningTime="2025-11-21 14:28:13.13238915 +0000 UTC m=+1170.416982625" Nov 21 14:28:13 crc kubenswrapper[4897]: I1121 14:28:13.177343 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-phs97" podStartSLOduration=4.694488823 podStartE2EDuration="1m1.177325294s" podCreationTimestamp="2025-11-21 14:27:12 +0000 UTC" firstStartedPulling="2025-11-21 14:27:15.28047284 +0000 UTC m=+1112.565066315" lastFinishedPulling="2025-11-21 14:28:11.763309311 +0000 UTC m=+1169.047902786" observedRunningTime="2025-11-21 14:28:13.17533152 +0000 UTC m=+1170.459924995" watchObservedRunningTime="2025-11-21 14:28:13.177325294 +0000 UTC m=+1170.461918769" Nov 21 14:28:13 crc kubenswrapper[4897]: I1121 14:28:13.193130 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-gzzl8" podStartSLOduration=4.683064807 podStartE2EDuration="1m1.193116106s" podCreationTimestamp="2025-11-21 14:27:12 +0000 UTC" firstStartedPulling="2025-11-21 14:27:15.256093738 +0000 UTC m=+1112.540687213" lastFinishedPulling="2025-11-21 14:28:11.766145037 +0000 UTC m=+1169.050738512" observedRunningTime="2025-11-21 14:28:13.192030367 +0000 UTC m=+1170.476623842" watchObservedRunningTime="2025-11-21 14:28:13.193116106 +0000 UTC m=+1170.477709581" Nov 21 14:28:13 crc kubenswrapper[4897]: I1121 14:28:13.235581 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-bh4rl" Nov 21 14:28:13 crc kubenswrapper[4897]: I1121 14:28:13.238581 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-2grhg" podStartSLOduration=4.733098797 podStartE2EDuration="1m1.238557593s" podCreationTimestamp="2025-11-21 14:27:12 +0000 UTC" firstStartedPulling="2025-11-21 14:27:15.257387292 +0000 UTC m=+1112.541980767" lastFinishedPulling="2025-11-21 14:28:11.762846088 +0000 UTC m=+1169.047439563" observedRunningTime="2025-11-21 14:28:13.225034632 +0000 UTC m=+1170.509628107" watchObservedRunningTime="2025-11-21 14:28:13.238557593 +0000 UTC m=+1170.523151068" Nov 21 14:28:13 crc kubenswrapper[4897]: I1121 14:28:13.268128 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-2sqhj" Nov 21 14:28:13 crc kubenswrapper[4897]: I1121 14:28:13.967386 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-b4c496f69-9fkd7" Nov 21 14:28:13 crc kubenswrapper[4897]: I1121 14:28:13.970203 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-b4c496f69-9fkd7" Nov 21 14:28:19 crc kubenswrapper[4897]: E1121 14:28:19.092349 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-2422v" podUID="ef9b1721-eadf-4f29-a175-008292188ef2" Nov 21 14:28:20 crc kubenswrapper[4897]: E1121 14:28:20.090753 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:4838402d41d42c56613d43dc5041aae475a2b18e6172491d6c4d4a78a580697f\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-jrlfk" podUID="40748f20-acda-4ced-ac8b-2e741dae0c66" Nov 21 14:28:22 crc kubenswrapper[4897]: E1121 14:28:22.223436 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.70:5001/openstack-k8s-operators/telemetry-operator:6280f54c4d86e239852669b9aa334e584f1fe080\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-77797865fc-9n8nt" podUID="1557d794-d052-44ad-b908-4766d1abe234" Nov 21 14:28:22 crc kubenswrapper[4897]: I1121 14:28:22.867404 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-7969689c84-xlbhh" Nov 21 14:28:22 crc kubenswrapper[4897]: I1121 14:28:22.966024 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-l2njs" Nov 21 14:28:23 crc kubenswrapper[4897]: I1121 14:28:23.044066 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-gzzl8" Nov 21 14:28:23 crc kubenswrapper[4897]: I1121 14:28:23.155131 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-58f887965d-dhc5w" Nov 21 14:28:23 crc kubenswrapper[4897]: I1121 14:28:23.210826 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-2grhg" Nov 21 14:28:23 crc kubenswrapper[4897]: I1121 14:28:23.235517 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-2dnk2" Nov 21 14:28:23 crc kubenswrapper[4897]: I1121 14:28:23.239990 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-bh4rl" Nov 21 14:28:23 crc kubenswrapper[4897]: I1121 14:28:23.295794 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-pt5vd" Nov 21 14:28:23 crc kubenswrapper[4897]: I1121 14:28:23.339093 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-phs97" Nov 21 14:28:23 crc kubenswrapper[4897]: I1121 14:28:23.504143 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-smlb7" Nov 21 14:28:23 crc kubenswrapper[4897]: I1121 14:28:23.724891 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-glmx7" Nov 21 14:28:34 crc kubenswrapper[4897]: I1121 14:28:34.370487 4897 patch_prober.go:28] interesting pod/machine-config-daemon-krv5b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 21 14:28:34 crc kubenswrapper[4897]: I1121 14:28:34.371188 4897 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 21 14:28:34 crc kubenswrapper[4897]: I1121 14:28:34.371242 4897 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" Nov 21 14:28:34 crc kubenswrapper[4897]: I1121 14:28:34.372030 4897 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"cdc84dcb39834c826868fb54c0e5d0aa88dbf0ae2b0ceea30db993577fc2bded"} pod="openshift-machine-config-operator/machine-config-daemon-krv5b" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 21 14:28:34 crc kubenswrapper[4897]: I1121 14:28:34.372094 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" containerName="machine-config-daemon" containerID="cri-o://cdc84dcb39834c826868fb54c0e5d0aa88dbf0ae2b0ceea30db993577fc2bded" gracePeriod=600 Nov 21 14:28:35 crc kubenswrapper[4897]: I1121 14:28:35.245813 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-2422v" event={"ID":"ef9b1721-eadf-4f29-a175-008292188ef2","Type":"ContainerStarted","Data":"e71fac9f0c0d8766bfce2767dc9068307216b05c7a4bacfd8a8e716a80ec284c"} Nov 21 14:28:35 crc kubenswrapper[4897]: I1121 14:28:35.251734 4897 generic.go:334] "Generic (PLEG): container finished" podID="e7670227-d280-4847-b882-754429f56b0f" containerID="cdc84dcb39834c826868fb54c0e5d0aa88dbf0ae2b0ceea30db993577fc2bded" exitCode=0 Nov 21 14:28:35 crc kubenswrapper[4897]: I1121 14:28:35.251779 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" event={"ID":"e7670227-d280-4847-b882-754429f56b0f","Type":"ContainerDied","Data":"cdc84dcb39834c826868fb54c0e5d0aa88dbf0ae2b0ceea30db993577fc2bded"} Nov 21 14:28:35 crc kubenswrapper[4897]: I1121 14:28:35.251802 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" event={"ID":"e7670227-d280-4847-b882-754429f56b0f","Type":"ContainerStarted","Data":"fbff5a4c19b1cfbe0c7e57eb05f0093aa3829fbabee820a190f916e0ce4fe6b9"} Nov 21 14:28:35 crc kubenswrapper[4897]: I1121 14:28:35.251818 4897 scope.go:117] "RemoveContainer" containerID="6f8d8438a69d75aad7ccb460842fd43be11b0196e6779f692d047f7ed8f74e83" Nov 21 14:28:35 crc kubenswrapper[4897]: I1121 14:28:35.267360 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-2422v" podStartSLOduration=3.347396717 podStartE2EDuration="1m22.267345687s" podCreationTimestamp="2025-11-21 14:27:13 +0000 UTC" firstStartedPulling="2025-11-21 14:27:15.828852688 +0000 UTC m=+1113.113446173" lastFinishedPulling="2025-11-21 14:28:34.748801668 +0000 UTC m=+1192.033395143" observedRunningTime="2025-11-21 14:28:35.264706926 +0000 UTC m=+1192.549300401" watchObservedRunningTime="2025-11-21 14:28:35.267345687 +0000 UTC m=+1192.551939162" Nov 21 14:28:37 crc kubenswrapper[4897]: I1121 14:28:37.283686 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-jrlfk" event={"ID":"40748f20-acda-4ced-ac8b-2e741dae0c66","Type":"ContainerStarted","Data":"00233cfc4f53cb22aa0120dc46e8c0b13d62fd9852bf9d6ea3cfc979aac427e9"} Nov 21 14:28:37 crc kubenswrapper[4897]: I1121 14:28:37.285383 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-jrlfk" Nov 21 14:28:37 crc kubenswrapper[4897]: I1121 14:28:37.304110 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-jrlfk" podStartSLOduration=5.122915337 podStartE2EDuration="1m25.304092818s" podCreationTimestamp="2025-11-21 14:27:12 +0000 UTC" firstStartedPulling="2025-11-21 14:27:15.795753261 +0000 UTC m=+1113.080346736" lastFinishedPulling="2025-11-21 14:28:35.976930722 +0000 UTC m=+1193.261524217" observedRunningTime="2025-11-21 14:28:37.302346942 +0000 UTC m=+1194.586940427" watchObservedRunningTime="2025-11-21 14:28:37.304092818 +0000 UTC m=+1194.588686303" Nov 21 14:28:38 crc kubenswrapper[4897]: I1121 14:28:38.294337 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-77797865fc-9n8nt" event={"ID":"1557d794-d052-44ad-b908-4766d1abe234","Type":"ContainerStarted","Data":"5e300122c388f379c23706ddd28a7bb6c4a437ac24830fd6f795c25c1b615018"} Nov 21 14:28:38 crc kubenswrapper[4897]: I1121 14:28:38.294987 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-77797865fc-9n8nt" Nov 21 14:28:38 crc kubenswrapper[4897]: I1121 14:28:38.308964 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-77797865fc-9n8nt" podStartSLOduration=4.322000496 podStartE2EDuration="1m26.308947312s" podCreationTimestamp="2025-11-21 14:27:12 +0000 UTC" firstStartedPulling="2025-11-21 14:27:15.749930604 +0000 UTC m=+1113.034524079" lastFinishedPulling="2025-11-21 14:28:37.73687741 +0000 UTC m=+1195.021470895" observedRunningTime="2025-11-21 14:28:38.308422089 +0000 UTC m=+1195.593015564" watchObservedRunningTime="2025-11-21 14:28:38.308947312 +0000 UTC m=+1195.593540787" Nov 21 14:28:43 crc kubenswrapper[4897]: I1121 14:28:43.688739 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-77797865fc-9n8nt" Nov 21 14:28:44 crc kubenswrapper[4897]: I1121 14:28:44.130118 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-jrlfk" Nov 21 14:29:00 crc kubenswrapper[4897]: I1121 14:29:00.217436 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-fpb49"] Nov 21 14:29:00 crc kubenswrapper[4897]: I1121 14:29:00.221886 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-fpb49" Nov 21 14:29:00 crc kubenswrapper[4897]: I1121 14:29:00.223983 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-7nj2j" Nov 21 14:29:00 crc kubenswrapper[4897]: I1121 14:29:00.224179 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Nov 21 14:29:00 crc kubenswrapper[4897]: I1121 14:29:00.224437 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Nov 21 14:29:00 crc kubenswrapper[4897]: I1121 14:29:00.224940 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Nov 21 14:29:00 crc kubenswrapper[4897]: I1121 14:29:00.235993 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-fpb49"] Nov 21 14:29:00 crc kubenswrapper[4897]: I1121 14:29:00.295685 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-whhtg"] Nov 21 14:29:00 crc kubenswrapper[4897]: I1121 14:29:00.299578 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-whhtg" Nov 21 14:29:00 crc kubenswrapper[4897]: I1121 14:29:00.302448 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Nov 21 14:29:00 crc kubenswrapper[4897]: I1121 14:29:00.305568 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-whhtg"] Nov 21 14:29:00 crc kubenswrapper[4897]: I1121 14:29:00.350598 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7522ba8b-cc71-459f-9f6a-fbd0f646dce3-config\") pod \"dnsmasq-dns-675f4bcbfc-fpb49\" (UID: \"7522ba8b-cc71-459f-9f6a-fbd0f646dce3\") " pod="openstack/dnsmasq-dns-675f4bcbfc-fpb49" Nov 21 14:29:00 crc kubenswrapper[4897]: I1121 14:29:00.350901 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bw8j8\" (UniqueName: \"kubernetes.io/projected/7522ba8b-cc71-459f-9f6a-fbd0f646dce3-kube-api-access-bw8j8\") pod \"dnsmasq-dns-675f4bcbfc-fpb49\" (UID: \"7522ba8b-cc71-459f-9f6a-fbd0f646dce3\") " pod="openstack/dnsmasq-dns-675f4bcbfc-fpb49" Nov 21 14:29:00 crc kubenswrapper[4897]: I1121 14:29:00.452029 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s2m5c\" (UniqueName: \"kubernetes.io/projected/f6215eb2-5fde-4880-b101-a3be3a015d93-kube-api-access-s2m5c\") pod \"dnsmasq-dns-78dd6ddcc-whhtg\" (UID: \"f6215eb2-5fde-4880-b101-a3be3a015d93\") " pod="openstack/dnsmasq-dns-78dd6ddcc-whhtg" Nov 21 14:29:00 crc kubenswrapper[4897]: I1121 14:29:00.452101 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bw8j8\" (UniqueName: \"kubernetes.io/projected/7522ba8b-cc71-459f-9f6a-fbd0f646dce3-kube-api-access-bw8j8\") pod \"dnsmasq-dns-675f4bcbfc-fpb49\" (UID: \"7522ba8b-cc71-459f-9f6a-fbd0f646dce3\") " pod="openstack/dnsmasq-dns-675f4bcbfc-fpb49" Nov 21 14:29:00 crc kubenswrapper[4897]: I1121 14:29:00.452272 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f6215eb2-5fde-4880-b101-a3be3a015d93-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-whhtg\" (UID: \"f6215eb2-5fde-4880-b101-a3be3a015d93\") " pod="openstack/dnsmasq-dns-78dd6ddcc-whhtg" Nov 21 14:29:00 crc kubenswrapper[4897]: I1121 14:29:00.452356 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f6215eb2-5fde-4880-b101-a3be3a015d93-config\") pod \"dnsmasq-dns-78dd6ddcc-whhtg\" (UID: \"f6215eb2-5fde-4880-b101-a3be3a015d93\") " pod="openstack/dnsmasq-dns-78dd6ddcc-whhtg" Nov 21 14:29:00 crc kubenswrapper[4897]: I1121 14:29:00.452563 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7522ba8b-cc71-459f-9f6a-fbd0f646dce3-config\") pod \"dnsmasq-dns-675f4bcbfc-fpb49\" (UID: \"7522ba8b-cc71-459f-9f6a-fbd0f646dce3\") " pod="openstack/dnsmasq-dns-675f4bcbfc-fpb49" Nov 21 14:29:00 crc kubenswrapper[4897]: I1121 14:29:00.453458 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7522ba8b-cc71-459f-9f6a-fbd0f646dce3-config\") pod \"dnsmasq-dns-675f4bcbfc-fpb49\" (UID: \"7522ba8b-cc71-459f-9f6a-fbd0f646dce3\") " pod="openstack/dnsmasq-dns-675f4bcbfc-fpb49" Nov 21 14:29:00 crc kubenswrapper[4897]: I1121 14:29:00.480939 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bw8j8\" (UniqueName: \"kubernetes.io/projected/7522ba8b-cc71-459f-9f6a-fbd0f646dce3-kube-api-access-bw8j8\") pod \"dnsmasq-dns-675f4bcbfc-fpb49\" (UID: \"7522ba8b-cc71-459f-9f6a-fbd0f646dce3\") " pod="openstack/dnsmasq-dns-675f4bcbfc-fpb49" Nov 21 14:29:00 crc kubenswrapper[4897]: I1121 14:29:00.548145 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-fpb49" Nov 21 14:29:00 crc kubenswrapper[4897]: I1121 14:29:00.553953 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f6215eb2-5fde-4880-b101-a3be3a015d93-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-whhtg\" (UID: \"f6215eb2-5fde-4880-b101-a3be3a015d93\") " pod="openstack/dnsmasq-dns-78dd6ddcc-whhtg" Nov 21 14:29:00 crc kubenswrapper[4897]: I1121 14:29:00.554011 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f6215eb2-5fde-4880-b101-a3be3a015d93-config\") pod \"dnsmasq-dns-78dd6ddcc-whhtg\" (UID: \"f6215eb2-5fde-4880-b101-a3be3a015d93\") " pod="openstack/dnsmasq-dns-78dd6ddcc-whhtg" Nov 21 14:29:00 crc kubenswrapper[4897]: I1121 14:29:00.554122 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2m5c\" (UniqueName: \"kubernetes.io/projected/f6215eb2-5fde-4880-b101-a3be3a015d93-kube-api-access-s2m5c\") pod \"dnsmasq-dns-78dd6ddcc-whhtg\" (UID: \"f6215eb2-5fde-4880-b101-a3be3a015d93\") " pod="openstack/dnsmasq-dns-78dd6ddcc-whhtg" Nov 21 14:29:00 crc kubenswrapper[4897]: I1121 14:29:00.556401 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f6215eb2-5fde-4880-b101-a3be3a015d93-config\") pod \"dnsmasq-dns-78dd6ddcc-whhtg\" (UID: \"f6215eb2-5fde-4880-b101-a3be3a015d93\") " pod="openstack/dnsmasq-dns-78dd6ddcc-whhtg" Nov 21 14:29:00 crc kubenswrapper[4897]: I1121 14:29:00.565577 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f6215eb2-5fde-4880-b101-a3be3a015d93-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-whhtg\" (UID: \"f6215eb2-5fde-4880-b101-a3be3a015d93\") " pod="openstack/dnsmasq-dns-78dd6ddcc-whhtg" Nov 21 14:29:00 crc kubenswrapper[4897]: I1121 14:29:00.575810 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2m5c\" (UniqueName: \"kubernetes.io/projected/f6215eb2-5fde-4880-b101-a3be3a015d93-kube-api-access-s2m5c\") pod \"dnsmasq-dns-78dd6ddcc-whhtg\" (UID: \"f6215eb2-5fde-4880-b101-a3be3a015d93\") " pod="openstack/dnsmasq-dns-78dd6ddcc-whhtg" Nov 21 14:29:00 crc kubenswrapper[4897]: I1121 14:29:00.618796 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-whhtg" Nov 21 14:29:01 crc kubenswrapper[4897]: I1121 14:29:01.067379 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-fpb49"] Nov 21 14:29:01 crc kubenswrapper[4897]: W1121 14:29:01.074656 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7522ba8b_cc71_459f_9f6a_fbd0f646dce3.slice/crio-5ef9f69ff7ec34c24c0d19c2ed4fa3ae7f9f06486e4607f69a6c793476e08763 WatchSource:0}: Error finding container 5ef9f69ff7ec34c24c0d19c2ed4fa3ae7f9f06486e4607f69a6c793476e08763: Status 404 returned error can't find the container with id 5ef9f69ff7ec34c24c0d19c2ed4fa3ae7f9f06486e4607f69a6c793476e08763 Nov 21 14:29:01 crc kubenswrapper[4897]: I1121 14:29:01.224566 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-whhtg"] Nov 21 14:29:01 crc kubenswrapper[4897]: W1121 14:29:01.227920 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf6215eb2_5fde_4880_b101_a3be3a015d93.slice/crio-753f2f00a3cf405c991e722c415ddc0a184c0c77525a31247e0f551ae714e181 WatchSource:0}: Error finding container 753f2f00a3cf405c991e722c415ddc0a184c0c77525a31247e0f551ae714e181: Status 404 returned error can't find the container with id 753f2f00a3cf405c991e722c415ddc0a184c0c77525a31247e0f551ae714e181 Nov 21 14:29:01 crc kubenswrapper[4897]: I1121 14:29:01.488698 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-fpb49" event={"ID":"7522ba8b-cc71-459f-9f6a-fbd0f646dce3","Type":"ContainerStarted","Data":"5ef9f69ff7ec34c24c0d19c2ed4fa3ae7f9f06486e4607f69a6c793476e08763"} Nov 21 14:29:01 crc kubenswrapper[4897]: I1121 14:29:01.491747 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-whhtg" event={"ID":"f6215eb2-5fde-4880-b101-a3be3a015d93","Type":"ContainerStarted","Data":"753f2f00a3cf405c991e722c415ddc0a184c0c77525a31247e0f551ae714e181"} Nov 21 14:29:03 crc kubenswrapper[4897]: I1121 14:29:03.371324 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-fpb49"] Nov 21 14:29:03 crc kubenswrapper[4897]: I1121 14:29:03.412392 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-khkvq"] Nov 21 14:29:03 crc kubenswrapper[4897]: I1121 14:29:03.414857 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-khkvq" Nov 21 14:29:03 crc kubenswrapper[4897]: I1121 14:29:03.440396 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-khkvq"] Nov 21 14:29:03 crc kubenswrapper[4897]: I1121 14:29:03.520856 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6bf04b77-0ef9-42f0-8a02-8c4c7568e390-dns-svc\") pod \"dnsmasq-dns-666b6646f7-khkvq\" (UID: \"6bf04b77-0ef9-42f0-8a02-8c4c7568e390\") " pod="openstack/dnsmasq-dns-666b6646f7-khkvq" Nov 21 14:29:03 crc kubenswrapper[4897]: I1121 14:29:03.520973 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6b5qf\" (UniqueName: \"kubernetes.io/projected/6bf04b77-0ef9-42f0-8a02-8c4c7568e390-kube-api-access-6b5qf\") pod \"dnsmasq-dns-666b6646f7-khkvq\" (UID: \"6bf04b77-0ef9-42f0-8a02-8c4c7568e390\") " pod="openstack/dnsmasq-dns-666b6646f7-khkvq" Nov 21 14:29:03 crc kubenswrapper[4897]: I1121 14:29:03.521262 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6bf04b77-0ef9-42f0-8a02-8c4c7568e390-config\") pod \"dnsmasq-dns-666b6646f7-khkvq\" (UID: \"6bf04b77-0ef9-42f0-8a02-8c4c7568e390\") " pod="openstack/dnsmasq-dns-666b6646f7-khkvq" Nov 21 14:29:03 crc kubenswrapper[4897]: I1121 14:29:03.623523 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6bf04b77-0ef9-42f0-8a02-8c4c7568e390-dns-svc\") pod \"dnsmasq-dns-666b6646f7-khkvq\" (UID: \"6bf04b77-0ef9-42f0-8a02-8c4c7568e390\") " pod="openstack/dnsmasq-dns-666b6646f7-khkvq" Nov 21 14:29:03 crc kubenswrapper[4897]: I1121 14:29:03.623627 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6b5qf\" (UniqueName: \"kubernetes.io/projected/6bf04b77-0ef9-42f0-8a02-8c4c7568e390-kube-api-access-6b5qf\") pod \"dnsmasq-dns-666b6646f7-khkvq\" (UID: \"6bf04b77-0ef9-42f0-8a02-8c4c7568e390\") " pod="openstack/dnsmasq-dns-666b6646f7-khkvq" Nov 21 14:29:03 crc kubenswrapper[4897]: I1121 14:29:03.623668 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6bf04b77-0ef9-42f0-8a02-8c4c7568e390-config\") pod \"dnsmasq-dns-666b6646f7-khkvq\" (UID: \"6bf04b77-0ef9-42f0-8a02-8c4c7568e390\") " pod="openstack/dnsmasq-dns-666b6646f7-khkvq" Nov 21 14:29:03 crc kubenswrapper[4897]: I1121 14:29:03.624683 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6bf04b77-0ef9-42f0-8a02-8c4c7568e390-config\") pod \"dnsmasq-dns-666b6646f7-khkvq\" (UID: \"6bf04b77-0ef9-42f0-8a02-8c4c7568e390\") " pod="openstack/dnsmasq-dns-666b6646f7-khkvq" Nov 21 14:29:03 crc kubenswrapper[4897]: I1121 14:29:03.625243 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6bf04b77-0ef9-42f0-8a02-8c4c7568e390-dns-svc\") pod \"dnsmasq-dns-666b6646f7-khkvq\" (UID: \"6bf04b77-0ef9-42f0-8a02-8c4c7568e390\") " pod="openstack/dnsmasq-dns-666b6646f7-khkvq" Nov 21 14:29:03 crc kubenswrapper[4897]: I1121 14:29:03.652648 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6b5qf\" (UniqueName: \"kubernetes.io/projected/6bf04b77-0ef9-42f0-8a02-8c4c7568e390-kube-api-access-6b5qf\") pod \"dnsmasq-dns-666b6646f7-khkvq\" (UID: \"6bf04b77-0ef9-42f0-8a02-8c4c7568e390\") " pod="openstack/dnsmasq-dns-666b6646f7-khkvq" Nov 21 14:29:03 crc kubenswrapper[4897]: I1121 14:29:03.772126 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-khkvq" Nov 21 14:29:03 crc kubenswrapper[4897]: I1121 14:29:03.933411 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-whhtg"] Nov 21 14:29:03 crc kubenswrapper[4897]: I1121 14:29:03.975956 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-8f9xc"] Nov 21 14:29:03 crc kubenswrapper[4897]: I1121 14:29:03.982805 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-8f9xc" Nov 21 14:29:04 crc kubenswrapper[4897]: I1121 14:29:04.004022 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-8f9xc"] Nov 21 14:29:04 crc kubenswrapper[4897]: I1121 14:29:04.138239 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ecceef5c-b1e5-4c3c-97b3-d805494cfba3-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-8f9xc\" (UID: \"ecceef5c-b1e5-4c3c-97b3-d805494cfba3\") " pod="openstack/dnsmasq-dns-57d769cc4f-8f9xc" Nov 21 14:29:04 crc kubenswrapper[4897]: I1121 14:29:04.139857 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ecceef5c-b1e5-4c3c-97b3-d805494cfba3-config\") pod \"dnsmasq-dns-57d769cc4f-8f9xc\" (UID: \"ecceef5c-b1e5-4c3c-97b3-d805494cfba3\") " pod="openstack/dnsmasq-dns-57d769cc4f-8f9xc" Nov 21 14:29:04 crc kubenswrapper[4897]: I1121 14:29:04.139900 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rqwlp\" (UniqueName: \"kubernetes.io/projected/ecceef5c-b1e5-4c3c-97b3-d805494cfba3-kube-api-access-rqwlp\") pod \"dnsmasq-dns-57d769cc4f-8f9xc\" (UID: \"ecceef5c-b1e5-4c3c-97b3-d805494cfba3\") " pod="openstack/dnsmasq-dns-57d769cc4f-8f9xc" Nov 21 14:29:04 crc kubenswrapper[4897]: I1121 14:29:04.241968 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ecceef5c-b1e5-4c3c-97b3-d805494cfba3-config\") pod \"dnsmasq-dns-57d769cc4f-8f9xc\" (UID: \"ecceef5c-b1e5-4c3c-97b3-d805494cfba3\") " pod="openstack/dnsmasq-dns-57d769cc4f-8f9xc" Nov 21 14:29:04 crc kubenswrapper[4897]: I1121 14:29:04.242026 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rqwlp\" (UniqueName: \"kubernetes.io/projected/ecceef5c-b1e5-4c3c-97b3-d805494cfba3-kube-api-access-rqwlp\") pod \"dnsmasq-dns-57d769cc4f-8f9xc\" (UID: \"ecceef5c-b1e5-4c3c-97b3-d805494cfba3\") " pod="openstack/dnsmasq-dns-57d769cc4f-8f9xc" Nov 21 14:29:04 crc kubenswrapper[4897]: I1121 14:29:04.242062 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ecceef5c-b1e5-4c3c-97b3-d805494cfba3-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-8f9xc\" (UID: \"ecceef5c-b1e5-4c3c-97b3-d805494cfba3\") " pod="openstack/dnsmasq-dns-57d769cc4f-8f9xc" Nov 21 14:29:04 crc kubenswrapper[4897]: I1121 14:29:04.244226 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ecceef5c-b1e5-4c3c-97b3-d805494cfba3-config\") pod \"dnsmasq-dns-57d769cc4f-8f9xc\" (UID: \"ecceef5c-b1e5-4c3c-97b3-d805494cfba3\") " pod="openstack/dnsmasq-dns-57d769cc4f-8f9xc" Nov 21 14:29:04 crc kubenswrapper[4897]: I1121 14:29:04.245112 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ecceef5c-b1e5-4c3c-97b3-d805494cfba3-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-8f9xc\" (UID: \"ecceef5c-b1e5-4c3c-97b3-d805494cfba3\") " pod="openstack/dnsmasq-dns-57d769cc4f-8f9xc" Nov 21 14:29:04 crc kubenswrapper[4897]: I1121 14:29:04.278800 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rqwlp\" (UniqueName: \"kubernetes.io/projected/ecceef5c-b1e5-4c3c-97b3-d805494cfba3-kube-api-access-rqwlp\") pod \"dnsmasq-dns-57d769cc4f-8f9xc\" (UID: \"ecceef5c-b1e5-4c3c-97b3-d805494cfba3\") " pod="openstack/dnsmasq-dns-57d769cc4f-8f9xc" Nov 21 14:29:04 crc kubenswrapper[4897]: I1121 14:29:04.340991 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-8f9xc" Nov 21 14:29:04 crc kubenswrapper[4897]: I1121 14:29:04.540485 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-khkvq"] Nov 21 14:29:04 crc kubenswrapper[4897]: I1121 14:29:04.613001 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Nov 21 14:29:04 crc kubenswrapper[4897]: I1121 14:29:04.614548 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 21 14:29:04 crc kubenswrapper[4897]: I1121 14:29:04.623359 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Nov 21 14:29:04 crc kubenswrapper[4897]: I1121 14:29:04.623670 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-kksgw" Nov 21 14:29:04 crc kubenswrapper[4897]: I1121 14:29:04.623869 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Nov 21 14:29:04 crc kubenswrapper[4897]: I1121 14:29:04.624012 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Nov 21 14:29:04 crc kubenswrapper[4897]: I1121 14:29:04.624221 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Nov 21 14:29:04 crc kubenswrapper[4897]: I1121 14:29:04.624458 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Nov 21 14:29:04 crc kubenswrapper[4897]: I1121 14:29:04.629585 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 21 14:29:04 crc kubenswrapper[4897]: I1121 14:29:04.633452 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Nov 21 14:29:04 crc kubenswrapper[4897]: I1121 14:29:04.752144 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b513e4ee-7c30-4942-b997-4d52414e1ec0-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"b513e4ee-7c30-4942-b997-4d52414e1ec0\") " pod="openstack/rabbitmq-server-0" Nov 21 14:29:04 crc kubenswrapper[4897]: I1121 14:29:04.752249 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b513e4ee-7c30-4942-b997-4d52414e1ec0-pod-info\") pod \"rabbitmq-server-0\" (UID: \"b513e4ee-7c30-4942-b997-4d52414e1ec0\") " pod="openstack/rabbitmq-server-0" Nov 21 14:29:04 crc kubenswrapper[4897]: I1121 14:29:04.752289 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b513e4ee-7c30-4942-b997-4d52414e1ec0-config-data\") pod \"rabbitmq-server-0\" (UID: \"b513e4ee-7c30-4942-b997-4d52414e1ec0\") " pod="openstack/rabbitmq-server-0" Nov 21 14:29:04 crc kubenswrapper[4897]: I1121 14:29:04.752427 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"b513e4ee-7c30-4942-b997-4d52414e1ec0\") " pod="openstack/rabbitmq-server-0" Nov 21 14:29:04 crc kubenswrapper[4897]: I1121 14:29:04.752551 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b513e4ee-7c30-4942-b997-4d52414e1ec0-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"b513e4ee-7c30-4942-b997-4d52414e1ec0\") " pod="openstack/rabbitmq-server-0" Nov 21 14:29:04 crc kubenswrapper[4897]: I1121 14:29:04.756827 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b513e4ee-7c30-4942-b997-4d52414e1ec0-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"b513e4ee-7c30-4942-b997-4d52414e1ec0\") " pod="openstack/rabbitmq-server-0" Nov 21 14:29:04 crc kubenswrapper[4897]: I1121 14:29:04.756905 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b513e4ee-7c30-4942-b997-4d52414e1ec0-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"b513e4ee-7c30-4942-b997-4d52414e1ec0\") " pod="openstack/rabbitmq-server-0" Nov 21 14:29:04 crc kubenswrapper[4897]: I1121 14:29:04.756960 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b513e4ee-7c30-4942-b997-4d52414e1ec0-server-conf\") pod \"rabbitmq-server-0\" (UID: \"b513e4ee-7c30-4942-b997-4d52414e1ec0\") " pod="openstack/rabbitmq-server-0" Nov 21 14:29:04 crc kubenswrapper[4897]: I1121 14:29:04.757030 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s2gt5\" (UniqueName: \"kubernetes.io/projected/b513e4ee-7c30-4942-b997-4d52414e1ec0-kube-api-access-s2gt5\") pod \"rabbitmq-server-0\" (UID: \"b513e4ee-7c30-4942-b997-4d52414e1ec0\") " pod="openstack/rabbitmq-server-0" Nov 21 14:29:04 crc kubenswrapper[4897]: I1121 14:29:04.757076 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b513e4ee-7c30-4942-b997-4d52414e1ec0-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"b513e4ee-7c30-4942-b997-4d52414e1ec0\") " pod="openstack/rabbitmq-server-0" Nov 21 14:29:04 crc kubenswrapper[4897]: I1121 14:29:04.757153 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b513e4ee-7c30-4942-b997-4d52414e1ec0-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"b513e4ee-7c30-4942-b997-4d52414e1ec0\") " pod="openstack/rabbitmq-server-0" Nov 21 14:29:04 crc kubenswrapper[4897]: I1121 14:29:04.858772 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"b513e4ee-7c30-4942-b997-4d52414e1ec0\") " pod="openstack/rabbitmq-server-0" Nov 21 14:29:04 crc kubenswrapper[4897]: I1121 14:29:04.858866 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b513e4ee-7c30-4942-b997-4d52414e1ec0-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"b513e4ee-7c30-4942-b997-4d52414e1ec0\") " pod="openstack/rabbitmq-server-0" Nov 21 14:29:04 crc kubenswrapper[4897]: I1121 14:29:04.858907 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b513e4ee-7c30-4942-b997-4d52414e1ec0-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"b513e4ee-7c30-4942-b997-4d52414e1ec0\") " pod="openstack/rabbitmq-server-0" Nov 21 14:29:04 crc kubenswrapper[4897]: I1121 14:29:04.858948 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b513e4ee-7c30-4942-b997-4d52414e1ec0-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"b513e4ee-7c30-4942-b997-4d52414e1ec0\") " pod="openstack/rabbitmq-server-0" Nov 21 14:29:04 crc kubenswrapper[4897]: I1121 14:29:04.858984 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b513e4ee-7c30-4942-b997-4d52414e1ec0-server-conf\") pod \"rabbitmq-server-0\" (UID: \"b513e4ee-7c30-4942-b997-4d52414e1ec0\") " pod="openstack/rabbitmq-server-0" Nov 21 14:29:04 crc kubenswrapper[4897]: I1121 14:29:04.859021 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2gt5\" (UniqueName: \"kubernetes.io/projected/b513e4ee-7c30-4942-b997-4d52414e1ec0-kube-api-access-s2gt5\") pod \"rabbitmq-server-0\" (UID: \"b513e4ee-7c30-4942-b997-4d52414e1ec0\") " pod="openstack/rabbitmq-server-0" Nov 21 14:29:04 crc kubenswrapper[4897]: I1121 14:29:04.859054 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b513e4ee-7c30-4942-b997-4d52414e1ec0-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"b513e4ee-7c30-4942-b997-4d52414e1ec0\") " pod="openstack/rabbitmq-server-0" Nov 21 14:29:04 crc kubenswrapper[4897]: I1121 14:29:04.859080 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b513e4ee-7c30-4942-b997-4d52414e1ec0-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"b513e4ee-7c30-4942-b997-4d52414e1ec0\") " pod="openstack/rabbitmq-server-0" Nov 21 14:29:04 crc kubenswrapper[4897]: I1121 14:29:04.859113 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b513e4ee-7c30-4942-b997-4d52414e1ec0-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"b513e4ee-7c30-4942-b997-4d52414e1ec0\") " pod="openstack/rabbitmq-server-0" Nov 21 14:29:04 crc kubenswrapper[4897]: I1121 14:29:04.859163 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b513e4ee-7c30-4942-b997-4d52414e1ec0-pod-info\") pod \"rabbitmq-server-0\" (UID: \"b513e4ee-7c30-4942-b997-4d52414e1ec0\") " pod="openstack/rabbitmq-server-0" Nov 21 14:29:04 crc kubenswrapper[4897]: I1121 14:29:04.859193 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b513e4ee-7c30-4942-b997-4d52414e1ec0-config-data\") pod \"rabbitmq-server-0\" (UID: \"b513e4ee-7c30-4942-b997-4d52414e1ec0\") " pod="openstack/rabbitmq-server-0" Nov 21 14:29:04 crc kubenswrapper[4897]: I1121 14:29:04.859916 4897 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"b513e4ee-7c30-4942-b997-4d52414e1ec0\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/rabbitmq-server-0" Nov 21 14:29:04 crc kubenswrapper[4897]: I1121 14:29:04.860029 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b513e4ee-7c30-4942-b997-4d52414e1ec0-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"b513e4ee-7c30-4942-b997-4d52414e1ec0\") " pod="openstack/rabbitmq-server-0" Nov 21 14:29:04 crc kubenswrapper[4897]: I1121 14:29:04.860302 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b513e4ee-7c30-4942-b997-4d52414e1ec0-config-data\") pod \"rabbitmq-server-0\" (UID: \"b513e4ee-7c30-4942-b997-4d52414e1ec0\") " pod="openstack/rabbitmq-server-0" Nov 21 14:29:04 crc kubenswrapper[4897]: I1121 14:29:04.860526 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b513e4ee-7c30-4942-b997-4d52414e1ec0-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"b513e4ee-7c30-4942-b997-4d52414e1ec0\") " pod="openstack/rabbitmq-server-0" Nov 21 14:29:04 crc kubenswrapper[4897]: I1121 14:29:04.861108 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b513e4ee-7c30-4942-b997-4d52414e1ec0-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"b513e4ee-7c30-4942-b997-4d52414e1ec0\") " pod="openstack/rabbitmq-server-0" Nov 21 14:29:04 crc kubenswrapper[4897]: I1121 14:29:04.862280 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b513e4ee-7c30-4942-b997-4d52414e1ec0-server-conf\") pod \"rabbitmq-server-0\" (UID: \"b513e4ee-7c30-4942-b997-4d52414e1ec0\") " pod="openstack/rabbitmq-server-0" Nov 21 14:29:04 crc kubenswrapper[4897]: I1121 14:29:04.875142 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b513e4ee-7c30-4942-b997-4d52414e1ec0-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"b513e4ee-7c30-4942-b997-4d52414e1ec0\") " pod="openstack/rabbitmq-server-0" Nov 21 14:29:04 crc kubenswrapper[4897]: I1121 14:29:04.876149 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b513e4ee-7c30-4942-b997-4d52414e1ec0-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"b513e4ee-7c30-4942-b997-4d52414e1ec0\") " pod="openstack/rabbitmq-server-0" Nov 21 14:29:04 crc kubenswrapper[4897]: I1121 14:29:04.883559 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b513e4ee-7c30-4942-b997-4d52414e1ec0-pod-info\") pod \"rabbitmq-server-0\" (UID: \"b513e4ee-7c30-4942-b997-4d52414e1ec0\") " pod="openstack/rabbitmq-server-0" Nov 21 14:29:04 crc kubenswrapper[4897]: I1121 14:29:04.891985 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b513e4ee-7c30-4942-b997-4d52414e1ec0-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"b513e4ee-7c30-4942-b997-4d52414e1ec0\") " pod="openstack/rabbitmq-server-0" Nov 21 14:29:04 crc kubenswrapper[4897]: I1121 14:29:04.896732 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-8f9xc"] Nov 21 14:29:04 crc kubenswrapper[4897]: I1121 14:29:04.899913 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"b513e4ee-7c30-4942-b997-4d52414e1ec0\") " pod="openstack/rabbitmq-server-0" Nov 21 14:29:04 crc kubenswrapper[4897]: I1121 14:29:04.943170 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2gt5\" (UniqueName: \"kubernetes.io/projected/b513e4ee-7c30-4942-b997-4d52414e1ec0-kube-api-access-s2gt5\") pod \"rabbitmq-server-0\" (UID: \"b513e4ee-7c30-4942-b997-4d52414e1ec0\") " pod="openstack/rabbitmq-server-0" Nov 21 14:29:04 crc kubenswrapper[4897]: I1121 14:29:04.953473 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 21 14:29:05 crc kubenswrapper[4897]: I1121 14:29:05.053366 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 21 14:29:05 crc kubenswrapper[4897]: I1121 14:29:05.056056 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 21 14:29:05 crc kubenswrapper[4897]: I1121 14:29:05.058650 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Nov 21 14:29:05 crc kubenswrapper[4897]: I1121 14:29:05.058893 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Nov 21 14:29:05 crc kubenswrapper[4897]: I1121 14:29:05.058948 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-w65rf" Nov 21 14:29:05 crc kubenswrapper[4897]: I1121 14:29:05.058955 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Nov 21 14:29:05 crc kubenswrapper[4897]: I1121 14:29:05.059014 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Nov 21 14:29:05 crc kubenswrapper[4897]: I1121 14:29:05.059160 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Nov 21 14:29:05 crc kubenswrapper[4897]: I1121 14:29:05.063678 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Nov 21 14:29:05 crc kubenswrapper[4897]: I1121 14:29:05.076922 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 21 14:29:05 crc kubenswrapper[4897]: I1121 14:29:05.165222 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/bd692188-6da6-4387-a46d-003a2da0d0c8-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"bd692188-6da6-4387-a46d-003a2da0d0c8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 21 14:29:05 crc kubenswrapper[4897]: I1121 14:29:05.165278 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/bd692188-6da6-4387-a46d-003a2da0d0c8-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"bd692188-6da6-4387-a46d-003a2da0d0c8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 21 14:29:05 crc kubenswrapper[4897]: I1121 14:29:05.165294 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/bd692188-6da6-4387-a46d-003a2da0d0c8-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"bd692188-6da6-4387-a46d-003a2da0d0c8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 21 14:29:05 crc kubenswrapper[4897]: I1121 14:29:05.165316 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/bd692188-6da6-4387-a46d-003a2da0d0c8-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"bd692188-6da6-4387-a46d-003a2da0d0c8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 21 14:29:05 crc kubenswrapper[4897]: I1121 14:29:05.165355 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kbnvn\" (UniqueName: \"kubernetes.io/projected/bd692188-6da6-4387-a46d-003a2da0d0c8-kube-api-access-kbnvn\") pod \"rabbitmq-cell1-server-0\" (UID: \"bd692188-6da6-4387-a46d-003a2da0d0c8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 21 14:29:05 crc kubenswrapper[4897]: I1121 14:29:05.165370 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bd692188-6da6-4387-a46d-003a2da0d0c8-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"bd692188-6da6-4387-a46d-003a2da0d0c8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 21 14:29:05 crc kubenswrapper[4897]: I1121 14:29:05.165407 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"bd692188-6da6-4387-a46d-003a2da0d0c8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 21 14:29:05 crc kubenswrapper[4897]: I1121 14:29:05.165448 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/bd692188-6da6-4387-a46d-003a2da0d0c8-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"bd692188-6da6-4387-a46d-003a2da0d0c8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 21 14:29:05 crc kubenswrapper[4897]: I1121 14:29:05.165473 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/bd692188-6da6-4387-a46d-003a2da0d0c8-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"bd692188-6da6-4387-a46d-003a2da0d0c8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 21 14:29:05 crc kubenswrapper[4897]: I1121 14:29:05.165491 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/bd692188-6da6-4387-a46d-003a2da0d0c8-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"bd692188-6da6-4387-a46d-003a2da0d0c8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 21 14:29:05 crc kubenswrapper[4897]: I1121 14:29:05.165549 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/bd692188-6da6-4387-a46d-003a2da0d0c8-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"bd692188-6da6-4387-a46d-003a2da0d0c8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 21 14:29:05 crc kubenswrapper[4897]: I1121 14:29:05.268007 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/bd692188-6da6-4387-a46d-003a2da0d0c8-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"bd692188-6da6-4387-a46d-003a2da0d0c8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 21 14:29:05 crc kubenswrapper[4897]: I1121 14:29:05.268063 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/bd692188-6da6-4387-a46d-003a2da0d0c8-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"bd692188-6da6-4387-a46d-003a2da0d0c8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 21 14:29:05 crc kubenswrapper[4897]: I1121 14:29:05.268085 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/bd692188-6da6-4387-a46d-003a2da0d0c8-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"bd692188-6da6-4387-a46d-003a2da0d0c8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 21 14:29:05 crc kubenswrapper[4897]: I1121 14:29:05.268112 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/bd692188-6da6-4387-a46d-003a2da0d0c8-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"bd692188-6da6-4387-a46d-003a2da0d0c8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 21 14:29:05 crc kubenswrapper[4897]: I1121 14:29:05.268162 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kbnvn\" (UniqueName: \"kubernetes.io/projected/bd692188-6da6-4387-a46d-003a2da0d0c8-kube-api-access-kbnvn\") pod \"rabbitmq-cell1-server-0\" (UID: \"bd692188-6da6-4387-a46d-003a2da0d0c8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 21 14:29:05 crc kubenswrapper[4897]: I1121 14:29:05.268185 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bd692188-6da6-4387-a46d-003a2da0d0c8-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"bd692188-6da6-4387-a46d-003a2da0d0c8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 21 14:29:05 crc kubenswrapper[4897]: I1121 14:29:05.268221 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"bd692188-6da6-4387-a46d-003a2da0d0c8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 21 14:29:05 crc kubenswrapper[4897]: I1121 14:29:05.268278 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/bd692188-6da6-4387-a46d-003a2da0d0c8-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"bd692188-6da6-4387-a46d-003a2da0d0c8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 21 14:29:05 crc kubenswrapper[4897]: I1121 14:29:05.268314 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/bd692188-6da6-4387-a46d-003a2da0d0c8-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"bd692188-6da6-4387-a46d-003a2da0d0c8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 21 14:29:05 crc kubenswrapper[4897]: I1121 14:29:05.268343 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/bd692188-6da6-4387-a46d-003a2da0d0c8-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"bd692188-6da6-4387-a46d-003a2da0d0c8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 21 14:29:05 crc kubenswrapper[4897]: I1121 14:29:05.268391 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/bd692188-6da6-4387-a46d-003a2da0d0c8-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"bd692188-6da6-4387-a46d-003a2da0d0c8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 21 14:29:05 crc kubenswrapper[4897]: I1121 14:29:05.268841 4897 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"bd692188-6da6-4387-a46d-003a2da0d0c8\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/rabbitmq-cell1-server-0" Nov 21 14:29:05 crc kubenswrapper[4897]: I1121 14:29:05.273124 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/bd692188-6da6-4387-a46d-003a2da0d0c8-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"bd692188-6da6-4387-a46d-003a2da0d0c8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 21 14:29:05 crc kubenswrapper[4897]: I1121 14:29:05.273913 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/bd692188-6da6-4387-a46d-003a2da0d0c8-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"bd692188-6da6-4387-a46d-003a2da0d0c8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 21 14:29:05 crc kubenswrapper[4897]: I1121 14:29:05.274674 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bd692188-6da6-4387-a46d-003a2da0d0c8-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"bd692188-6da6-4387-a46d-003a2da0d0c8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 21 14:29:05 crc kubenswrapper[4897]: I1121 14:29:05.276366 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/bd692188-6da6-4387-a46d-003a2da0d0c8-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"bd692188-6da6-4387-a46d-003a2da0d0c8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 21 14:29:05 crc kubenswrapper[4897]: I1121 14:29:05.276662 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/bd692188-6da6-4387-a46d-003a2da0d0c8-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"bd692188-6da6-4387-a46d-003a2da0d0c8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 21 14:29:05 crc kubenswrapper[4897]: I1121 14:29:05.289445 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/bd692188-6da6-4387-a46d-003a2da0d0c8-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"bd692188-6da6-4387-a46d-003a2da0d0c8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 21 14:29:05 crc kubenswrapper[4897]: I1121 14:29:05.291901 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/bd692188-6da6-4387-a46d-003a2da0d0c8-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"bd692188-6da6-4387-a46d-003a2da0d0c8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 21 14:29:05 crc kubenswrapper[4897]: I1121 14:29:05.301420 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/bd692188-6da6-4387-a46d-003a2da0d0c8-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"bd692188-6da6-4387-a46d-003a2da0d0c8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 21 14:29:05 crc kubenswrapper[4897]: I1121 14:29:05.301446 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kbnvn\" (UniqueName: \"kubernetes.io/projected/bd692188-6da6-4387-a46d-003a2da0d0c8-kube-api-access-kbnvn\") pod \"rabbitmq-cell1-server-0\" (UID: \"bd692188-6da6-4387-a46d-003a2da0d0c8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 21 14:29:05 crc kubenswrapper[4897]: I1121 14:29:05.309280 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/bd692188-6da6-4387-a46d-003a2da0d0c8-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"bd692188-6da6-4387-a46d-003a2da0d0c8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 21 14:29:05 crc kubenswrapper[4897]: I1121 14:29:05.344557 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"bd692188-6da6-4387-a46d-003a2da0d0c8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 21 14:29:05 crc kubenswrapper[4897]: I1121 14:29:05.389546 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 21 14:29:05 crc kubenswrapper[4897]: I1121 14:29:05.570682 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 21 14:29:05 crc kubenswrapper[4897]: W1121 14:29:05.578016 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb513e4ee_7c30_4942_b997_4d52414e1ec0.slice/crio-d57051f4a2d9e61d4d7988ce611e9c0574d0451a505583e0be5558754171da2b WatchSource:0}: Error finding container d57051f4a2d9e61d4d7988ce611e9c0574d0451a505583e0be5558754171da2b: Status 404 returned error can't find the container with id d57051f4a2d9e61d4d7988ce611e9c0574d0451a505583e0be5558754171da2b Nov 21 14:29:05 crc kubenswrapper[4897]: I1121 14:29:05.596855 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-8f9xc" event={"ID":"ecceef5c-b1e5-4c3c-97b3-d805494cfba3","Type":"ContainerStarted","Data":"33669f0e7adb0fc32c9cd659ca9c5f306af98c8a6ed1b09cd5baad14c63df0fd"} Nov 21 14:29:05 crc kubenswrapper[4897]: I1121 14:29:05.606586 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-khkvq" event={"ID":"6bf04b77-0ef9-42f0-8a02-8c4c7568e390","Type":"ContainerStarted","Data":"a72fb2283c9a1f64135fea02b1846c00268b9c8f14cf5808328ead7e0788d548"} Nov 21 14:29:05 crc kubenswrapper[4897]: I1121 14:29:05.989621 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 21 14:29:06 crc kubenswrapper[4897]: W1121 14:29:06.002079 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbd692188_6da6_4387_a46d_003a2da0d0c8.slice/crio-a8d57c480eb257dc71e9bdee5d9c69bc7b91ca6a3572a2555ff0fd0a19f8543e WatchSource:0}: Error finding container a8d57c480eb257dc71e9bdee5d9c69bc7b91ca6a3572a2555ff0fd0a19f8543e: Status 404 returned error can't find the container with id a8d57c480eb257dc71e9bdee5d9c69bc7b91ca6a3572a2555ff0fd0a19f8543e Nov 21 14:29:06 crc kubenswrapper[4897]: I1121 14:29:06.268428 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Nov 21 14:29:06 crc kubenswrapper[4897]: I1121 14:29:06.270856 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Nov 21 14:29:06 crc kubenswrapper[4897]: I1121 14:29:06.274479 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-7vj4k" Nov 21 14:29:06 crc kubenswrapper[4897]: I1121 14:29:06.275338 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Nov 21 14:29:06 crc kubenswrapper[4897]: I1121 14:29:06.276324 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Nov 21 14:29:06 crc kubenswrapper[4897]: I1121 14:29:06.278812 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Nov 21 14:29:06 crc kubenswrapper[4897]: I1121 14:29:06.290068 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Nov 21 14:29:06 crc kubenswrapper[4897]: I1121 14:29:06.294031 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Nov 21 14:29:06 crc kubenswrapper[4897]: I1121 14:29:06.309210 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tfr5n\" (UniqueName: \"kubernetes.io/projected/fce5a726-bd06-473b-a9d0-f812ffe83693-kube-api-access-tfr5n\") pod \"openstack-galera-0\" (UID: \"fce5a726-bd06-473b-a9d0-f812ffe83693\") " pod="openstack/openstack-galera-0" Nov 21 14:29:06 crc kubenswrapper[4897]: I1121 14:29:06.309273 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fce5a726-bd06-473b-a9d0-f812ffe83693-operator-scripts\") pod \"openstack-galera-0\" (UID: \"fce5a726-bd06-473b-a9d0-f812ffe83693\") " pod="openstack/openstack-galera-0" Nov 21 14:29:06 crc kubenswrapper[4897]: I1121 14:29:06.309302 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fce5a726-bd06-473b-a9d0-f812ffe83693-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"fce5a726-bd06-473b-a9d0-f812ffe83693\") " pod="openstack/openstack-galera-0" Nov 21 14:29:06 crc kubenswrapper[4897]: I1121 14:29:06.309357 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/fce5a726-bd06-473b-a9d0-f812ffe83693-kolla-config\") pod \"openstack-galera-0\" (UID: \"fce5a726-bd06-473b-a9d0-f812ffe83693\") " pod="openstack/openstack-galera-0" Nov 21 14:29:06 crc kubenswrapper[4897]: I1121 14:29:06.309393 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/fce5a726-bd06-473b-a9d0-f812ffe83693-config-data-generated\") pod \"openstack-galera-0\" (UID: \"fce5a726-bd06-473b-a9d0-f812ffe83693\") " pod="openstack/openstack-galera-0" Nov 21 14:29:06 crc kubenswrapper[4897]: I1121 14:29:06.309412 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-galera-0\" (UID: \"fce5a726-bd06-473b-a9d0-f812ffe83693\") " pod="openstack/openstack-galera-0" Nov 21 14:29:06 crc kubenswrapper[4897]: I1121 14:29:06.309466 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/fce5a726-bd06-473b-a9d0-f812ffe83693-config-data-default\") pod \"openstack-galera-0\" (UID: \"fce5a726-bd06-473b-a9d0-f812ffe83693\") " pod="openstack/openstack-galera-0" Nov 21 14:29:06 crc kubenswrapper[4897]: I1121 14:29:06.309584 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/fce5a726-bd06-473b-a9d0-f812ffe83693-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"fce5a726-bd06-473b-a9d0-f812ffe83693\") " pod="openstack/openstack-galera-0" Nov 21 14:29:06 crc kubenswrapper[4897]: I1121 14:29:06.411962 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tfr5n\" (UniqueName: \"kubernetes.io/projected/fce5a726-bd06-473b-a9d0-f812ffe83693-kube-api-access-tfr5n\") pod \"openstack-galera-0\" (UID: \"fce5a726-bd06-473b-a9d0-f812ffe83693\") " pod="openstack/openstack-galera-0" Nov 21 14:29:06 crc kubenswrapper[4897]: I1121 14:29:06.412016 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fce5a726-bd06-473b-a9d0-f812ffe83693-operator-scripts\") pod \"openstack-galera-0\" (UID: \"fce5a726-bd06-473b-a9d0-f812ffe83693\") " pod="openstack/openstack-galera-0" Nov 21 14:29:06 crc kubenswrapper[4897]: I1121 14:29:06.412080 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fce5a726-bd06-473b-a9d0-f812ffe83693-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"fce5a726-bd06-473b-a9d0-f812ffe83693\") " pod="openstack/openstack-galera-0" Nov 21 14:29:06 crc kubenswrapper[4897]: I1121 14:29:06.412110 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/fce5a726-bd06-473b-a9d0-f812ffe83693-kolla-config\") pod \"openstack-galera-0\" (UID: \"fce5a726-bd06-473b-a9d0-f812ffe83693\") " pod="openstack/openstack-galera-0" Nov 21 14:29:06 crc kubenswrapper[4897]: I1121 14:29:06.412135 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/fce5a726-bd06-473b-a9d0-f812ffe83693-config-data-generated\") pod \"openstack-galera-0\" (UID: \"fce5a726-bd06-473b-a9d0-f812ffe83693\") " pod="openstack/openstack-galera-0" Nov 21 14:29:06 crc kubenswrapper[4897]: I1121 14:29:06.412154 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-galera-0\" (UID: \"fce5a726-bd06-473b-a9d0-f812ffe83693\") " pod="openstack/openstack-galera-0" Nov 21 14:29:06 crc kubenswrapper[4897]: I1121 14:29:06.412193 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/fce5a726-bd06-473b-a9d0-f812ffe83693-config-data-default\") pod \"openstack-galera-0\" (UID: \"fce5a726-bd06-473b-a9d0-f812ffe83693\") " pod="openstack/openstack-galera-0" Nov 21 14:29:06 crc kubenswrapper[4897]: I1121 14:29:06.412241 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/fce5a726-bd06-473b-a9d0-f812ffe83693-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"fce5a726-bd06-473b-a9d0-f812ffe83693\") " pod="openstack/openstack-galera-0" Nov 21 14:29:06 crc kubenswrapper[4897]: I1121 14:29:06.416331 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/fce5a726-bd06-473b-a9d0-f812ffe83693-kolla-config\") pod \"openstack-galera-0\" (UID: \"fce5a726-bd06-473b-a9d0-f812ffe83693\") " pod="openstack/openstack-galera-0" Nov 21 14:29:06 crc kubenswrapper[4897]: I1121 14:29:06.417301 4897 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-galera-0\" (UID: \"fce5a726-bd06-473b-a9d0-f812ffe83693\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/openstack-galera-0" Nov 21 14:29:06 crc kubenswrapper[4897]: I1121 14:29:06.417849 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/fce5a726-bd06-473b-a9d0-f812ffe83693-config-data-default\") pod \"openstack-galera-0\" (UID: \"fce5a726-bd06-473b-a9d0-f812ffe83693\") " pod="openstack/openstack-galera-0" Nov 21 14:29:06 crc kubenswrapper[4897]: I1121 14:29:06.418399 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/fce5a726-bd06-473b-a9d0-f812ffe83693-config-data-generated\") pod \"openstack-galera-0\" (UID: \"fce5a726-bd06-473b-a9d0-f812ffe83693\") " pod="openstack/openstack-galera-0" Nov 21 14:29:06 crc kubenswrapper[4897]: I1121 14:29:06.419963 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fce5a726-bd06-473b-a9d0-f812ffe83693-operator-scripts\") pod \"openstack-galera-0\" (UID: \"fce5a726-bd06-473b-a9d0-f812ffe83693\") " pod="openstack/openstack-galera-0" Nov 21 14:29:06 crc kubenswrapper[4897]: I1121 14:29:06.437461 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fce5a726-bd06-473b-a9d0-f812ffe83693-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"fce5a726-bd06-473b-a9d0-f812ffe83693\") " pod="openstack/openstack-galera-0" Nov 21 14:29:06 crc kubenswrapper[4897]: I1121 14:29:06.457180 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/fce5a726-bd06-473b-a9d0-f812ffe83693-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"fce5a726-bd06-473b-a9d0-f812ffe83693\") " pod="openstack/openstack-galera-0" Nov 21 14:29:06 crc kubenswrapper[4897]: I1121 14:29:06.463150 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tfr5n\" (UniqueName: \"kubernetes.io/projected/fce5a726-bd06-473b-a9d0-f812ffe83693-kube-api-access-tfr5n\") pod \"openstack-galera-0\" (UID: \"fce5a726-bd06-473b-a9d0-f812ffe83693\") " pod="openstack/openstack-galera-0" Nov 21 14:29:06 crc kubenswrapper[4897]: I1121 14:29:06.463718 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-galera-0\" (UID: \"fce5a726-bd06-473b-a9d0-f812ffe83693\") " pod="openstack/openstack-galera-0" Nov 21 14:29:06 crc kubenswrapper[4897]: I1121 14:29:06.595020 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Nov 21 14:29:06 crc kubenswrapper[4897]: I1121 14:29:06.730482 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"bd692188-6da6-4387-a46d-003a2da0d0c8","Type":"ContainerStarted","Data":"a8d57c480eb257dc71e9bdee5d9c69bc7b91ca6a3572a2555ff0fd0a19f8543e"} Nov 21 14:29:06 crc kubenswrapper[4897]: I1121 14:29:06.734630 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"b513e4ee-7c30-4942-b997-4d52414e1ec0","Type":"ContainerStarted","Data":"d57051f4a2d9e61d4d7988ce611e9c0574d0451a505583e0be5558754171da2b"} Nov 21 14:29:07 crc kubenswrapper[4897]: I1121 14:29:07.441319 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Nov 21 14:29:07 crc kubenswrapper[4897]: W1121 14:29:07.460258 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfce5a726_bd06_473b_a9d0_f812ffe83693.slice/crio-e37fa8e7437e180798d7dfeb454483e16f3e1cfa8f030be302d4a6e6b5e1dc02 WatchSource:0}: Error finding container e37fa8e7437e180798d7dfeb454483e16f3e1cfa8f030be302d4a6e6b5e1dc02: Status 404 returned error can't find the container with id e37fa8e7437e180798d7dfeb454483e16f3e1cfa8f030be302d4a6e6b5e1dc02 Nov 21 14:29:07 crc kubenswrapper[4897]: I1121 14:29:07.539298 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 21 14:29:07 crc kubenswrapper[4897]: I1121 14:29:07.546124 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Nov 21 14:29:07 crc kubenswrapper[4897]: I1121 14:29:07.552012 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-4pl4d" Nov 21 14:29:07 crc kubenswrapper[4897]: I1121 14:29:07.552369 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Nov 21 14:29:07 crc kubenswrapper[4897]: I1121 14:29:07.552788 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Nov 21 14:29:07 crc kubenswrapper[4897]: I1121 14:29:07.552935 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Nov 21 14:29:07 crc kubenswrapper[4897]: I1121 14:29:07.557473 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 21 14:29:07 crc kubenswrapper[4897]: I1121 14:29:07.652419 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/d3691270-398c-46e4-9db9-de8662274fd3-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"d3691270-398c-46e4-9db9-de8662274fd3\") " pod="openstack/openstack-cell1-galera-0" Nov 21 14:29:07 crc kubenswrapper[4897]: I1121 14:29:07.653172 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/d3691270-398c-46e4-9db9-de8662274fd3-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"d3691270-398c-46e4-9db9-de8662274fd3\") " pod="openstack/openstack-cell1-galera-0" Nov 21 14:29:07 crc kubenswrapper[4897]: I1121 14:29:07.653380 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3691270-398c-46e4-9db9-de8662274fd3-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"d3691270-398c-46e4-9db9-de8662274fd3\") " pod="openstack/openstack-cell1-galera-0" Nov 21 14:29:07 crc kubenswrapper[4897]: I1121 14:29:07.653433 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d3691270-398c-46e4-9db9-de8662274fd3-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"d3691270-398c-46e4-9db9-de8662274fd3\") " pod="openstack/openstack-cell1-galera-0" Nov 21 14:29:07 crc kubenswrapper[4897]: I1121 14:29:07.653545 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/d3691270-398c-46e4-9db9-de8662274fd3-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"d3691270-398c-46e4-9db9-de8662274fd3\") " pod="openstack/openstack-cell1-galera-0" Nov 21 14:29:07 crc kubenswrapper[4897]: I1121 14:29:07.653595 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-cell1-galera-0\" (UID: \"d3691270-398c-46e4-9db9-de8662274fd3\") " pod="openstack/openstack-cell1-galera-0" Nov 21 14:29:07 crc kubenswrapper[4897]: I1121 14:29:07.653664 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d3691270-398c-46e4-9db9-de8662274fd3-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"d3691270-398c-46e4-9db9-de8662274fd3\") " pod="openstack/openstack-cell1-galera-0" Nov 21 14:29:07 crc kubenswrapper[4897]: I1121 14:29:07.653702 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lpwq2\" (UniqueName: \"kubernetes.io/projected/d3691270-398c-46e4-9db9-de8662274fd3-kube-api-access-lpwq2\") pod \"openstack-cell1-galera-0\" (UID: \"d3691270-398c-46e4-9db9-de8662274fd3\") " pod="openstack/openstack-cell1-galera-0" Nov 21 14:29:07 crc kubenswrapper[4897]: I1121 14:29:07.754265 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Nov 21 14:29:07 crc kubenswrapper[4897]: I1121 14:29:07.754905 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d3691270-398c-46e4-9db9-de8662274fd3-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"d3691270-398c-46e4-9db9-de8662274fd3\") " pod="openstack/openstack-cell1-galera-0" Nov 21 14:29:07 crc kubenswrapper[4897]: I1121 14:29:07.754965 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lpwq2\" (UniqueName: \"kubernetes.io/projected/d3691270-398c-46e4-9db9-de8662274fd3-kube-api-access-lpwq2\") pod \"openstack-cell1-galera-0\" (UID: \"d3691270-398c-46e4-9db9-de8662274fd3\") " pod="openstack/openstack-cell1-galera-0" Nov 21 14:29:07 crc kubenswrapper[4897]: I1121 14:29:07.755040 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/d3691270-398c-46e4-9db9-de8662274fd3-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"d3691270-398c-46e4-9db9-de8662274fd3\") " pod="openstack/openstack-cell1-galera-0" Nov 21 14:29:07 crc kubenswrapper[4897]: I1121 14:29:07.755082 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/d3691270-398c-46e4-9db9-de8662274fd3-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"d3691270-398c-46e4-9db9-de8662274fd3\") " pod="openstack/openstack-cell1-galera-0" Nov 21 14:29:07 crc kubenswrapper[4897]: I1121 14:29:07.755119 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3691270-398c-46e4-9db9-de8662274fd3-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"d3691270-398c-46e4-9db9-de8662274fd3\") " pod="openstack/openstack-cell1-galera-0" Nov 21 14:29:07 crc kubenswrapper[4897]: I1121 14:29:07.755160 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d3691270-398c-46e4-9db9-de8662274fd3-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"d3691270-398c-46e4-9db9-de8662274fd3\") " pod="openstack/openstack-cell1-galera-0" Nov 21 14:29:07 crc kubenswrapper[4897]: I1121 14:29:07.755224 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/d3691270-398c-46e4-9db9-de8662274fd3-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"d3691270-398c-46e4-9db9-de8662274fd3\") " pod="openstack/openstack-cell1-galera-0" Nov 21 14:29:07 crc kubenswrapper[4897]: I1121 14:29:07.755266 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-cell1-galera-0\" (UID: \"d3691270-398c-46e4-9db9-de8662274fd3\") " pod="openstack/openstack-cell1-galera-0" Nov 21 14:29:07 crc kubenswrapper[4897]: I1121 14:29:07.755707 4897 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-cell1-galera-0\" (UID: \"d3691270-398c-46e4-9db9-de8662274fd3\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/openstack-cell1-galera-0" Nov 21 14:29:07 crc kubenswrapper[4897]: I1121 14:29:07.757068 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Nov 21 14:29:07 crc kubenswrapper[4897]: I1121 14:29:07.759922 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"fce5a726-bd06-473b-a9d0-f812ffe83693","Type":"ContainerStarted","Data":"e37fa8e7437e180798d7dfeb454483e16f3e1cfa8f030be302d4a6e6b5e1dc02"} Nov 21 14:29:07 crc kubenswrapper[4897]: I1121 14:29:07.760621 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d3691270-398c-46e4-9db9-de8662274fd3-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"d3691270-398c-46e4-9db9-de8662274fd3\") " pod="openstack/openstack-cell1-galera-0" Nov 21 14:29:07 crc kubenswrapper[4897]: I1121 14:29:07.760863 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/d3691270-398c-46e4-9db9-de8662274fd3-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"d3691270-398c-46e4-9db9-de8662274fd3\") " pod="openstack/openstack-cell1-galera-0" Nov 21 14:29:07 crc kubenswrapper[4897]: I1121 14:29:07.760863 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/d3691270-398c-46e4-9db9-de8662274fd3-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"d3691270-398c-46e4-9db9-de8662274fd3\") " pod="openstack/openstack-cell1-galera-0" Nov 21 14:29:07 crc kubenswrapper[4897]: I1121 14:29:07.761771 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/d3691270-398c-46e4-9db9-de8662274fd3-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"d3691270-398c-46e4-9db9-de8662274fd3\") " pod="openstack/openstack-cell1-galera-0" Nov 21 14:29:07 crc kubenswrapper[4897]: I1121 14:29:07.762355 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d3691270-398c-46e4-9db9-de8662274fd3-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"d3691270-398c-46e4-9db9-de8662274fd3\") " pod="openstack/openstack-cell1-galera-0" Nov 21 14:29:07 crc kubenswrapper[4897]: I1121 14:29:07.766062 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3691270-398c-46e4-9db9-de8662274fd3-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"d3691270-398c-46e4-9db9-de8662274fd3\") " pod="openstack/openstack-cell1-galera-0" Nov 21 14:29:07 crc kubenswrapper[4897]: I1121 14:29:07.766887 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Nov 21 14:29:07 crc kubenswrapper[4897]: I1121 14:29:07.767020 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-jbpdl" Nov 21 14:29:07 crc kubenswrapper[4897]: I1121 14:29:07.767134 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Nov 21 14:29:07 crc kubenswrapper[4897]: I1121 14:29:07.778267 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Nov 21 14:29:07 crc kubenswrapper[4897]: I1121 14:29:07.790706 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lpwq2\" (UniqueName: \"kubernetes.io/projected/d3691270-398c-46e4-9db9-de8662274fd3-kube-api-access-lpwq2\") pod \"openstack-cell1-galera-0\" (UID: \"d3691270-398c-46e4-9db9-de8662274fd3\") " pod="openstack/openstack-cell1-galera-0" Nov 21 14:29:07 crc kubenswrapper[4897]: I1121 14:29:07.855205 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-cell1-galera-0\" (UID: \"d3691270-398c-46e4-9db9-de8662274fd3\") " pod="openstack/openstack-cell1-galera-0" Nov 21 14:29:07 crc kubenswrapper[4897]: I1121 14:29:07.858452 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jgrn7\" (UniqueName: \"kubernetes.io/projected/fc30a810-d114-48ee-97c2-414f1ed23d22-kube-api-access-jgrn7\") pod \"memcached-0\" (UID: \"fc30a810-d114-48ee-97c2-414f1ed23d22\") " pod="openstack/memcached-0" Nov 21 14:29:07 crc kubenswrapper[4897]: I1121 14:29:07.858497 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/fc30a810-d114-48ee-97c2-414f1ed23d22-memcached-tls-certs\") pod \"memcached-0\" (UID: \"fc30a810-d114-48ee-97c2-414f1ed23d22\") " pod="openstack/memcached-0" Nov 21 14:29:07 crc kubenswrapper[4897]: I1121 14:29:07.858533 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/fc30a810-d114-48ee-97c2-414f1ed23d22-kolla-config\") pod \"memcached-0\" (UID: \"fc30a810-d114-48ee-97c2-414f1ed23d22\") " pod="openstack/memcached-0" Nov 21 14:29:07 crc kubenswrapper[4897]: I1121 14:29:07.858719 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc30a810-d114-48ee-97c2-414f1ed23d22-combined-ca-bundle\") pod \"memcached-0\" (UID: \"fc30a810-d114-48ee-97c2-414f1ed23d22\") " pod="openstack/memcached-0" Nov 21 14:29:07 crc kubenswrapper[4897]: I1121 14:29:07.858794 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fc30a810-d114-48ee-97c2-414f1ed23d22-config-data\") pod \"memcached-0\" (UID: \"fc30a810-d114-48ee-97c2-414f1ed23d22\") " pod="openstack/memcached-0" Nov 21 14:29:07 crc kubenswrapper[4897]: I1121 14:29:07.917289 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Nov 21 14:29:07 crc kubenswrapper[4897]: I1121 14:29:07.960165 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc30a810-d114-48ee-97c2-414f1ed23d22-combined-ca-bundle\") pod \"memcached-0\" (UID: \"fc30a810-d114-48ee-97c2-414f1ed23d22\") " pod="openstack/memcached-0" Nov 21 14:29:07 crc kubenswrapper[4897]: I1121 14:29:07.960276 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fc30a810-d114-48ee-97c2-414f1ed23d22-config-data\") pod \"memcached-0\" (UID: \"fc30a810-d114-48ee-97c2-414f1ed23d22\") " pod="openstack/memcached-0" Nov 21 14:29:07 crc kubenswrapper[4897]: I1121 14:29:07.960337 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jgrn7\" (UniqueName: \"kubernetes.io/projected/fc30a810-d114-48ee-97c2-414f1ed23d22-kube-api-access-jgrn7\") pod \"memcached-0\" (UID: \"fc30a810-d114-48ee-97c2-414f1ed23d22\") " pod="openstack/memcached-0" Nov 21 14:29:07 crc kubenswrapper[4897]: I1121 14:29:07.960358 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/fc30a810-d114-48ee-97c2-414f1ed23d22-memcached-tls-certs\") pod \"memcached-0\" (UID: \"fc30a810-d114-48ee-97c2-414f1ed23d22\") " pod="openstack/memcached-0" Nov 21 14:29:07 crc kubenswrapper[4897]: I1121 14:29:07.960373 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/fc30a810-d114-48ee-97c2-414f1ed23d22-kolla-config\") pod \"memcached-0\" (UID: \"fc30a810-d114-48ee-97c2-414f1ed23d22\") " pod="openstack/memcached-0" Nov 21 14:29:07 crc kubenswrapper[4897]: I1121 14:29:07.961350 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fc30a810-d114-48ee-97c2-414f1ed23d22-config-data\") pod \"memcached-0\" (UID: \"fc30a810-d114-48ee-97c2-414f1ed23d22\") " pod="openstack/memcached-0" Nov 21 14:29:07 crc kubenswrapper[4897]: I1121 14:29:07.961469 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/fc30a810-d114-48ee-97c2-414f1ed23d22-kolla-config\") pod \"memcached-0\" (UID: \"fc30a810-d114-48ee-97c2-414f1ed23d22\") " pod="openstack/memcached-0" Nov 21 14:29:07 crc kubenswrapper[4897]: I1121 14:29:07.965739 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/fc30a810-d114-48ee-97c2-414f1ed23d22-memcached-tls-certs\") pod \"memcached-0\" (UID: \"fc30a810-d114-48ee-97c2-414f1ed23d22\") " pod="openstack/memcached-0" Nov 21 14:29:07 crc kubenswrapper[4897]: I1121 14:29:07.988390 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jgrn7\" (UniqueName: \"kubernetes.io/projected/fc30a810-d114-48ee-97c2-414f1ed23d22-kube-api-access-jgrn7\") pod \"memcached-0\" (UID: \"fc30a810-d114-48ee-97c2-414f1ed23d22\") " pod="openstack/memcached-0" Nov 21 14:29:07 crc kubenswrapper[4897]: I1121 14:29:07.988670 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc30a810-d114-48ee-97c2-414f1ed23d22-combined-ca-bundle\") pod \"memcached-0\" (UID: \"fc30a810-d114-48ee-97c2-414f1ed23d22\") " pod="openstack/memcached-0" Nov 21 14:29:08 crc kubenswrapper[4897]: I1121 14:29:08.228962 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Nov 21 14:29:08 crc kubenswrapper[4897]: I1121 14:29:08.629807 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 21 14:29:08 crc kubenswrapper[4897]: W1121 14:29:08.709984 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd3691270_398c_46e4_9db9_de8662274fd3.slice/crio-396073c942179fb760fe7ddf4a4e67e72e3924c21639440dac0ed0f81641b945 WatchSource:0}: Error finding container 396073c942179fb760fe7ddf4a4e67e72e3924c21639440dac0ed0f81641b945: Status 404 returned error can't find the container with id 396073c942179fb760fe7ddf4a4e67e72e3924c21639440dac0ed0f81641b945 Nov 21 14:29:08 crc kubenswrapper[4897]: I1121 14:29:08.774154 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"d3691270-398c-46e4-9db9-de8662274fd3","Type":"ContainerStarted","Data":"396073c942179fb760fe7ddf4a4e67e72e3924c21639440dac0ed0f81641b945"} Nov 21 14:29:08 crc kubenswrapper[4897]: I1121 14:29:08.818373 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Nov 21 14:29:10 crc kubenswrapper[4897]: I1121 14:29:10.024959 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Nov 21 14:29:10 crc kubenswrapper[4897]: I1121 14:29:10.031814 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 21 14:29:10 crc kubenswrapper[4897]: I1121 14:29:10.035655 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-m2cgp" Nov 21 14:29:10 crc kubenswrapper[4897]: I1121 14:29:10.043331 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 21 14:29:10 crc kubenswrapper[4897]: I1121 14:29:10.111418 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pnbx9\" (UniqueName: \"kubernetes.io/projected/95ad29e4-b4b6-4cec-b978-23f7ea25edc6-kube-api-access-pnbx9\") pod \"kube-state-metrics-0\" (UID: \"95ad29e4-b4b6-4cec-b978-23f7ea25edc6\") " pod="openstack/kube-state-metrics-0" Nov 21 14:29:10 crc kubenswrapper[4897]: I1121 14:29:10.215963 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pnbx9\" (UniqueName: \"kubernetes.io/projected/95ad29e4-b4b6-4cec-b978-23f7ea25edc6-kube-api-access-pnbx9\") pod \"kube-state-metrics-0\" (UID: \"95ad29e4-b4b6-4cec-b978-23f7ea25edc6\") " pod="openstack/kube-state-metrics-0" Nov 21 14:29:10 crc kubenswrapper[4897]: I1121 14:29:10.255886 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pnbx9\" (UniqueName: \"kubernetes.io/projected/95ad29e4-b4b6-4cec-b978-23f7ea25edc6-kube-api-access-pnbx9\") pod \"kube-state-metrics-0\" (UID: \"95ad29e4-b4b6-4cec-b978-23f7ea25edc6\") " pod="openstack/kube-state-metrics-0" Nov 21 14:29:10 crc kubenswrapper[4897]: I1121 14:29:10.363199 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 21 14:29:10 crc kubenswrapper[4897]: I1121 14:29:10.619875 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-ui-dashboards-7d5fb4cbfb-mt4vk"] Nov 21 14:29:10 crc kubenswrapper[4897]: I1121 14:29:10.621004 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-mt4vk" Nov 21 14:29:10 crc kubenswrapper[4897]: I1121 14:29:10.627760 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-ui-dashboards" Nov 21 14:29:10 crc kubenswrapper[4897]: I1121 14:29:10.634416 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-ui-dashboards-sa-dockercfg-qg84s" Nov 21 14:29:10 crc kubenswrapper[4897]: I1121 14:29:10.655564 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-ui-dashboards-7d5fb4cbfb-mt4vk"] Nov 21 14:29:10 crc kubenswrapper[4897]: I1121 14:29:10.734907 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6af809de-112c-4988-a06e-4838ef479ace-serving-cert\") pod \"observability-ui-dashboards-7d5fb4cbfb-mt4vk\" (UID: \"6af809de-112c-4988-a06e-4838ef479ace\") " pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-mt4vk" Nov 21 14:29:10 crc kubenswrapper[4897]: I1121 14:29:10.734956 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xcxrn\" (UniqueName: \"kubernetes.io/projected/6af809de-112c-4988-a06e-4838ef479ace-kube-api-access-xcxrn\") pod \"observability-ui-dashboards-7d5fb4cbfb-mt4vk\" (UID: \"6af809de-112c-4988-a06e-4838ef479ace\") " pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-mt4vk" Nov 21 14:29:10 crc kubenswrapper[4897]: I1121 14:29:10.841834 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6af809de-112c-4988-a06e-4838ef479ace-serving-cert\") pod \"observability-ui-dashboards-7d5fb4cbfb-mt4vk\" (UID: \"6af809de-112c-4988-a06e-4838ef479ace\") " pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-mt4vk" Nov 21 14:29:10 crc kubenswrapper[4897]: I1121 14:29:10.842088 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xcxrn\" (UniqueName: \"kubernetes.io/projected/6af809de-112c-4988-a06e-4838ef479ace-kube-api-access-xcxrn\") pod \"observability-ui-dashboards-7d5fb4cbfb-mt4vk\" (UID: \"6af809de-112c-4988-a06e-4838ef479ace\") " pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-mt4vk" Nov 21 14:29:10 crc kubenswrapper[4897]: I1121 14:29:10.864543 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6af809de-112c-4988-a06e-4838ef479ace-serving-cert\") pod \"observability-ui-dashboards-7d5fb4cbfb-mt4vk\" (UID: \"6af809de-112c-4988-a06e-4838ef479ace\") " pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-mt4vk" Nov 21 14:29:10 crc kubenswrapper[4897]: I1121 14:29:10.909535 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xcxrn\" (UniqueName: \"kubernetes.io/projected/6af809de-112c-4988-a06e-4838ef479ace-kube-api-access-xcxrn\") pod \"observability-ui-dashboards-7d5fb4cbfb-mt4vk\" (UID: \"6af809de-112c-4988-a06e-4838ef479ace\") " pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-mt4vk" Nov 21 14:29:10 crc kubenswrapper[4897]: I1121 14:29:10.946797 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-mt4vk" Nov 21 14:29:11 crc kubenswrapper[4897]: I1121 14:29:11.009974 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-84f7c79c46-td7jx"] Nov 21 14:29:11 crc kubenswrapper[4897]: I1121 14:29:11.011281 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-84f7c79c46-td7jx" Nov 21 14:29:11 crc kubenswrapper[4897]: I1121 14:29:11.073770 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-84f7c79c46-td7jx"] Nov 21 14:29:11 crc kubenswrapper[4897]: I1121 14:29:11.154823 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kzh9t\" (UniqueName: \"kubernetes.io/projected/dc6ff181-277d-4265-b2bb-bed01dc666b4-kube-api-access-kzh9t\") pod \"console-84f7c79c46-td7jx\" (UID: \"dc6ff181-277d-4265-b2bb-bed01dc666b4\") " pod="openshift-console/console-84f7c79c46-td7jx" Nov 21 14:29:11 crc kubenswrapper[4897]: I1121 14:29:11.155004 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/dc6ff181-277d-4265-b2bb-bed01dc666b4-console-serving-cert\") pod \"console-84f7c79c46-td7jx\" (UID: \"dc6ff181-277d-4265-b2bb-bed01dc666b4\") " pod="openshift-console/console-84f7c79c46-td7jx" Nov 21 14:29:11 crc kubenswrapper[4897]: I1121 14:29:11.155129 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/dc6ff181-277d-4265-b2bb-bed01dc666b4-oauth-serving-cert\") pod \"console-84f7c79c46-td7jx\" (UID: \"dc6ff181-277d-4265-b2bb-bed01dc666b4\") " pod="openshift-console/console-84f7c79c46-td7jx" Nov 21 14:29:11 crc kubenswrapper[4897]: I1121 14:29:11.155179 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/dc6ff181-277d-4265-b2bb-bed01dc666b4-service-ca\") pod \"console-84f7c79c46-td7jx\" (UID: \"dc6ff181-277d-4265-b2bb-bed01dc666b4\") " pod="openshift-console/console-84f7c79c46-td7jx" Nov 21 14:29:11 crc kubenswrapper[4897]: I1121 14:29:11.155237 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/dc6ff181-277d-4265-b2bb-bed01dc666b4-console-oauth-config\") pod \"console-84f7c79c46-td7jx\" (UID: \"dc6ff181-277d-4265-b2bb-bed01dc666b4\") " pod="openshift-console/console-84f7c79c46-td7jx" Nov 21 14:29:11 crc kubenswrapper[4897]: I1121 14:29:11.155328 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/dc6ff181-277d-4265-b2bb-bed01dc666b4-console-config\") pod \"console-84f7c79c46-td7jx\" (UID: \"dc6ff181-277d-4265-b2bb-bed01dc666b4\") " pod="openshift-console/console-84f7c79c46-td7jx" Nov 21 14:29:11 crc kubenswrapper[4897]: I1121 14:29:11.155388 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dc6ff181-277d-4265-b2bb-bed01dc666b4-trusted-ca-bundle\") pod \"console-84f7c79c46-td7jx\" (UID: \"dc6ff181-277d-4265-b2bb-bed01dc666b4\") " pod="openshift-console/console-84f7c79c46-td7jx" Nov 21 14:29:11 crc kubenswrapper[4897]: I1121 14:29:11.164969 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 21 14:29:11 crc kubenswrapper[4897]: I1121 14:29:11.167859 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Nov 21 14:29:11 crc kubenswrapper[4897]: I1121 14:29:11.169955 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Nov 21 14:29:11 crc kubenswrapper[4897]: I1121 14:29:11.170560 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Nov 21 14:29:11 crc kubenswrapper[4897]: I1121 14:29:11.171392 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Nov 21 14:29:11 crc kubenswrapper[4897]: I1121 14:29:11.171597 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Nov 21 14:29:11 crc kubenswrapper[4897]: I1121 14:29:11.172397 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-vshbl" Nov 21 14:29:11 crc kubenswrapper[4897]: I1121 14:29:11.181934 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Nov 21 14:29:11 crc kubenswrapper[4897]: I1121 14:29:11.190301 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 21 14:29:11 crc kubenswrapper[4897]: I1121 14:29:11.258095 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/dc6ff181-277d-4265-b2bb-bed01dc666b4-console-config\") pod \"console-84f7c79c46-td7jx\" (UID: \"dc6ff181-277d-4265-b2bb-bed01dc666b4\") " pod="openshift-console/console-84f7c79c46-td7jx" Nov 21 14:29:11 crc kubenswrapper[4897]: I1121 14:29:11.258147 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/91aa2270-92a7-4af4-9a92-141ea2b24ee1-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"91aa2270-92a7-4af4-9a92-141ea2b24ee1\") " pod="openstack/prometheus-metric-storage-0" Nov 21 14:29:11 crc kubenswrapper[4897]: I1121 14:29:11.258182 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dc6ff181-277d-4265-b2bb-bed01dc666b4-trusted-ca-bundle\") pod \"console-84f7c79c46-td7jx\" (UID: \"dc6ff181-277d-4265-b2bb-bed01dc666b4\") " pod="openshift-console/console-84f7c79c46-td7jx" Nov 21 14:29:11 crc kubenswrapper[4897]: I1121 14:29:11.258209 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/91aa2270-92a7-4af4-9a92-141ea2b24ee1-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"91aa2270-92a7-4af4-9a92-141ea2b24ee1\") " pod="openstack/prometheus-metric-storage-0" Nov 21 14:29:11 crc kubenswrapper[4897]: I1121 14:29:11.258256 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kzh9t\" (UniqueName: \"kubernetes.io/projected/dc6ff181-277d-4265-b2bb-bed01dc666b4-kube-api-access-kzh9t\") pod \"console-84f7c79c46-td7jx\" (UID: \"dc6ff181-277d-4265-b2bb-bed01dc666b4\") " pod="openshift-console/console-84f7c79c46-td7jx" Nov 21 14:29:11 crc kubenswrapper[4897]: I1121 14:29:11.258285 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-7060bc90-dec6-47cf-93b2-5e0ead1b71e7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7060bc90-dec6-47cf-93b2-5e0ead1b71e7\") pod \"prometheus-metric-storage-0\" (UID: \"91aa2270-92a7-4af4-9a92-141ea2b24ee1\") " pod="openstack/prometheus-metric-storage-0" Nov 21 14:29:11 crc kubenswrapper[4897]: I1121 14:29:11.258308 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/91aa2270-92a7-4af4-9a92-141ea2b24ee1-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"91aa2270-92a7-4af4-9a92-141ea2b24ee1\") " pod="openstack/prometheus-metric-storage-0" Nov 21 14:29:11 crc kubenswrapper[4897]: I1121 14:29:11.258341 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/91aa2270-92a7-4af4-9a92-141ea2b24ee1-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"91aa2270-92a7-4af4-9a92-141ea2b24ee1\") " pod="openstack/prometheus-metric-storage-0" Nov 21 14:29:11 crc kubenswrapper[4897]: I1121 14:29:11.258371 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/91aa2270-92a7-4af4-9a92-141ea2b24ee1-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"91aa2270-92a7-4af4-9a92-141ea2b24ee1\") " pod="openstack/prometheus-metric-storage-0" Nov 21 14:29:11 crc kubenswrapper[4897]: I1121 14:29:11.258388 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hxhfg\" (UniqueName: \"kubernetes.io/projected/91aa2270-92a7-4af4-9a92-141ea2b24ee1-kube-api-access-hxhfg\") pod \"prometheus-metric-storage-0\" (UID: \"91aa2270-92a7-4af4-9a92-141ea2b24ee1\") " pod="openstack/prometheus-metric-storage-0" Nov 21 14:29:11 crc kubenswrapper[4897]: I1121 14:29:11.258421 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/dc6ff181-277d-4265-b2bb-bed01dc666b4-console-serving-cert\") pod \"console-84f7c79c46-td7jx\" (UID: \"dc6ff181-277d-4265-b2bb-bed01dc666b4\") " pod="openshift-console/console-84f7c79c46-td7jx" Nov 21 14:29:11 crc kubenswrapper[4897]: I1121 14:29:11.258489 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/dc6ff181-277d-4265-b2bb-bed01dc666b4-oauth-serving-cert\") pod \"console-84f7c79c46-td7jx\" (UID: \"dc6ff181-277d-4265-b2bb-bed01dc666b4\") " pod="openshift-console/console-84f7c79c46-td7jx" Nov 21 14:29:11 crc kubenswrapper[4897]: I1121 14:29:11.258635 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/dc6ff181-277d-4265-b2bb-bed01dc666b4-service-ca\") pod \"console-84f7c79c46-td7jx\" (UID: \"dc6ff181-277d-4265-b2bb-bed01dc666b4\") " pod="openshift-console/console-84f7c79c46-td7jx" Nov 21 14:29:11 crc kubenswrapper[4897]: I1121 14:29:11.258728 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/dc6ff181-277d-4265-b2bb-bed01dc666b4-console-oauth-config\") pod \"console-84f7c79c46-td7jx\" (UID: \"dc6ff181-277d-4265-b2bb-bed01dc666b4\") " pod="openshift-console/console-84f7c79c46-td7jx" Nov 21 14:29:11 crc kubenswrapper[4897]: I1121 14:29:11.259891 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/91aa2270-92a7-4af4-9a92-141ea2b24ee1-config\") pod \"prometheus-metric-storage-0\" (UID: \"91aa2270-92a7-4af4-9a92-141ea2b24ee1\") " pod="openstack/prometheus-metric-storage-0" Nov 21 14:29:11 crc kubenswrapper[4897]: I1121 14:29:11.261664 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/dc6ff181-277d-4265-b2bb-bed01dc666b4-service-ca\") pod \"console-84f7c79c46-td7jx\" (UID: \"dc6ff181-277d-4265-b2bb-bed01dc666b4\") " pod="openshift-console/console-84f7c79c46-td7jx" Nov 21 14:29:11 crc kubenswrapper[4897]: I1121 14:29:11.262074 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/dc6ff181-277d-4265-b2bb-bed01dc666b4-oauth-serving-cert\") pod \"console-84f7c79c46-td7jx\" (UID: \"dc6ff181-277d-4265-b2bb-bed01dc666b4\") " pod="openshift-console/console-84f7c79c46-td7jx" Nov 21 14:29:11 crc kubenswrapper[4897]: I1121 14:29:11.262612 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/dc6ff181-277d-4265-b2bb-bed01dc666b4-console-config\") pod \"console-84f7c79c46-td7jx\" (UID: \"dc6ff181-277d-4265-b2bb-bed01dc666b4\") " pod="openshift-console/console-84f7c79c46-td7jx" Nov 21 14:29:11 crc kubenswrapper[4897]: I1121 14:29:11.263400 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dc6ff181-277d-4265-b2bb-bed01dc666b4-trusted-ca-bundle\") pod \"console-84f7c79c46-td7jx\" (UID: \"dc6ff181-277d-4265-b2bb-bed01dc666b4\") " pod="openshift-console/console-84f7c79c46-td7jx" Nov 21 14:29:11 crc kubenswrapper[4897]: I1121 14:29:11.275956 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/dc6ff181-277d-4265-b2bb-bed01dc666b4-console-serving-cert\") pod \"console-84f7c79c46-td7jx\" (UID: \"dc6ff181-277d-4265-b2bb-bed01dc666b4\") " pod="openshift-console/console-84f7c79c46-td7jx" Nov 21 14:29:11 crc kubenswrapper[4897]: I1121 14:29:11.294812 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/dc6ff181-277d-4265-b2bb-bed01dc666b4-console-oauth-config\") pod \"console-84f7c79c46-td7jx\" (UID: \"dc6ff181-277d-4265-b2bb-bed01dc666b4\") " pod="openshift-console/console-84f7c79c46-td7jx" Nov 21 14:29:11 crc kubenswrapper[4897]: I1121 14:29:11.305441 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kzh9t\" (UniqueName: \"kubernetes.io/projected/dc6ff181-277d-4265-b2bb-bed01dc666b4-kube-api-access-kzh9t\") pod \"console-84f7c79c46-td7jx\" (UID: \"dc6ff181-277d-4265-b2bb-bed01dc666b4\") " pod="openshift-console/console-84f7c79c46-td7jx" Nov 21 14:29:11 crc kubenswrapper[4897]: I1121 14:29:11.380978 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/91aa2270-92a7-4af4-9a92-141ea2b24ee1-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"91aa2270-92a7-4af4-9a92-141ea2b24ee1\") " pod="openstack/prometheus-metric-storage-0" Nov 21 14:29:11 crc kubenswrapper[4897]: I1121 14:29:11.381180 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-7060bc90-dec6-47cf-93b2-5e0ead1b71e7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7060bc90-dec6-47cf-93b2-5e0ead1b71e7\") pod \"prometheus-metric-storage-0\" (UID: \"91aa2270-92a7-4af4-9a92-141ea2b24ee1\") " pod="openstack/prometheus-metric-storage-0" Nov 21 14:29:11 crc kubenswrapper[4897]: I1121 14:29:11.381227 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/91aa2270-92a7-4af4-9a92-141ea2b24ee1-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"91aa2270-92a7-4af4-9a92-141ea2b24ee1\") " pod="openstack/prometheus-metric-storage-0" Nov 21 14:29:11 crc kubenswrapper[4897]: I1121 14:29:11.381303 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/91aa2270-92a7-4af4-9a92-141ea2b24ee1-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"91aa2270-92a7-4af4-9a92-141ea2b24ee1\") " pod="openstack/prometheus-metric-storage-0" Nov 21 14:29:11 crc kubenswrapper[4897]: I1121 14:29:11.381337 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/91aa2270-92a7-4af4-9a92-141ea2b24ee1-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"91aa2270-92a7-4af4-9a92-141ea2b24ee1\") " pod="openstack/prometheus-metric-storage-0" Nov 21 14:29:11 crc kubenswrapper[4897]: I1121 14:29:11.381362 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hxhfg\" (UniqueName: \"kubernetes.io/projected/91aa2270-92a7-4af4-9a92-141ea2b24ee1-kube-api-access-hxhfg\") pod \"prometheus-metric-storage-0\" (UID: \"91aa2270-92a7-4af4-9a92-141ea2b24ee1\") " pod="openstack/prometheus-metric-storage-0" Nov 21 14:29:11 crc kubenswrapper[4897]: I1121 14:29:11.381569 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/91aa2270-92a7-4af4-9a92-141ea2b24ee1-config\") pod \"prometheus-metric-storage-0\" (UID: \"91aa2270-92a7-4af4-9a92-141ea2b24ee1\") " pod="openstack/prometheus-metric-storage-0" Nov 21 14:29:11 crc kubenswrapper[4897]: I1121 14:29:11.381662 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/91aa2270-92a7-4af4-9a92-141ea2b24ee1-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"91aa2270-92a7-4af4-9a92-141ea2b24ee1\") " pod="openstack/prometheus-metric-storage-0" Nov 21 14:29:11 crc kubenswrapper[4897]: I1121 14:29:11.382758 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/91aa2270-92a7-4af4-9a92-141ea2b24ee1-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"91aa2270-92a7-4af4-9a92-141ea2b24ee1\") " pod="openstack/prometheus-metric-storage-0" Nov 21 14:29:11 crc kubenswrapper[4897]: I1121 14:29:11.391850 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-84f7c79c46-td7jx" Nov 21 14:29:11 crc kubenswrapper[4897]: I1121 14:29:11.392040 4897 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 21 14:29:11 crc kubenswrapper[4897]: I1121 14:29:11.392099 4897 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-7060bc90-dec6-47cf-93b2-5e0ead1b71e7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7060bc90-dec6-47cf-93b2-5e0ead1b71e7\") pod \"prometheus-metric-storage-0\" (UID: \"91aa2270-92a7-4af4-9a92-141ea2b24ee1\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/90579abefb2729890b6253e9dfb2d9cca2541535dfd7807baf8b98857bf5cb7e/globalmount\"" pod="openstack/prometheus-metric-storage-0" Nov 21 14:29:11 crc kubenswrapper[4897]: I1121 14:29:11.399979 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/91aa2270-92a7-4af4-9a92-141ea2b24ee1-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"91aa2270-92a7-4af4-9a92-141ea2b24ee1\") " pod="openstack/prometheus-metric-storage-0" Nov 21 14:29:11 crc kubenswrapper[4897]: I1121 14:29:11.401228 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/91aa2270-92a7-4af4-9a92-141ea2b24ee1-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"91aa2270-92a7-4af4-9a92-141ea2b24ee1\") " pod="openstack/prometheus-metric-storage-0" Nov 21 14:29:11 crc kubenswrapper[4897]: I1121 14:29:11.404366 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/91aa2270-92a7-4af4-9a92-141ea2b24ee1-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"91aa2270-92a7-4af4-9a92-141ea2b24ee1\") " pod="openstack/prometheus-metric-storage-0" Nov 21 14:29:11 crc kubenswrapper[4897]: I1121 14:29:11.407777 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hxhfg\" (UniqueName: \"kubernetes.io/projected/91aa2270-92a7-4af4-9a92-141ea2b24ee1-kube-api-access-hxhfg\") pod \"prometheus-metric-storage-0\" (UID: \"91aa2270-92a7-4af4-9a92-141ea2b24ee1\") " pod="openstack/prometheus-metric-storage-0" Nov 21 14:29:11 crc kubenswrapper[4897]: I1121 14:29:11.408397 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/91aa2270-92a7-4af4-9a92-141ea2b24ee1-config\") pod \"prometheus-metric-storage-0\" (UID: \"91aa2270-92a7-4af4-9a92-141ea2b24ee1\") " pod="openstack/prometheus-metric-storage-0" Nov 21 14:29:11 crc kubenswrapper[4897]: I1121 14:29:11.408694 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/91aa2270-92a7-4af4-9a92-141ea2b24ee1-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"91aa2270-92a7-4af4-9a92-141ea2b24ee1\") " pod="openstack/prometheus-metric-storage-0" Nov 21 14:29:11 crc kubenswrapper[4897]: I1121 14:29:11.452636 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-7060bc90-dec6-47cf-93b2-5e0ead1b71e7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7060bc90-dec6-47cf-93b2-5e0ead1b71e7\") pod \"prometheus-metric-storage-0\" (UID: \"91aa2270-92a7-4af4-9a92-141ea2b24ee1\") " pod="openstack/prometheus-metric-storage-0" Nov 21 14:29:11 crc kubenswrapper[4897]: I1121 14:29:11.499064 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Nov 21 14:29:13 crc kubenswrapper[4897]: I1121 14:29:13.067447 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-mm6tz"] Nov 21 14:29:13 crc kubenswrapper[4897]: I1121 14:29:13.069160 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-mm6tz" Nov 21 14:29:13 crc kubenswrapper[4897]: I1121 14:29:13.072156 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Nov 21 14:29:13 crc kubenswrapper[4897]: I1121 14:29:13.072281 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-792x9" Nov 21 14:29:13 crc kubenswrapper[4897]: I1121 14:29:13.072359 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Nov 21 14:29:13 crc kubenswrapper[4897]: I1121 14:29:13.075461 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-k28mx"] Nov 21 14:29:13 crc kubenswrapper[4897]: I1121 14:29:13.079586 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-k28mx" Nov 21 14:29:13 crc kubenswrapper[4897]: I1121 14:29:13.083300 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-mm6tz"] Nov 21 14:29:13 crc kubenswrapper[4897]: I1121 14:29:13.105479 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-k28mx"] Nov 21 14:29:13 crc kubenswrapper[4897]: I1121 14:29:13.114083 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e0c27889-eea7-4a06-9195-3401929cf6b2-var-run-ovn\") pod \"ovn-controller-mm6tz\" (UID: \"e0c27889-eea7-4a06-9195-3401929cf6b2\") " pod="openstack/ovn-controller-mm6tz" Nov 21 14:29:13 crc kubenswrapper[4897]: I1121 14:29:13.114127 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ae71c0a8-809d-4e15-af09-b3a8b326b5a2-scripts\") pod \"ovn-controller-ovs-k28mx\" (UID: \"ae71c0a8-809d-4e15-af09-b3a8b326b5a2\") " pod="openstack/ovn-controller-ovs-k28mx" Nov 21 14:29:13 crc kubenswrapper[4897]: I1121 14:29:13.114147 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/ae71c0a8-809d-4e15-af09-b3a8b326b5a2-etc-ovs\") pod \"ovn-controller-ovs-k28mx\" (UID: \"ae71c0a8-809d-4e15-af09-b3a8b326b5a2\") " pod="openstack/ovn-controller-ovs-k28mx" Nov 21 14:29:13 crc kubenswrapper[4897]: I1121 14:29:13.114161 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/ae71c0a8-809d-4e15-af09-b3a8b326b5a2-var-log\") pod \"ovn-controller-ovs-k28mx\" (UID: \"ae71c0a8-809d-4e15-af09-b3a8b326b5a2\") " pod="openstack/ovn-controller-ovs-k28mx" Nov 21 14:29:13 crc kubenswrapper[4897]: I1121 14:29:13.114218 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/e0c27889-eea7-4a06-9195-3401929cf6b2-ovn-controller-tls-certs\") pod \"ovn-controller-mm6tz\" (UID: \"e0c27889-eea7-4a06-9195-3401929cf6b2\") " pod="openstack/ovn-controller-mm6tz" Nov 21 14:29:13 crc kubenswrapper[4897]: I1121 14:29:13.114238 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e0c27889-eea7-4a06-9195-3401929cf6b2-var-log-ovn\") pod \"ovn-controller-mm6tz\" (UID: \"e0c27889-eea7-4a06-9195-3401929cf6b2\") " pod="openstack/ovn-controller-mm6tz" Nov 21 14:29:13 crc kubenswrapper[4897]: I1121 14:29:13.114264 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e0c27889-eea7-4a06-9195-3401929cf6b2-scripts\") pod \"ovn-controller-mm6tz\" (UID: \"e0c27889-eea7-4a06-9195-3401929cf6b2\") " pod="openstack/ovn-controller-mm6tz" Nov 21 14:29:13 crc kubenswrapper[4897]: I1121 14:29:13.114286 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0c27889-eea7-4a06-9195-3401929cf6b2-combined-ca-bundle\") pod \"ovn-controller-mm6tz\" (UID: \"e0c27889-eea7-4a06-9195-3401929cf6b2\") " pod="openstack/ovn-controller-mm6tz" Nov 21 14:29:13 crc kubenswrapper[4897]: I1121 14:29:13.114315 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e0c27889-eea7-4a06-9195-3401929cf6b2-var-run\") pod \"ovn-controller-mm6tz\" (UID: \"e0c27889-eea7-4a06-9195-3401929cf6b2\") " pod="openstack/ovn-controller-mm6tz" Nov 21 14:29:13 crc kubenswrapper[4897]: I1121 14:29:13.114345 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/ae71c0a8-809d-4e15-af09-b3a8b326b5a2-var-lib\") pod \"ovn-controller-ovs-k28mx\" (UID: \"ae71c0a8-809d-4e15-af09-b3a8b326b5a2\") " pod="openstack/ovn-controller-ovs-k28mx" Nov 21 14:29:13 crc kubenswrapper[4897]: I1121 14:29:13.114366 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7btwt\" (UniqueName: \"kubernetes.io/projected/e0c27889-eea7-4a06-9195-3401929cf6b2-kube-api-access-7btwt\") pod \"ovn-controller-mm6tz\" (UID: \"e0c27889-eea7-4a06-9195-3401929cf6b2\") " pod="openstack/ovn-controller-mm6tz" Nov 21 14:29:13 crc kubenswrapper[4897]: I1121 14:29:13.114421 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zgwkj\" (UniqueName: \"kubernetes.io/projected/ae71c0a8-809d-4e15-af09-b3a8b326b5a2-kube-api-access-zgwkj\") pod \"ovn-controller-ovs-k28mx\" (UID: \"ae71c0a8-809d-4e15-af09-b3a8b326b5a2\") " pod="openstack/ovn-controller-ovs-k28mx" Nov 21 14:29:13 crc kubenswrapper[4897]: I1121 14:29:13.114586 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/ae71c0a8-809d-4e15-af09-b3a8b326b5a2-var-run\") pod \"ovn-controller-ovs-k28mx\" (UID: \"ae71c0a8-809d-4e15-af09-b3a8b326b5a2\") " pod="openstack/ovn-controller-ovs-k28mx" Nov 21 14:29:13 crc kubenswrapper[4897]: I1121 14:29:13.216737 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zgwkj\" (UniqueName: \"kubernetes.io/projected/ae71c0a8-809d-4e15-af09-b3a8b326b5a2-kube-api-access-zgwkj\") pod \"ovn-controller-ovs-k28mx\" (UID: \"ae71c0a8-809d-4e15-af09-b3a8b326b5a2\") " pod="openstack/ovn-controller-ovs-k28mx" Nov 21 14:29:13 crc kubenswrapper[4897]: I1121 14:29:13.216803 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/ae71c0a8-809d-4e15-af09-b3a8b326b5a2-var-run\") pod \"ovn-controller-ovs-k28mx\" (UID: \"ae71c0a8-809d-4e15-af09-b3a8b326b5a2\") " pod="openstack/ovn-controller-ovs-k28mx" Nov 21 14:29:13 crc kubenswrapper[4897]: I1121 14:29:13.217578 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/ae71c0a8-809d-4e15-af09-b3a8b326b5a2-var-run\") pod \"ovn-controller-ovs-k28mx\" (UID: \"ae71c0a8-809d-4e15-af09-b3a8b326b5a2\") " pod="openstack/ovn-controller-ovs-k28mx" Nov 21 14:29:13 crc kubenswrapper[4897]: I1121 14:29:13.216866 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e0c27889-eea7-4a06-9195-3401929cf6b2-var-run-ovn\") pod \"ovn-controller-mm6tz\" (UID: \"e0c27889-eea7-4a06-9195-3401929cf6b2\") " pod="openstack/ovn-controller-mm6tz" Nov 21 14:29:13 crc kubenswrapper[4897]: I1121 14:29:13.217634 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e0c27889-eea7-4a06-9195-3401929cf6b2-var-run-ovn\") pod \"ovn-controller-mm6tz\" (UID: \"e0c27889-eea7-4a06-9195-3401929cf6b2\") " pod="openstack/ovn-controller-mm6tz" Nov 21 14:29:13 crc kubenswrapper[4897]: I1121 14:29:13.217667 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ae71c0a8-809d-4e15-af09-b3a8b326b5a2-scripts\") pod \"ovn-controller-ovs-k28mx\" (UID: \"ae71c0a8-809d-4e15-af09-b3a8b326b5a2\") " pod="openstack/ovn-controller-ovs-k28mx" Nov 21 14:29:13 crc kubenswrapper[4897]: I1121 14:29:13.217691 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/ae71c0a8-809d-4e15-af09-b3a8b326b5a2-etc-ovs\") pod \"ovn-controller-ovs-k28mx\" (UID: \"ae71c0a8-809d-4e15-af09-b3a8b326b5a2\") " pod="openstack/ovn-controller-ovs-k28mx" Nov 21 14:29:13 crc kubenswrapper[4897]: I1121 14:29:13.217710 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/ae71c0a8-809d-4e15-af09-b3a8b326b5a2-var-log\") pod \"ovn-controller-ovs-k28mx\" (UID: \"ae71c0a8-809d-4e15-af09-b3a8b326b5a2\") " pod="openstack/ovn-controller-ovs-k28mx" Nov 21 14:29:13 crc kubenswrapper[4897]: I1121 14:29:13.217795 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/e0c27889-eea7-4a06-9195-3401929cf6b2-ovn-controller-tls-certs\") pod \"ovn-controller-mm6tz\" (UID: \"e0c27889-eea7-4a06-9195-3401929cf6b2\") " pod="openstack/ovn-controller-mm6tz" Nov 21 14:29:13 crc kubenswrapper[4897]: I1121 14:29:13.217825 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e0c27889-eea7-4a06-9195-3401929cf6b2-var-log-ovn\") pod \"ovn-controller-mm6tz\" (UID: \"e0c27889-eea7-4a06-9195-3401929cf6b2\") " pod="openstack/ovn-controller-mm6tz" Nov 21 14:29:13 crc kubenswrapper[4897]: I1121 14:29:13.217857 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e0c27889-eea7-4a06-9195-3401929cf6b2-scripts\") pod \"ovn-controller-mm6tz\" (UID: \"e0c27889-eea7-4a06-9195-3401929cf6b2\") " pod="openstack/ovn-controller-mm6tz" Nov 21 14:29:13 crc kubenswrapper[4897]: I1121 14:29:13.217890 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0c27889-eea7-4a06-9195-3401929cf6b2-combined-ca-bundle\") pod \"ovn-controller-mm6tz\" (UID: \"e0c27889-eea7-4a06-9195-3401929cf6b2\") " pod="openstack/ovn-controller-mm6tz" Nov 21 14:29:13 crc kubenswrapper[4897]: I1121 14:29:13.217929 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e0c27889-eea7-4a06-9195-3401929cf6b2-var-run\") pod \"ovn-controller-mm6tz\" (UID: \"e0c27889-eea7-4a06-9195-3401929cf6b2\") " pod="openstack/ovn-controller-mm6tz" Nov 21 14:29:13 crc kubenswrapper[4897]: I1121 14:29:13.217982 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/ae71c0a8-809d-4e15-af09-b3a8b326b5a2-var-lib\") pod \"ovn-controller-ovs-k28mx\" (UID: \"ae71c0a8-809d-4e15-af09-b3a8b326b5a2\") " pod="openstack/ovn-controller-ovs-k28mx" Nov 21 14:29:13 crc kubenswrapper[4897]: I1121 14:29:13.218043 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7btwt\" (UniqueName: \"kubernetes.io/projected/e0c27889-eea7-4a06-9195-3401929cf6b2-kube-api-access-7btwt\") pod \"ovn-controller-mm6tz\" (UID: \"e0c27889-eea7-4a06-9195-3401929cf6b2\") " pod="openstack/ovn-controller-mm6tz" Nov 21 14:29:13 crc kubenswrapper[4897]: I1121 14:29:13.218583 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/ae71c0a8-809d-4e15-af09-b3a8b326b5a2-etc-ovs\") pod \"ovn-controller-ovs-k28mx\" (UID: \"ae71c0a8-809d-4e15-af09-b3a8b326b5a2\") " pod="openstack/ovn-controller-ovs-k28mx" Nov 21 14:29:13 crc kubenswrapper[4897]: I1121 14:29:13.218736 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/ae71c0a8-809d-4e15-af09-b3a8b326b5a2-var-log\") pod \"ovn-controller-ovs-k28mx\" (UID: \"ae71c0a8-809d-4e15-af09-b3a8b326b5a2\") " pod="openstack/ovn-controller-ovs-k28mx" Nov 21 14:29:13 crc kubenswrapper[4897]: I1121 14:29:13.219155 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e0c27889-eea7-4a06-9195-3401929cf6b2-var-run\") pod \"ovn-controller-mm6tz\" (UID: \"e0c27889-eea7-4a06-9195-3401929cf6b2\") " pod="openstack/ovn-controller-mm6tz" Nov 21 14:29:13 crc kubenswrapper[4897]: I1121 14:29:13.219211 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e0c27889-eea7-4a06-9195-3401929cf6b2-var-log-ovn\") pod \"ovn-controller-mm6tz\" (UID: \"e0c27889-eea7-4a06-9195-3401929cf6b2\") " pod="openstack/ovn-controller-mm6tz" Nov 21 14:29:13 crc kubenswrapper[4897]: I1121 14:29:13.219246 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/ae71c0a8-809d-4e15-af09-b3a8b326b5a2-var-lib\") pod \"ovn-controller-ovs-k28mx\" (UID: \"ae71c0a8-809d-4e15-af09-b3a8b326b5a2\") " pod="openstack/ovn-controller-ovs-k28mx" Nov 21 14:29:13 crc kubenswrapper[4897]: I1121 14:29:13.220434 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ae71c0a8-809d-4e15-af09-b3a8b326b5a2-scripts\") pod \"ovn-controller-ovs-k28mx\" (UID: \"ae71c0a8-809d-4e15-af09-b3a8b326b5a2\") " pod="openstack/ovn-controller-ovs-k28mx" Nov 21 14:29:13 crc kubenswrapper[4897]: I1121 14:29:13.222980 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e0c27889-eea7-4a06-9195-3401929cf6b2-scripts\") pod \"ovn-controller-mm6tz\" (UID: \"e0c27889-eea7-4a06-9195-3401929cf6b2\") " pod="openstack/ovn-controller-mm6tz" Nov 21 14:29:13 crc kubenswrapper[4897]: I1121 14:29:13.227658 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0c27889-eea7-4a06-9195-3401929cf6b2-combined-ca-bundle\") pod \"ovn-controller-mm6tz\" (UID: \"e0c27889-eea7-4a06-9195-3401929cf6b2\") " pod="openstack/ovn-controller-mm6tz" Nov 21 14:29:13 crc kubenswrapper[4897]: I1121 14:29:13.229576 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/e0c27889-eea7-4a06-9195-3401929cf6b2-ovn-controller-tls-certs\") pod \"ovn-controller-mm6tz\" (UID: \"e0c27889-eea7-4a06-9195-3401929cf6b2\") " pod="openstack/ovn-controller-mm6tz" Nov 21 14:29:13 crc kubenswrapper[4897]: I1121 14:29:13.233885 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zgwkj\" (UniqueName: \"kubernetes.io/projected/ae71c0a8-809d-4e15-af09-b3a8b326b5a2-kube-api-access-zgwkj\") pod \"ovn-controller-ovs-k28mx\" (UID: \"ae71c0a8-809d-4e15-af09-b3a8b326b5a2\") " pod="openstack/ovn-controller-ovs-k28mx" Nov 21 14:29:13 crc kubenswrapper[4897]: I1121 14:29:13.238818 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7btwt\" (UniqueName: \"kubernetes.io/projected/e0c27889-eea7-4a06-9195-3401929cf6b2-kube-api-access-7btwt\") pod \"ovn-controller-mm6tz\" (UID: \"e0c27889-eea7-4a06-9195-3401929cf6b2\") " pod="openstack/ovn-controller-mm6tz" Nov 21 14:29:13 crc kubenswrapper[4897]: I1121 14:29:13.394214 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-mm6tz" Nov 21 14:29:13 crc kubenswrapper[4897]: I1121 14:29:13.406657 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-k28mx" Nov 21 14:29:16 crc kubenswrapper[4897]: I1121 14:29:16.152662 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 21 14:29:16 crc kubenswrapper[4897]: I1121 14:29:16.154541 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Nov 21 14:29:16 crc kubenswrapper[4897]: I1121 14:29:16.159911 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Nov 21 14:29:16 crc kubenswrapper[4897]: I1121 14:29:16.159936 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-z4lcf" Nov 21 14:29:16 crc kubenswrapper[4897]: I1121 14:29:16.159969 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Nov 21 14:29:16 crc kubenswrapper[4897]: I1121 14:29:16.160803 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Nov 21 14:29:16 crc kubenswrapper[4897]: I1121 14:29:16.160940 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Nov 21 14:29:16 crc kubenswrapper[4897]: I1121 14:29:16.166544 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 21 14:29:16 crc kubenswrapper[4897]: I1121 14:29:16.284873 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dabe882c-8166-4930-86ad-e2296a7dbf04-config\") pod \"ovsdbserver-nb-0\" (UID: \"dabe882c-8166-4930-86ad-e2296a7dbf04\") " pod="openstack/ovsdbserver-nb-0" Nov 21 14:29:16 crc kubenswrapper[4897]: I1121 14:29:16.284915 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dabe882c-8166-4930-86ad-e2296a7dbf04-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"dabe882c-8166-4930-86ad-e2296a7dbf04\") " pod="openstack/ovsdbserver-nb-0" Nov 21 14:29:16 crc kubenswrapper[4897]: I1121 14:29:16.284935 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dabe882c-8166-4930-86ad-e2296a7dbf04-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"dabe882c-8166-4930-86ad-e2296a7dbf04\") " pod="openstack/ovsdbserver-nb-0" Nov 21 14:29:16 crc kubenswrapper[4897]: I1121 14:29:16.284976 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nfmd8\" (UniqueName: \"kubernetes.io/projected/dabe882c-8166-4930-86ad-e2296a7dbf04-kube-api-access-nfmd8\") pod \"ovsdbserver-nb-0\" (UID: \"dabe882c-8166-4930-86ad-e2296a7dbf04\") " pod="openstack/ovsdbserver-nb-0" Nov 21 14:29:16 crc kubenswrapper[4897]: I1121 14:29:16.285341 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/dabe882c-8166-4930-86ad-e2296a7dbf04-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"dabe882c-8166-4930-86ad-e2296a7dbf04\") " pod="openstack/ovsdbserver-nb-0" Nov 21 14:29:16 crc kubenswrapper[4897]: I1121 14:29:16.285429 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/dabe882c-8166-4930-86ad-e2296a7dbf04-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"dabe882c-8166-4930-86ad-e2296a7dbf04\") " pod="openstack/ovsdbserver-nb-0" Nov 21 14:29:16 crc kubenswrapper[4897]: I1121 14:29:16.285551 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-nb-0\" (UID: \"dabe882c-8166-4930-86ad-e2296a7dbf04\") " pod="openstack/ovsdbserver-nb-0" Nov 21 14:29:16 crc kubenswrapper[4897]: I1121 14:29:16.285659 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/dabe882c-8166-4930-86ad-e2296a7dbf04-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"dabe882c-8166-4930-86ad-e2296a7dbf04\") " pod="openstack/ovsdbserver-nb-0" Nov 21 14:29:16 crc kubenswrapper[4897]: I1121 14:29:16.387801 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/dabe882c-8166-4930-86ad-e2296a7dbf04-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"dabe882c-8166-4930-86ad-e2296a7dbf04\") " pod="openstack/ovsdbserver-nb-0" Nov 21 14:29:16 crc kubenswrapper[4897]: I1121 14:29:16.387928 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dabe882c-8166-4930-86ad-e2296a7dbf04-config\") pod \"ovsdbserver-nb-0\" (UID: \"dabe882c-8166-4930-86ad-e2296a7dbf04\") " pod="openstack/ovsdbserver-nb-0" Nov 21 14:29:16 crc kubenswrapper[4897]: I1121 14:29:16.387966 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dabe882c-8166-4930-86ad-e2296a7dbf04-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"dabe882c-8166-4930-86ad-e2296a7dbf04\") " pod="openstack/ovsdbserver-nb-0" Nov 21 14:29:16 crc kubenswrapper[4897]: I1121 14:29:16.387986 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dabe882c-8166-4930-86ad-e2296a7dbf04-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"dabe882c-8166-4930-86ad-e2296a7dbf04\") " pod="openstack/ovsdbserver-nb-0" Nov 21 14:29:16 crc kubenswrapper[4897]: I1121 14:29:16.388043 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nfmd8\" (UniqueName: \"kubernetes.io/projected/dabe882c-8166-4930-86ad-e2296a7dbf04-kube-api-access-nfmd8\") pod \"ovsdbserver-nb-0\" (UID: \"dabe882c-8166-4930-86ad-e2296a7dbf04\") " pod="openstack/ovsdbserver-nb-0" Nov 21 14:29:16 crc kubenswrapper[4897]: I1121 14:29:16.388175 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/dabe882c-8166-4930-86ad-e2296a7dbf04-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"dabe882c-8166-4930-86ad-e2296a7dbf04\") " pod="openstack/ovsdbserver-nb-0" Nov 21 14:29:16 crc kubenswrapper[4897]: I1121 14:29:16.388213 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/dabe882c-8166-4930-86ad-e2296a7dbf04-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"dabe882c-8166-4930-86ad-e2296a7dbf04\") " pod="openstack/ovsdbserver-nb-0" Nov 21 14:29:16 crc kubenswrapper[4897]: I1121 14:29:16.388261 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-nb-0\" (UID: \"dabe882c-8166-4930-86ad-e2296a7dbf04\") " pod="openstack/ovsdbserver-nb-0" Nov 21 14:29:16 crc kubenswrapper[4897]: I1121 14:29:16.388653 4897 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-nb-0\" (UID: \"dabe882c-8166-4930-86ad-e2296a7dbf04\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/ovsdbserver-nb-0" Nov 21 14:29:16 crc kubenswrapper[4897]: I1121 14:29:16.389391 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dabe882c-8166-4930-86ad-e2296a7dbf04-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"dabe882c-8166-4930-86ad-e2296a7dbf04\") " pod="openstack/ovsdbserver-nb-0" Nov 21 14:29:16 crc kubenswrapper[4897]: I1121 14:29:16.389438 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/dabe882c-8166-4930-86ad-e2296a7dbf04-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"dabe882c-8166-4930-86ad-e2296a7dbf04\") " pod="openstack/ovsdbserver-nb-0" Nov 21 14:29:16 crc kubenswrapper[4897]: I1121 14:29:16.389551 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dabe882c-8166-4930-86ad-e2296a7dbf04-config\") pod \"ovsdbserver-nb-0\" (UID: \"dabe882c-8166-4930-86ad-e2296a7dbf04\") " pod="openstack/ovsdbserver-nb-0" Nov 21 14:29:16 crc kubenswrapper[4897]: I1121 14:29:16.394212 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dabe882c-8166-4930-86ad-e2296a7dbf04-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"dabe882c-8166-4930-86ad-e2296a7dbf04\") " pod="openstack/ovsdbserver-nb-0" Nov 21 14:29:16 crc kubenswrapper[4897]: I1121 14:29:16.400131 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/dabe882c-8166-4930-86ad-e2296a7dbf04-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"dabe882c-8166-4930-86ad-e2296a7dbf04\") " pod="openstack/ovsdbserver-nb-0" Nov 21 14:29:16 crc kubenswrapper[4897]: I1121 14:29:16.401074 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/dabe882c-8166-4930-86ad-e2296a7dbf04-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"dabe882c-8166-4930-86ad-e2296a7dbf04\") " pod="openstack/ovsdbserver-nb-0" Nov 21 14:29:16 crc kubenswrapper[4897]: I1121 14:29:16.404131 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nfmd8\" (UniqueName: \"kubernetes.io/projected/dabe882c-8166-4930-86ad-e2296a7dbf04-kube-api-access-nfmd8\") pod \"ovsdbserver-nb-0\" (UID: \"dabe882c-8166-4930-86ad-e2296a7dbf04\") " pod="openstack/ovsdbserver-nb-0" Nov 21 14:29:16 crc kubenswrapper[4897]: I1121 14:29:16.426836 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-nb-0\" (UID: \"dabe882c-8166-4930-86ad-e2296a7dbf04\") " pod="openstack/ovsdbserver-nb-0" Nov 21 14:29:16 crc kubenswrapper[4897]: I1121 14:29:16.484770 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Nov 21 14:29:16 crc kubenswrapper[4897]: I1121 14:29:16.885646 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"fc30a810-d114-48ee-97c2-414f1ed23d22","Type":"ContainerStarted","Data":"396044007fce4debb20b289db599107c5e336e63572f5faaaed4ece8f6b4a31e"} Nov 21 14:29:17 crc kubenswrapper[4897]: I1121 14:29:17.141296 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 21 14:29:17 crc kubenswrapper[4897]: I1121 14:29:17.142800 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Nov 21 14:29:17 crc kubenswrapper[4897]: I1121 14:29:17.145262 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Nov 21 14:29:17 crc kubenswrapper[4897]: I1121 14:29:17.151968 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Nov 21 14:29:17 crc kubenswrapper[4897]: I1121 14:29:17.152151 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-px67j" Nov 21 14:29:17 crc kubenswrapper[4897]: I1121 14:29:17.152421 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Nov 21 14:29:17 crc kubenswrapper[4897]: I1121 14:29:17.160626 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 21 14:29:17 crc kubenswrapper[4897]: I1121 14:29:17.210754 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8c8fd3ef-d89c-429e-a82b-454894334b2d-config\") pod \"ovsdbserver-sb-0\" (UID: \"8c8fd3ef-d89c-429e-a82b-454894334b2d\") " pod="openstack/ovsdbserver-sb-0" Nov 21 14:29:17 crc kubenswrapper[4897]: I1121 14:29:17.210804 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/8c8fd3ef-d89c-429e-a82b-454894334b2d-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"8c8fd3ef-d89c-429e-a82b-454894334b2d\") " pod="openstack/ovsdbserver-sb-0" Nov 21 14:29:17 crc kubenswrapper[4897]: I1121 14:29:17.210850 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-krmfs\" (UniqueName: \"kubernetes.io/projected/8c8fd3ef-d89c-429e-a82b-454894334b2d-kube-api-access-krmfs\") pod \"ovsdbserver-sb-0\" (UID: \"8c8fd3ef-d89c-429e-a82b-454894334b2d\") " pod="openstack/ovsdbserver-sb-0" Nov 21 14:29:17 crc kubenswrapper[4897]: I1121 14:29:17.210890 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c8fd3ef-d89c-429e-a82b-454894334b2d-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"8c8fd3ef-d89c-429e-a82b-454894334b2d\") " pod="openstack/ovsdbserver-sb-0" Nov 21 14:29:17 crc kubenswrapper[4897]: I1121 14:29:17.210918 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8c8fd3ef-d89c-429e-a82b-454894334b2d-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"8c8fd3ef-d89c-429e-a82b-454894334b2d\") " pod="openstack/ovsdbserver-sb-0" Nov 21 14:29:17 crc kubenswrapper[4897]: I1121 14:29:17.210972 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8c8fd3ef-d89c-429e-a82b-454894334b2d-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"8c8fd3ef-d89c-429e-a82b-454894334b2d\") " pod="openstack/ovsdbserver-sb-0" Nov 21 14:29:17 crc kubenswrapper[4897]: I1121 14:29:17.211026 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-sb-0\" (UID: \"8c8fd3ef-d89c-429e-a82b-454894334b2d\") " pod="openstack/ovsdbserver-sb-0" Nov 21 14:29:17 crc kubenswrapper[4897]: I1121 14:29:17.211052 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8c8fd3ef-d89c-429e-a82b-454894334b2d-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"8c8fd3ef-d89c-429e-a82b-454894334b2d\") " pod="openstack/ovsdbserver-sb-0" Nov 21 14:29:17 crc kubenswrapper[4897]: I1121 14:29:17.312359 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8c8fd3ef-d89c-429e-a82b-454894334b2d-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"8c8fd3ef-d89c-429e-a82b-454894334b2d\") " pod="openstack/ovsdbserver-sb-0" Nov 21 14:29:17 crc kubenswrapper[4897]: I1121 14:29:17.312763 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8c8fd3ef-d89c-429e-a82b-454894334b2d-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"8c8fd3ef-d89c-429e-a82b-454894334b2d\") " pod="openstack/ovsdbserver-sb-0" Nov 21 14:29:17 crc kubenswrapper[4897]: I1121 14:29:17.312844 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-sb-0\" (UID: \"8c8fd3ef-d89c-429e-a82b-454894334b2d\") " pod="openstack/ovsdbserver-sb-0" Nov 21 14:29:17 crc kubenswrapper[4897]: I1121 14:29:17.312878 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8c8fd3ef-d89c-429e-a82b-454894334b2d-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"8c8fd3ef-d89c-429e-a82b-454894334b2d\") " pod="openstack/ovsdbserver-sb-0" Nov 21 14:29:17 crc kubenswrapper[4897]: I1121 14:29:17.312951 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8c8fd3ef-d89c-429e-a82b-454894334b2d-config\") pod \"ovsdbserver-sb-0\" (UID: \"8c8fd3ef-d89c-429e-a82b-454894334b2d\") " pod="openstack/ovsdbserver-sb-0" Nov 21 14:29:17 crc kubenswrapper[4897]: I1121 14:29:17.312980 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/8c8fd3ef-d89c-429e-a82b-454894334b2d-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"8c8fd3ef-d89c-429e-a82b-454894334b2d\") " pod="openstack/ovsdbserver-sb-0" Nov 21 14:29:17 crc kubenswrapper[4897]: I1121 14:29:17.313034 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-krmfs\" (UniqueName: \"kubernetes.io/projected/8c8fd3ef-d89c-429e-a82b-454894334b2d-kube-api-access-krmfs\") pod \"ovsdbserver-sb-0\" (UID: \"8c8fd3ef-d89c-429e-a82b-454894334b2d\") " pod="openstack/ovsdbserver-sb-0" Nov 21 14:29:17 crc kubenswrapper[4897]: I1121 14:29:17.313063 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c8fd3ef-d89c-429e-a82b-454894334b2d-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"8c8fd3ef-d89c-429e-a82b-454894334b2d\") " pod="openstack/ovsdbserver-sb-0" Nov 21 14:29:17 crc kubenswrapper[4897]: I1121 14:29:17.314218 4897 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-sb-0\" (UID: \"8c8fd3ef-d89c-429e-a82b-454894334b2d\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/ovsdbserver-sb-0" Nov 21 14:29:17 crc kubenswrapper[4897]: I1121 14:29:17.316322 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/8c8fd3ef-d89c-429e-a82b-454894334b2d-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"8c8fd3ef-d89c-429e-a82b-454894334b2d\") " pod="openstack/ovsdbserver-sb-0" Nov 21 14:29:17 crc kubenswrapper[4897]: I1121 14:29:17.316856 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8c8fd3ef-d89c-429e-a82b-454894334b2d-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"8c8fd3ef-d89c-429e-a82b-454894334b2d\") " pod="openstack/ovsdbserver-sb-0" Nov 21 14:29:17 crc kubenswrapper[4897]: I1121 14:29:17.316944 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8c8fd3ef-d89c-429e-a82b-454894334b2d-config\") pod \"ovsdbserver-sb-0\" (UID: \"8c8fd3ef-d89c-429e-a82b-454894334b2d\") " pod="openstack/ovsdbserver-sb-0" Nov 21 14:29:17 crc kubenswrapper[4897]: I1121 14:29:17.319377 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8c8fd3ef-d89c-429e-a82b-454894334b2d-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"8c8fd3ef-d89c-429e-a82b-454894334b2d\") " pod="openstack/ovsdbserver-sb-0" Nov 21 14:29:17 crc kubenswrapper[4897]: I1121 14:29:17.319420 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8c8fd3ef-d89c-429e-a82b-454894334b2d-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"8c8fd3ef-d89c-429e-a82b-454894334b2d\") " pod="openstack/ovsdbserver-sb-0" Nov 21 14:29:17 crc kubenswrapper[4897]: I1121 14:29:17.320292 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c8fd3ef-d89c-429e-a82b-454894334b2d-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"8c8fd3ef-d89c-429e-a82b-454894334b2d\") " pod="openstack/ovsdbserver-sb-0" Nov 21 14:29:17 crc kubenswrapper[4897]: I1121 14:29:17.335890 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-krmfs\" (UniqueName: \"kubernetes.io/projected/8c8fd3ef-d89c-429e-a82b-454894334b2d-kube-api-access-krmfs\") pod \"ovsdbserver-sb-0\" (UID: \"8c8fd3ef-d89c-429e-a82b-454894334b2d\") " pod="openstack/ovsdbserver-sb-0" Nov 21 14:29:17 crc kubenswrapper[4897]: I1121 14:29:17.358931 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-sb-0\" (UID: \"8c8fd3ef-d89c-429e-a82b-454894334b2d\") " pod="openstack/ovsdbserver-sb-0" Nov 21 14:29:17 crc kubenswrapper[4897]: I1121 14:29:17.481409 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Nov 21 14:29:38 crc kubenswrapper[4897]: E1121 14:29:38.151863 4897 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified" Nov 21 14:29:38 crc kubenswrapper[4897]: E1121 14:29:38.152907 4897 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s2gt5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-server-0_openstack(b513e4ee-7c30-4942-b997-4d52414e1ec0): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 21 14:29:38 crc kubenswrapper[4897]: E1121 14:29:38.154426 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-server-0" podUID="b513e4ee-7c30-4942-b997-4d52414e1ec0" Nov 21 14:29:38 crc kubenswrapper[4897]: E1121 14:29:38.197239 4897 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified" Nov 21 14:29:38 crc kubenswrapper[4897]: E1121 14:29:38.197412 4897 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-kbnvn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cell1-server-0_openstack(bd692188-6da6-4387-a46d-003a2da0d0c8): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 21 14:29:38 crc kubenswrapper[4897]: E1121 14:29:38.198722 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-cell1-server-0" podUID="bd692188-6da6-4387-a46d-003a2da0d0c8" Nov 21 14:29:39 crc kubenswrapper[4897]: E1121 14:29:39.108772 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified\\\"\"" pod="openstack/rabbitmq-server-0" podUID="b513e4ee-7c30-4942-b997-4d52414e1ec0" Nov 21 14:29:39 crc kubenswrapper[4897]: E1121 14:29:39.108930 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified\\\"\"" pod="openstack/rabbitmq-cell1-server-0" podUID="bd692188-6da6-4387-a46d-003a2da0d0c8" Nov 21 14:29:39 crc kubenswrapper[4897]: E1121 14:29:39.700561 4897 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-mariadb:current-podified" Nov 21 14:29:39 crc kubenswrapper[4897]: E1121 14:29:39.700734 4897 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:mysql-bootstrap,Image:quay.io/podified-antelope-centos9/openstack-mariadb:current-podified,Command:[bash /var/lib/operator-scripts/mysql_bootstrap.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:True,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:mysql-db,ReadOnly:false,MountPath:/var/lib/mysql,SubPath:mysql,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-default,ReadOnly:true,MountPath:/var/lib/config-data/default,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-generated,ReadOnly:false,MountPath:/var/lib/config-data/generated,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:operator-scripts,ReadOnly:true,MountPath:/var/lib/operator-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tfr5n,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-galera-0_openstack(fce5a726-bd06-473b-a9d0-f812ffe83693): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 21 14:29:39 crc kubenswrapper[4897]: E1121 14:29:39.701946 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/openstack-galera-0" podUID="fce5a726-bd06-473b-a9d0-f812ffe83693" Nov 21 14:29:40 crc kubenswrapper[4897]: E1121 14:29:40.129968 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-mariadb:current-podified\\\"\"" pod="openstack/openstack-galera-0" podUID="fce5a726-bd06-473b-a9d0-f812ffe83693" Nov 21 14:29:40 crc kubenswrapper[4897]: E1121 14:29:40.681772 4897 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Nov 21 14:29:40 crc kubenswrapper[4897]: E1121 14:29:40.682079 4897 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bw8j8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-fpb49_openstack(7522ba8b-cc71-459f-9f6a-fbd0f646dce3): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 21 14:29:40 crc kubenswrapper[4897]: E1121 14:29:40.683309 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-fpb49" podUID="7522ba8b-cc71-459f-9f6a-fbd0f646dce3" Nov 21 14:29:40 crc kubenswrapper[4897]: E1121 14:29:40.766564 4897 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Nov 21 14:29:40 crc kubenswrapper[4897]: E1121 14:29:40.766754 4897 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rqwlp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-57d769cc4f-8f9xc_openstack(ecceef5c-b1e5-4c3c-97b3-d805494cfba3): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 21 14:29:40 crc kubenswrapper[4897]: E1121 14:29:40.767969 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-57d769cc4f-8f9xc" podUID="ecceef5c-b1e5-4c3c-97b3-d805494cfba3" Nov 21 14:29:40 crc kubenswrapper[4897]: E1121 14:29:40.771422 4897 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Nov 21 14:29:40 crc kubenswrapper[4897]: E1121 14:29:40.771569 4897 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6b5qf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-666b6646f7-khkvq_openstack(6bf04b77-0ef9-42f0-8a02-8c4c7568e390): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 21 14:29:40 crc kubenswrapper[4897]: E1121 14:29:40.773625 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-666b6646f7-khkvq" podUID="6bf04b77-0ef9-42f0-8a02-8c4c7568e390" Nov 21 14:29:41 crc kubenswrapper[4897]: E1121 14:29:41.137854 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-57d769cc4f-8f9xc" podUID="ecceef5c-b1e5-4c3c-97b3-d805494cfba3" Nov 21 14:29:41 crc kubenswrapper[4897]: E1121 14:29:41.137876 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-666b6646f7-khkvq" podUID="6bf04b77-0ef9-42f0-8a02-8c4c7568e390" Nov 21 14:29:42 crc kubenswrapper[4897]: I1121 14:29:42.012433 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-fpb49" Nov 21 14:29:42 crc kubenswrapper[4897]: E1121 14:29:42.041718 4897 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Nov 21 14:29:42 crc kubenswrapper[4897]: E1121 14:29:42.041881 4897 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s2m5c,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-whhtg_openstack(f6215eb2-5fde-4880-b101-a3be3a015d93): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 21 14:29:42 crc kubenswrapper[4897]: E1121 14:29:42.044127 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-whhtg" podUID="f6215eb2-5fde-4880-b101-a3be3a015d93" Nov 21 14:29:42 crc kubenswrapper[4897]: I1121 14:29:42.079600 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7522ba8b-cc71-459f-9f6a-fbd0f646dce3-config\") pod \"7522ba8b-cc71-459f-9f6a-fbd0f646dce3\" (UID: \"7522ba8b-cc71-459f-9f6a-fbd0f646dce3\") " Nov 21 14:29:42 crc kubenswrapper[4897]: I1121 14:29:42.079799 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bw8j8\" (UniqueName: \"kubernetes.io/projected/7522ba8b-cc71-459f-9f6a-fbd0f646dce3-kube-api-access-bw8j8\") pod \"7522ba8b-cc71-459f-9f6a-fbd0f646dce3\" (UID: \"7522ba8b-cc71-459f-9f6a-fbd0f646dce3\") " Nov 21 14:29:42 crc kubenswrapper[4897]: I1121 14:29:42.080140 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7522ba8b-cc71-459f-9f6a-fbd0f646dce3-config" (OuterVolumeSpecName: "config") pod "7522ba8b-cc71-459f-9f6a-fbd0f646dce3" (UID: "7522ba8b-cc71-459f-9f6a-fbd0f646dce3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:29:42 crc kubenswrapper[4897]: I1121 14:29:42.080279 4897 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7522ba8b-cc71-459f-9f6a-fbd0f646dce3-config\") on node \"crc\" DevicePath \"\"" Nov 21 14:29:42 crc kubenswrapper[4897]: I1121 14:29:42.083566 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7522ba8b-cc71-459f-9f6a-fbd0f646dce3-kube-api-access-bw8j8" (OuterVolumeSpecName: "kube-api-access-bw8j8") pod "7522ba8b-cc71-459f-9f6a-fbd0f646dce3" (UID: "7522ba8b-cc71-459f-9f6a-fbd0f646dce3"). InnerVolumeSpecName "kube-api-access-bw8j8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:29:42 crc kubenswrapper[4897]: I1121 14:29:42.157738 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-fpb49" Nov 21 14:29:42 crc kubenswrapper[4897]: I1121 14:29:42.157781 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-fpb49" event={"ID":"7522ba8b-cc71-459f-9f6a-fbd0f646dce3","Type":"ContainerDied","Data":"5ef9f69ff7ec34c24c0d19c2ed4fa3ae7f9f06486e4607f69a6c793476e08763"} Nov 21 14:29:42 crc kubenswrapper[4897]: I1121 14:29:42.165338 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 21 14:29:42 crc kubenswrapper[4897]: W1121 14:29:42.167183 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod95ad29e4_b4b6_4cec_b978_23f7ea25edc6.slice/crio-f02f019457b52e5712c17087ebfe728d1d845fff91f68d9f881356575fb8cb4a WatchSource:0}: Error finding container f02f019457b52e5712c17087ebfe728d1d845fff91f68d9f881356575fb8cb4a: Status 404 returned error can't find the container with id f02f019457b52e5712c17087ebfe728d1d845fff91f68d9f881356575fb8cb4a Nov 21 14:29:42 crc kubenswrapper[4897]: I1121 14:29:42.187885 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bw8j8\" (UniqueName: \"kubernetes.io/projected/7522ba8b-cc71-459f-9f6a-fbd0f646dce3-kube-api-access-bw8j8\") on node \"crc\" DevicePath \"\"" Nov 21 14:29:42 crc kubenswrapper[4897]: I1121 14:29:42.191603 4897 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 21 14:29:42 crc kubenswrapper[4897]: I1121 14:29:42.226364 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-fpb49"] Nov 21 14:29:42 crc kubenswrapper[4897]: I1121 14:29:42.232790 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-fpb49"] Nov 21 14:29:42 crc kubenswrapper[4897]: I1121 14:29:42.393633 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 21 14:29:42 crc kubenswrapper[4897]: W1121 14:29:42.396856 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod91aa2270_92a7_4af4_9a92_141ea2b24ee1.slice/crio-751b0fdc381aeca542be156ad6c66f3e7d3df396c03bdd04b0819a816b3e00e9 WatchSource:0}: Error finding container 751b0fdc381aeca542be156ad6c66f3e7d3df396c03bdd04b0819a816b3e00e9: Status 404 returned error can't find the container with id 751b0fdc381aeca542be156ad6c66f3e7d3df396c03bdd04b0819a816b3e00e9 Nov 21 14:29:42 crc kubenswrapper[4897]: I1121 14:29:42.646786 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-whhtg" Nov 21 14:29:42 crc kubenswrapper[4897]: I1121 14:29:42.698187 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s2m5c\" (UniqueName: \"kubernetes.io/projected/f6215eb2-5fde-4880-b101-a3be3a015d93-kube-api-access-s2m5c\") pod \"f6215eb2-5fde-4880-b101-a3be3a015d93\" (UID: \"f6215eb2-5fde-4880-b101-a3be3a015d93\") " Nov 21 14:29:42 crc kubenswrapper[4897]: I1121 14:29:42.698305 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f6215eb2-5fde-4880-b101-a3be3a015d93-config\") pod \"f6215eb2-5fde-4880-b101-a3be3a015d93\" (UID: \"f6215eb2-5fde-4880-b101-a3be3a015d93\") " Nov 21 14:29:42 crc kubenswrapper[4897]: I1121 14:29:42.698394 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f6215eb2-5fde-4880-b101-a3be3a015d93-dns-svc\") pod \"f6215eb2-5fde-4880-b101-a3be3a015d93\" (UID: \"f6215eb2-5fde-4880-b101-a3be3a015d93\") " Nov 21 14:29:42 crc kubenswrapper[4897]: I1121 14:29:42.715742 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f6215eb2-5fde-4880-b101-a3be3a015d93-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f6215eb2-5fde-4880-b101-a3be3a015d93" (UID: "f6215eb2-5fde-4880-b101-a3be3a015d93"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:29:42 crc kubenswrapper[4897]: I1121 14:29:42.716474 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f6215eb2-5fde-4880-b101-a3be3a015d93-config" (OuterVolumeSpecName: "config") pod "f6215eb2-5fde-4880-b101-a3be3a015d93" (UID: "f6215eb2-5fde-4880-b101-a3be3a015d93"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:29:42 crc kubenswrapper[4897]: I1121 14:29:42.722367 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6215eb2-5fde-4880-b101-a3be3a015d93-kube-api-access-s2m5c" (OuterVolumeSpecName: "kube-api-access-s2m5c") pod "f6215eb2-5fde-4880-b101-a3be3a015d93" (UID: "f6215eb2-5fde-4880-b101-a3be3a015d93"). InnerVolumeSpecName "kube-api-access-s2m5c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:29:42 crc kubenswrapper[4897]: I1121 14:29:42.762615 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-84f7c79c46-td7jx"] Nov 21 14:29:42 crc kubenswrapper[4897]: W1121 14:29:42.771682 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6af809de_112c_4988_a06e_4838ef479ace.slice/crio-de5df286662eed1a58d32ebf0a5c55484dd16a0a9d834dcb3654b189633f8104 WatchSource:0}: Error finding container de5df286662eed1a58d32ebf0a5c55484dd16a0a9d834dcb3654b189633f8104: Status 404 returned error can't find the container with id de5df286662eed1a58d32ebf0a5c55484dd16a0a9d834dcb3654b189633f8104 Nov 21 14:29:42 crc kubenswrapper[4897]: I1121 14:29:42.781022 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-mm6tz"] Nov 21 14:29:42 crc kubenswrapper[4897]: I1121 14:29:42.793326 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-ui-dashboards-7d5fb4cbfb-mt4vk"] Nov 21 14:29:42 crc kubenswrapper[4897]: I1121 14:29:42.816887 4897 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f6215eb2-5fde-4880-b101-a3be3a015d93-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 21 14:29:42 crc kubenswrapper[4897]: I1121 14:29:42.816921 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s2m5c\" (UniqueName: \"kubernetes.io/projected/f6215eb2-5fde-4880-b101-a3be3a015d93-kube-api-access-s2m5c\") on node \"crc\" DevicePath \"\"" Nov 21 14:29:42 crc kubenswrapper[4897]: I1121 14:29:42.816933 4897 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f6215eb2-5fde-4880-b101-a3be3a015d93-config\") on node \"crc\" DevicePath \"\"" Nov 21 14:29:43 crc kubenswrapper[4897]: I1121 14:29:43.169257 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-mm6tz" event={"ID":"e0c27889-eea7-4a06-9195-3401929cf6b2","Type":"ContainerStarted","Data":"432e78a238d8aba699f8e445111df40f54257459d1226e3d0c2a4bdb2fe600d1"} Nov 21 14:29:43 crc kubenswrapper[4897]: I1121 14:29:43.170915 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-whhtg" Nov 21 14:29:43 crc kubenswrapper[4897]: I1121 14:29:43.170886 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-whhtg" event={"ID":"f6215eb2-5fde-4880-b101-a3be3a015d93","Type":"ContainerDied","Data":"753f2f00a3cf405c991e722c415ddc0a184c0c77525a31247e0f551ae714e181"} Nov 21 14:29:43 crc kubenswrapper[4897]: I1121 14:29:43.172790 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-84f7c79c46-td7jx" event={"ID":"dc6ff181-277d-4265-b2bb-bed01dc666b4","Type":"ContainerStarted","Data":"ddb696bf9ffa762e6b29de95943091ba397d6097cf8700143d88870209d7221d"} Nov 21 14:29:43 crc kubenswrapper[4897]: I1121 14:29:43.172856 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-84f7c79c46-td7jx" event={"ID":"dc6ff181-277d-4265-b2bb-bed01dc666b4","Type":"ContainerStarted","Data":"52cd85db13eff7bdcb6d262a083a3192c9fe5fe9c384c4b3560e0a0a55db9472"} Nov 21 14:29:43 crc kubenswrapper[4897]: I1121 14:29:43.177788 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"fc30a810-d114-48ee-97c2-414f1ed23d22","Type":"ContainerStarted","Data":"635453603cba392256bda2ba216dffaf01ed30a029ea19e59b04ef3ddef37bf6"} Nov 21 14:29:43 crc kubenswrapper[4897]: I1121 14:29:43.177914 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Nov 21 14:29:43 crc kubenswrapper[4897]: I1121 14:29:43.179678 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"91aa2270-92a7-4af4-9a92-141ea2b24ee1","Type":"ContainerStarted","Data":"751b0fdc381aeca542be156ad6c66f3e7d3df396c03bdd04b0819a816b3e00e9"} Nov 21 14:29:43 crc kubenswrapper[4897]: I1121 14:29:43.180776 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"95ad29e4-b4b6-4cec-b978-23f7ea25edc6","Type":"ContainerStarted","Data":"f02f019457b52e5712c17087ebfe728d1d845fff91f68d9f881356575fb8cb4a"} Nov 21 14:29:43 crc kubenswrapper[4897]: I1121 14:29:43.181774 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-mt4vk" event={"ID":"6af809de-112c-4988-a06e-4838ef479ace","Type":"ContainerStarted","Data":"de5df286662eed1a58d32ebf0a5c55484dd16a0a9d834dcb3654b189633f8104"} Nov 21 14:29:43 crc kubenswrapper[4897]: I1121 14:29:43.183057 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"d3691270-398c-46e4-9db9-de8662274fd3","Type":"ContainerStarted","Data":"d9743b6acb206f8ed36f42d844d924a7f8d7953ced2da330e452478112dbd4f5"} Nov 21 14:29:43 crc kubenswrapper[4897]: I1121 14:29:43.216888 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-84f7c79c46-td7jx" podStartSLOduration=33.216869092 podStartE2EDuration="33.216869092s" podCreationTimestamp="2025-11-21 14:29:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:29:43.204162771 +0000 UTC m=+1260.488756276" watchObservedRunningTime="2025-11-21 14:29:43.216869092 +0000 UTC m=+1260.501462587" Nov 21 14:29:43 crc kubenswrapper[4897]: I1121 14:29:43.268796 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-whhtg"] Nov 21 14:29:43 crc kubenswrapper[4897]: I1121 14:29:43.279983 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-whhtg"] Nov 21 14:29:43 crc kubenswrapper[4897]: I1121 14:29:43.287529 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=10.40605929 podStartE2EDuration="36.287488995s" podCreationTimestamp="2025-11-21 14:29:07 +0000 UTC" firstStartedPulling="2025-11-21 14:29:16.213692298 +0000 UTC m=+1233.498285773" lastFinishedPulling="2025-11-21 14:29:42.095122003 +0000 UTC m=+1259.379715478" observedRunningTime="2025-11-21 14:29:43.262031533 +0000 UTC m=+1260.546625028" watchObservedRunningTime="2025-11-21 14:29:43.287488995 +0000 UTC m=+1260.572082470" Nov 21 14:29:43 crc kubenswrapper[4897]: W1121 14:29:43.493865 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8c8fd3ef_d89c_429e_a82b_454894334b2d.slice/crio-075678a2614cadbe6ceefd420f5bb8abe09393796c67a06e6fa16980e92b0f8c WatchSource:0}: Error finding container 075678a2614cadbe6ceefd420f5bb8abe09393796c67a06e6fa16980e92b0f8c: Status 404 returned error can't find the container with id 075678a2614cadbe6ceefd420f5bb8abe09393796c67a06e6fa16980e92b0f8c Nov 21 14:29:43 crc kubenswrapper[4897]: I1121 14:29:43.494231 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 21 14:29:44 crc kubenswrapper[4897]: I1121 14:29:44.078840 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-k28mx"] Nov 21 14:29:44 crc kubenswrapper[4897]: I1121 14:29:44.110744 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7522ba8b-cc71-459f-9f6a-fbd0f646dce3" path="/var/lib/kubelet/pods/7522ba8b-cc71-459f-9f6a-fbd0f646dce3/volumes" Nov 21 14:29:44 crc kubenswrapper[4897]: I1121 14:29:44.111231 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f6215eb2-5fde-4880-b101-a3be3a015d93" path="/var/lib/kubelet/pods/f6215eb2-5fde-4880-b101-a3be3a015d93/volumes" Nov 21 14:29:44 crc kubenswrapper[4897]: I1121 14:29:44.203233 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"8c8fd3ef-d89c-429e-a82b-454894334b2d","Type":"ContainerStarted","Data":"075678a2614cadbe6ceefd420f5bb8abe09393796c67a06e6fa16980e92b0f8c"} Nov 21 14:29:44 crc kubenswrapper[4897]: I1121 14:29:44.220360 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 21 14:29:44 crc kubenswrapper[4897]: W1121 14:29:44.443596 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podae71c0a8_809d_4e15_af09_b3a8b326b5a2.slice/crio-63dfb6ce5220feb534d3c3db9c1abc8bc6dece4e243e64c8d30ccf0118bef7aa WatchSource:0}: Error finding container 63dfb6ce5220feb534d3c3db9c1abc8bc6dece4e243e64c8d30ccf0118bef7aa: Status 404 returned error can't find the container with id 63dfb6ce5220feb534d3c3db9c1abc8bc6dece4e243e64c8d30ccf0118bef7aa Nov 21 14:29:45 crc kubenswrapper[4897]: I1121 14:29:45.216671 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"dabe882c-8166-4930-86ad-e2296a7dbf04","Type":"ContainerStarted","Data":"f319485fd10f472024667851686f663d2855f5e81150d9b45eed063e7c499213"} Nov 21 14:29:45 crc kubenswrapper[4897]: I1121 14:29:45.217682 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-k28mx" event={"ID":"ae71c0a8-809d-4e15-af09-b3a8b326b5a2","Type":"ContainerStarted","Data":"63dfb6ce5220feb534d3c3db9c1abc8bc6dece4e243e64c8d30ccf0118bef7aa"} Nov 21 14:29:48 crc kubenswrapper[4897]: I1121 14:29:48.232991 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Nov 21 14:29:49 crc kubenswrapper[4897]: I1121 14:29:49.272981 4897 generic.go:334] "Generic (PLEG): container finished" podID="d3691270-398c-46e4-9db9-de8662274fd3" containerID="d9743b6acb206f8ed36f42d844d924a7f8d7953ced2da330e452478112dbd4f5" exitCode=0 Nov 21 14:29:49 crc kubenswrapper[4897]: I1121 14:29:49.273232 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"d3691270-398c-46e4-9db9-de8662274fd3","Type":"ContainerDied","Data":"d9743b6acb206f8ed36f42d844d924a7f8d7953ced2da330e452478112dbd4f5"} Nov 21 14:29:50 crc kubenswrapper[4897]: I1121 14:29:50.069640 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-8f9xc"] Nov 21 14:29:50 crc kubenswrapper[4897]: I1121 14:29:50.151205 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-b42cn"] Nov 21 14:29:50 crc kubenswrapper[4897]: I1121 14:29:50.153230 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb5889db5-b42cn" Nov 21 14:29:50 crc kubenswrapper[4897]: I1121 14:29:50.163972 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-b42cn"] Nov 21 14:29:50 crc kubenswrapper[4897]: I1121 14:29:50.190390 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9306e5a3-58c4-462b-bad5-0e1f876bf832-dns-svc\") pod \"dnsmasq-dns-7cb5889db5-b42cn\" (UID: \"9306e5a3-58c4-462b-bad5-0e1f876bf832\") " pod="openstack/dnsmasq-dns-7cb5889db5-b42cn" Nov 21 14:29:50 crc kubenswrapper[4897]: I1121 14:29:50.190536 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9306e5a3-58c4-462b-bad5-0e1f876bf832-config\") pod \"dnsmasq-dns-7cb5889db5-b42cn\" (UID: \"9306e5a3-58c4-462b-bad5-0e1f876bf832\") " pod="openstack/dnsmasq-dns-7cb5889db5-b42cn" Nov 21 14:29:50 crc kubenswrapper[4897]: I1121 14:29:50.190618 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m9ttz\" (UniqueName: \"kubernetes.io/projected/9306e5a3-58c4-462b-bad5-0e1f876bf832-kube-api-access-m9ttz\") pod \"dnsmasq-dns-7cb5889db5-b42cn\" (UID: \"9306e5a3-58c4-462b-bad5-0e1f876bf832\") " pod="openstack/dnsmasq-dns-7cb5889db5-b42cn" Nov 21 14:29:50 crc kubenswrapper[4897]: I1121 14:29:50.292421 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9306e5a3-58c4-462b-bad5-0e1f876bf832-dns-svc\") pod \"dnsmasq-dns-7cb5889db5-b42cn\" (UID: \"9306e5a3-58c4-462b-bad5-0e1f876bf832\") " pod="openstack/dnsmasq-dns-7cb5889db5-b42cn" Nov 21 14:29:50 crc kubenswrapper[4897]: I1121 14:29:50.292754 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9306e5a3-58c4-462b-bad5-0e1f876bf832-config\") pod \"dnsmasq-dns-7cb5889db5-b42cn\" (UID: \"9306e5a3-58c4-462b-bad5-0e1f876bf832\") " pod="openstack/dnsmasq-dns-7cb5889db5-b42cn" Nov 21 14:29:50 crc kubenswrapper[4897]: I1121 14:29:50.292801 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m9ttz\" (UniqueName: \"kubernetes.io/projected/9306e5a3-58c4-462b-bad5-0e1f876bf832-kube-api-access-m9ttz\") pod \"dnsmasq-dns-7cb5889db5-b42cn\" (UID: \"9306e5a3-58c4-462b-bad5-0e1f876bf832\") " pod="openstack/dnsmasq-dns-7cb5889db5-b42cn" Nov 21 14:29:50 crc kubenswrapper[4897]: I1121 14:29:50.293319 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9306e5a3-58c4-462b-bad5-0e1f876bf832-dns-svc\") pod \"dnsmasq-dns-7cb5889db5-b42cn\" (UID: \"9306e5a3-58c4-462b-bad5-0e1f876bf832\") " pod="openstack/dnsmasq-dns-7cb5889db5-b42cn" Nov 21 14:29:50 crc kubenswrapper[4897]: I1121 14:29:50.293493 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9306e5a3-58c4-462b-bad5-0e1f876bf832-config\") pod \"dnsmasq-dns-7cb5889db5-b42cn\" (UID: \"9306e5a3-58c4-462b-bad5-0e1f876bf832\") " pod="openstack/dnsmasq-dns-7cb5889db5-b42cn" Nov 21 14:29:50 crc kubenswrapper[4897]: I1121 14:29:50.303747 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"d3691270-398c-46e4-9db9-de8662274fd3","Type":"ContainerStarted","Data":"15dc8b84dd86e3bad5f3a95a4b59c8afe7a4f94dfbe440d3263ef192d22cbc57"} Nov 21 14:29:50 crc kubenswrapper[4897]: I1121 14:29:50.326316 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m9ttz\" (UniqueName: \"kubernetes.io/projected/9306e5a3-58c4-462b-bad5-0e1f876bf832-kube-api-access-m9ttz\") pod \"dnsmasq-dns-7cb5889db5-b42cn\" (UID: \"9306e5a3-58c4-462b-bad5-0e1f876bf832\") " pod="openstack/dnsmasq-dns-7cb5889db5-b42cn" Nov 21 14:29:50 crc kubenswrapper[4897]: I1121 14:29:50.338036 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=11.012553999 podStartE2EDuration="44.337986693s" podCreationTimestamp="2025-11-21 14:29:06 +0000 UTC" firstStartedPulling="2025-11-21 14:29:08.716598135 +0000 UTC m=+1226.001191620" lastFinishedPulling="2025-11-21 14:29:42.042030839 +0000 UTC m=+1259.326624314" observedRunningTime="2025-11-21 14:29:50.330272145 +0000 UTC m=+1267.614865640" watchObservedRunningTime="2025-11-21 14:29:50.337986693 +0000 UTC m=+1267.622580168" Nov 21 14:29:50 crc kubenswrapper[4897]: I1121 14:29:50.501914 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-8f9xc" Nov 21 14:29:50 crc kubenswrapper[4897]: I1121 14:29:50.509044 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb5889db5-b42cn" Nov 21 14:29:50 crc kubenswrapper[4897]: I1121 14:29:50.608560 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rqwlp\" (UniqueName: \"kubernetes.io/projected/ecceef5c-b1e5-4c3c-97b3-d805494cfba3-kube-api-access-rqwlp\") pod \"ecceef5c-b1e5-4c3c-97b3-d805494cfba3\" (UID: \"ecceef5c-b1e5-4c3c-97b3-d805494cfba3\") " Nov 21 14:29:50 crc kubenswrapper[4897]: I1121 14:29:50.608709 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ecceef5c-b1e5-4c3c-97b3-d805494cfba3-dns-svc\") pod \"ecceef5c-b1e5-4c3c-97b3-d805494cfba3\" (UID: \"ecceef5c-b1e5-4c3c-97b3-d805494cfba3\") " Nov 21 14:29:50 crc kubenswrapper[4897]: I1121 14:29:50.608736 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ecceef5c-b1e5-4c3c-97b3-d805494cfba3-config\") pod \"ecceef5c-b1e5-4c3c-97b3-d805494cfba3\" (UID: \"ecceef5c-b1e5-4c3c-97b3-d805494cfba3\") " Nov 21 14:29:50 crc kubenswrapper[4897]: I1121 14:29:50.610079 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ecceef5c-b1e5-4c3c-97b3-d805494cfba3-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ecceef5c-b1e5-4c3c-97b3-d805494cfba3" (UID: "ecceef5c-b1e5-4c3c-97b3-d805494cfba3"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:29:50 crc kubenswrapper[4897]: I1121 14:29:50.610368 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ecceef5c-b1e5-4c3c-97b3-d805494cfba3-config" (OuterVolumeSpecName: "config") pod "ecceef5c-b1e5-4c3c-97b3-d805494cfba3" (UID: "ecceef5c-b1e5-4c3c-97b3-d805494cfba3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:29:50 crc kubenswrapper[4897]: I1121 14:29:50.610887 4897 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ecceef5c-b1e5-4c3c-97b3-d805494cfba3-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 21 14:29:50 crc kubenswrapper[4897]: I1121 14:29:50.610908 4897 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ecceef5c-b1e5-4c3c-97b3-d805494cfba3-config\") on node \"crc\" DevicePath \"\"" Nov 21 14:29:50 crc kubenswrapper[4897]: I1121 14:29:50.616857 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ecceef5c-b1e5-4c3c-97b3-d805494cfba3-kube-api-access-rqwlp" (OuterVolumeSpecName: "kube-api-access-rqwlp") pod "ecceef5c-b1e5-4c3c-97b3-d805494cfba3" (UID: "ecceef5c-b1e5-4c3c-97b3-d805494cfba3"). InnerVolumeSpecName "kube-api-access-rqwlp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:29:50 crc kubenswrapper[4897]: I1121 14:29:50.713124 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rqwlp\" (UniqueName: \"kubernetes.io/projected/ecceef5c-b1e5-4c3c-97b3-d805494cfba3-kube-api-access-rqwlp\") on node \"crc\" DevicePath \"\"" Nov 21 14:29:51 crc kubenswrapper[4897]: I1121 14:29:51.087950 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-b42cn"] Nov 21 14:29:51 crc kubenswrapper[4897]: I1121 14:29:51.189120 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Nov 21 14:29:51 crc kubenswrapper[4897]: I1121 14:29:51.197422 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Nov 21 14:29:51 crc kubenswrapper[4897]: I1121 14:29:51.199865 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Nov 21 14:29:51 crc kubenswrapper[4897]: I1121 14:29:51.199879 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Nov 21 14:29:51 crc kubenswrapper[4897]: I1121 14:29:51.199980 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-fkmrw" Nov 21 14:29:51 crc kubenswrapper[4897]: I1121 14:29:51.200773 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Nov 21 14:29:51 crc kubenswrapper[4897]: I1121 14:29:51.207238 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Nov 21 14:29:51 crc kubenswrapper[4897]: I1121 14:29:51.319669 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-mm6tz" event={"ID":"e0c27889-eea7-4a06-9195-3401929cf6b2","Type":"ContainerStarted","Data":"54c79acf28d0e593faf93df3af072d3d67454df8f4dfaf6269db7b62577e64ad"} Nov 21 14:29:51 crc kubenswrapper[4897]: I1121 14:29:51.320076 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-mm6tz" Nov 21 14:29:51 crc kubenswrapper[4897]: I1121 14:29:51.323709 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vkpqf\" (UniqueName: \"kubernetes.io/projected/5f9bf5d1-dcbd-4d19-b46d-22810ebf17a2-kube-api-access-vkpqf\") pod \"swift-storage-0\" (UID: \"5f9bf5d1-dcbd-4d19-b46d-22810ebf17a2\") " pod="openstack/swift-storage-0" Nov 21 14:29:51 crc kubenswrapper[4897]: I1121 14:29:51.323776 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/5f9bf5d1-dcbd-4d19-b46d-22810ebf17a2-cache\") pod \"swift-storage-0\" (UID: \"5f9bf5d1-dcbd-4d19-b46d-22810ebf17a2\") " pod="openstack/swift-storage-0" Nov 21 14:29:51 crc kubenswrapper[4897]: I1121 14:29:51.323792 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/5f9bf5d1-dcbd-4d19-b46d-22810ebf17a2-lock\") pod \"swift-storage-0\" (UID: \"5f9bf5d1-dcbd-4d19-b46d-22810ebf17a2\") " pod="openstack/swift-storage-0" Nov 21 14:29:51 crc kubenswrapper[4897]: I1121 14:29:51.324152 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/5f9bf5d1-dcbd-4d19-b46d-22810ebf17a2-etc-swift\") pod \"swift-storage-0\" (UID: \"5f9bf5d1-dcbd-4d19-b46d-22810ebf17a2\") " pod="openstack/swift-storage-0" Nov 21 14:29:51 crc kubenswrapper[4897]: I1121 14:29:51.324239 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"swift-storage-0\" (UID: \"5f9bf5d1-dcbd-4d19-b46d-22810ebf17a2\") " pod="openstack/swift-storage-0" Nov 21 14:29:51 crc kubenswrapper[4897]: I1121 14:29:51.325175 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-8f9xc" event={"ID":"ecceef5c-b1e5-4c3c-97b3-d805494cfba3","Type":"ContainerDied","Data":"33669f0e7adb0fc32c9cd659ca9c5f306af98c8a6ed1b09cd5baad14c63df0fd"} Nov 21 14:29:51 crc kubenswrapper[4897]: I1121 14:29:51.326149 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-8f9xc" Nov 21 14:29:51 crc kubenswrapper[4897]: I1121 14:29:51.330080 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"8c8fd3ef-d89c-429e-a82b-454894334b2d","Type":"ContainerStarted","Data":"afa0ca6c2aa00dadca2fa057de21ef957986dfe38d17fddf6405cab4d028554a"} Nov 21 14:29:51 crc kubenswrapper[4897]: I1121 14:29:51.334100 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"dabe882c-8166-4930-86ad-e2296a7dbf04","Type":"ContainerStarted","Data":"a3ed6c0acbf784febc45743c3ea701457c2af483da359215053a671d7e00efa5"} Nov 21 14:29:51 crc kubenswrapper[4897]: I1121 14:29:51.337441 4897 generic.go:334] "Generic (PLEG): container finished" podID="ae71c0a8-809d-4e15-af09-b3a8b326b5a2" containerID="44e53b2f96366580a44d52f69781e9e70bdf6edddd27406037fbeb6236553355" exitCode=0 Nov 21 14:29:51 crc kubenswrapper[4897]: I1121 14:29:51.337595 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-k28mx" event={"ID":"ae71c0a8-809d-4e15-af09-b3a8b326b5a2","Type":"ContainerDied","Data":"44e53b2f96366580a44d52f69781e9e70bdf6edddd27406037fbeb6236553355"} Nov 21 14:29:51 crc kubenswrapper[4897]: I1121 14:29:51.350083 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"95ad29e4-b4b6-4cec-b978-23f7ea25edc6","Type":"ContainerStarted","Data":"b82c8eaa87df5cc93b49764cd8875d1ee658727483ebf80e00a24ca9b4bece30"} Nov 21 14:29:51 crc kubenswrapper[4897]: I1121 14:29:51.350708 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Nov 21 14:29:51 crc kubenswrapper[4897]: I1121 14:29:51.352375 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-b42cn" event={"ID":"9306e5a3-58c4-462b-bad5-0e1f876bf832","Type":"ContainerStarted","Data":"5ed5f13738252f6e1b569777b5470c4550fd6da8fc267f2a110323041455fc7d"} Nov 21 14:29:51 crc kubenswrapper[4897]: I1121 14:29:51.354244 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-mt4vk" event={"ID":"6af809de-112c-4988-a06e-4838ef479ace","Type":"ContainerStarted","Data":"a163069b3283b1a9730c2680dad146163f3765fb86b052b4c42201b58240052e"} Nov 21 14:29:51 crc kubenswrapper[4897]: I1121 14:29:51.365004 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-mm6tz" podStartSLOduration=31.293014316 podStartE2EDuration="38.364968839s" podCreationTimestamp="2025-11-21 14:29:13 +0000 UTC" firstStartedPulling="2025-11-21 14:29:42.755003723 +0000 UTC m=+1260.039597208" lastFinishedPulling="2025-11-21 14:29:49.826958256 +0000 UTC m=+1267.111551731" observedRunningTime="2025-11-21 14:29:51.344190703 +0000 UTC m=+1268.628784198" watchObservedRunningTime="2025-11-21 14:29:51.364968839 +0000 UTC m=+1268.649562334" Nov 21 14:29:51 crc kubenswrapper[4897]: I1121 14:29:51.392535 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-84f7c79c46-td7jx" Nov 21 14:29:51 crc kubenswrapper[4897]: I1121 14:29:51.393673 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-84f7c79c46-td7jx" Nov 21 14:29:51 crc kubenswrapper[4897]: I1121 14:29:51.400273 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-mt4vk" podStartSLOduration=34.79680906 podStartE2EDuration="41.400253397s" podCreationTimestamp="2025-11-21 14:29:10 +0000 UTC" firstStartedPulling="2025-11-21 14:29:42.775088971 +0000 UTC m=+1260.059682466" lastFinishedPulling="2025-11-21 14:29:49.378533328 +0000 UTC m=+1266.663126803" observedRunningTime="2025-11-21 14:29:51.380390063 +0000 UTC m=+1268.664983538" watchObservedRunningTime="2025-11-21 14:29:51.400253397 +0000 UTC m=+1268.684846862" Nov 21 14:29:51 crc kubenswrapper[4897]: I1121 14:29:51.414170 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-84f7c79c46-td7jx" Nov 21 14:29:51 crc kubenswrapper[4897]: I1121 14:29:51.426590 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/5f9bf5d1-dcbd-4d19-b46d-22810ebf17a2-etc-swift\") pod \"swift-storage-0\" (UID: \"5f9bf5d1-dcbd-4d19-b46d-22810ebf17a2\") " pod="openstack/swift-storage-0" Nov 21 14:29:51 crc kubenswrapper[4897]: I1121 14:29:51.426732 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"swift-storage-0\" (UID: \"5f9bf5d1-dcbd-4d19-b46d-22810ebf17a2\") " pod="openstack/swift-storage-0" Nov 21 14:29:51 crc kubenswrapper[4897]: I1121 14:29:51.426857 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vkpqf\" (UniqueName: \"kubernetes.io/projected/5f9bf5d1-dcbd-4d19-b46d-22810ebf17a2-kube-api-access-vkpqf\") pod \"swift-storage-0\" (UID: \"5f9bf5d1-dcbd-4d19-b46d-22810ebf17a2\") " pod="openstack/swift-storage-0" Nov 21 14:29:51 crc kubenswrapper[4897]: I1121 14:29:51.426947 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/5f9bf5d1-dcbd-4d19-b46d-22810ebf17a2-cache\") pod \"swift-storage-0\" (UID: \"5f9bf5d1-dcbd-4d19-b46d-22810ebf17a2\") " pod="openstack/swift-storage-0" Nov 21 14:29:51 crc kubenswrapper[4897]: I1121 14:29:51.426969 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/5f9bf5d1-dcbd-4d19-b46d-22810ebf17a2-lock\") pod \"swift-storage-0\" (UID: \"5f9bf5d1-dcbd-4d19-b46d-22810ebf17a2\") " pod="openstack/swift-storage-0" Nov 21 14:29:51 crc kubenswrapper[4897]: I1121 14:29:51.428696 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=34.791138485 podStartE2EDuration="42.428681109s" podCreationTimestamp="2025-11-21 14:29:09 +0000 UTC" firstStartedPulling="2025-11-21 14:29:42.191292642 +0000 UTC m=+1259.475886117" lastFinishedPulling="2025-11-21 14:29:49.828835266 +0000 UTC m=+1267.113428741" observedRunningTime="2025-11-21 14:29:51.400619456 +0000 UTC m=+1268.685212961" watchObservedRunningTime="2025-11-21 14:29:51.428681109 +0000 UTC m=+1268.713274594" Nov 21 14:29:51 crc kubenswrapper[4897]: I1121 14:29:51.433422 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/5f9bf5d1-dcbd-4d19-b46d-22810ebf17a2-lock\") pod \"swift-storage-0\" (UID: \"5f9bf5d1-dcbd-4d19-b46d-22810ebf17a2\") " pod="openstack/swift-storage-0" Nov 21 14:29:51 crc kubenswrapper[4897]: E1121 14:29:51.433620 4897 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 21 14:29:51 crc kubenswrapper[4897]: E1121 14:29:51.433645 4897 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 21 14:29:51 crc kubenswrapper[4897]: E1121 14:29:51.433687 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/5f9bf5d1-dcbd-4d19-b46d-22810ebf17a2-etc-swift podName:5f9bf5d1-dcbd-4d19-b46d-22810ebf17a2 nodeName:}" failed. No retries permitted until 2025-11-21 14:29:51.933669843 +0000 UTC m=+1269.218263318 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/5f9bf5d1-dcbd-4d19-b46d-22810ebf17a2-etc-swift") pod "swift-storage-0" (UID: "5f9bf5d1-dcbd-4d19-b46d-22810ebf17a2") : configmap "swift-ring-files" not found Nov 21 14:29:51 crc kubenswrapper[4897]: I1121 14:29:51.435495 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/5f9bf5d1-dcbd-4d19-b46d-22810ebf17a2-cache\") pod \"swift-storage-0\" (UID: \"5f9bf5d1-dcbd-4d19-b46d-22810ebf17a2\") " pod="openstack/swift-storage-0" Nov 21 14:29:51 crc kubenswrapper[4897]: I1121 14:29:51.436811 4897 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"swift-storage-0\" (UID: \"5f9bf5d1-dcbd-4d19-b46d-22810ebf17a2\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/swift-storage-0" Nov 21 14:29:51 crc kubenswrapper[4897]: I1121 14:29:51.456776 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-8f9xc"] Nov 21 14:29:51 crc kubenswrapper[4897]: I1121 14:29:51.458697 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vkpqf\" (UniqueName: \"kubernetes.io/projected/5f9bf5d1-dcbd-4d19-b46d-22810ebf17a2-kube-api-access-vkpqf\") pod \"swift-storage-0\" (UID: \"5f9bf5d1-dcbd-4d19-b46d-22810ebf17a2\") " pod="openstack/swift-storage-0" Nov 21 14:29:51 crc kubenswrapper[4897]: I1121 14:29:51.467588 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-8f9xc"] Nov 21 14:29:51 crc kubenswrapper[4897]: I1121 14:29:51.736546 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"swift-storage-0\" (UID: \"5f9bf5d1-dcbd-4d19-b46d-22810ebf17a2\") " pod="openstack/swift-storage-0" Nov 21 14:29:51 crc kubenswrapper[4897]: I1121 14:29:51.937051 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/5f9bf5d1-dcbd-4d19-b46d-22810ebf17a2-etc-swift\") pod \"swift-storage-0\" (UID: \"5f9bf5d1-dcbd-4d19-b46d-22810ebf17a2\") " pod="openstack/swift-storage-0" Nov 21 14:29:51 crc kubenswrapper[4897]: E1121 14:29:51.937292 4897 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 21 14:29:51 crc kubenswrapper[4897]: E1121 14:29:51.937336 4897 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 21 14:29:51 crc kubenswrapper[4897]: E1121 14:29:51.937404 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/5f9bf5d1-dcbd-4d19-b46d-22810ebf17a2-etc-swift podName:5f9bf5d1-dcbd-4d19-b46d-22810ebf17a2 nodeName:}" failed. No retries permitted until 2025-11-21 14:29:52.937381254 +0000 UTC m=+1270.221974729 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/5f9bf5d1-dcbd-4d19-b46d-22810ebf17a2-etc-swift") pod "swift-storage-0" (UID: "5f9bf5d1-dcbd-4d19-b46d-22810ebf17a2") : configmap "swift-ring-files" not found Nov 21 14:29:52 crc kubenswrapper[4897]: I1121 14:29:52.102870 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ecceef5c-b1e5-4c3c-97b3-d805494cfba3" path="/var/lib/kubelet/pods/ecceef5c-b1e5-4c3c-97b3-d805494cfba3/volumes" Nov 21 14:29:52 crc kubenswrapper[4897]: I1121 14:29:52.375982 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-84f7c79c46-td7jx" Nov 21 14:29:52 crc kubenswrapper[4897]: I1121 14:29:52.455195 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f58d5cd65-2pc2p"] Nov 21 14:29:52 crc kubenswrapper[4897]: I1121 14:29:52.957549 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/5f9bf5d1-dcbd-4d19-b46d-22810ebf17a2-etc-swift\") pod \"swift-storage-0\" (UID: \"5f9bf5d1-dcbd-4d19-b46d-22810ebf17a2\") " pod="openstack/swift-storage-0" Nov 21 14:29:52 crc kubenswrapper[4897]: E1121 14:29:52.957817 4897 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 21 14:29:52 crc kubenswrapper[4897]: E1121 14:29:52.958127 4897 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 21 14:29:52 crc kubenswrapper[4897]: E1121 14:29:52.958187 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/5f9bf5d1-dcbd-4d19-b46d-22810ebf17a2-etc-swift podName:5f9bf5d1-dcbd-4d19-b46d-22810ebf17a2 nodeName:}" failed. No retries permitted until 2025-11-21 14:29:54.958167784 +0000 UTC m=+1272.242761269 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/5f9bf5d1-dcbd-4d19-b46d-22810ebf17a2-etc-swift") pod "swift-storage-0" (UID: "5f9bf5d1-dcbd-4d19-b46d-22810ebf17a2") : configmap "swift-ring-files" not found Nov 21 14:29:53 crc kubenswrapper[4897]: I1121 14:29:53.381168 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"b513e4ee-7c30-4942-b997-4d52414e1ec0","Type":"ContainerStarted","Data":"3140f08ac469f99b17c2aa56cad1a0ccf3db6d5b4c5110f9454241ae6294b460"} Nov 21 14:29:53 crc kubenswrapper[4897]: I1121 14:29:53.383406 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-k28mx" event={"ID":"ae71c0a8-809d-4e15-af09-b3a8b326b5a2","Type":"ContainerStarted","Data":"03fca8f89d1e4b92fe01a06f2ac7498cae13a8fde828ec089580725a2c429bb1"} Nov 21 14:29:53 crc kubenswrapper[4897]: I1121 14:29:53.386153 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"91aa2270-92a7-4af4-9a92-141ea2b24ee1","Type":"ContainerStarted","Data":"ae9d2f28e206da418bad9d39e2d6c5076e3d741d413ec6b6cb6c3b13dfa61274"} Nov 21 14:29:53 crc kubenswrapper[4897]: I1121 14:29:53.388002 4897 generic.go:334] "Generic (PLEG): container finished" podID="9306e5a3-58c4-462b-bad5-0e1f876bf832" containerID="206ce3c58802a46d150063b061c85011963c120a26ec12245f2b3374b87a68ab" exitCode=0 Nov 21 14:29:53 crc kubenswrapper[4897]: I1121 14:29:53.388349 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-b42cn" event={"ID":"9306e5a3-58c4-462b-bad5-0e1f876bf832","Type":"ContainerDied","Data":"206ce3c58802a46d150063b061c85011963c120a26ec12245f2b3374b87a68ab"} Nov 21 14:29:53 crc kubenswrapper[4897]: I1121 14:29:53.390068 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"fce5a726-bd06-473b-a9d0-f812ffe83693","Type":"ContainerStarted","Data":"8a7ca999bf42855f24b3ac03dac8be32e2bdc1d5214ae873b55de75611e26376"} Nov 21 14:29:55 crc kubenswrapper[4897]: I1121 14:29:55.023656 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/5f9bf5d1-dcbd-4d19-b46d-22810ebf17a2-etc-swift\") pod \"swift-storage-0\" (UID: \"5f9bf5d1-dcbd-4d19-b46d-22810ebf17a2\") " pod="openstack/swift-storage-0" Nov 21 14:29:55 crc kubenswrapper[4897]: E1121 14:29:55.023843 4897 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 21 14:29:55 crc kubenswrapper[4897]: E1121 14:29:55.024212 4897 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 21 14:29:55 crc kubenswrapper[4897]: E1121 14:29:55.024259 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/5f9bf5d1-dcbd-4d19-b46d-22810ebf17a2-etc-swift podName:5f9bf5d1-dcbd-4d19-b46d-22810ebf17a2 nodeName:}" failed. No retries permitted until 2025-11-21 14:29:59.024244154 +0000 UTC m=+1276.308837629 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/5f9bf5d1-dcbd-4d19-b46d-22810ebf17a2-etc-swift") pod "swift-storage-0" (UID: "5f9bf5d1-dcbd-4d19-b46d-22810ebf17a2") : configmap "swift-ring-files" not found Nov 21 14:29:55 crc kubenswrapper[4897]: I1121 14:29:55.141441 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-t9nzd"] Nov 21 14:29:55 crc kubenswrapper[4897]: I1121 14:29:55.144690 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-t9nzd" Nov 21 14:29:55 crc kubenswrapper[4897]: I1121 14:29:55.152897 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Nov 21 14:29:55 crc kubenswrapper[4897]: I1121 14:29:55.153910 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Nov 21 14:29:55 crc kubenswrapper[4897]: I1121 14:29:55.153922 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Nov 21 14:29:55 crc kubenswrapper[4897]: I1121 14:29:55.157708 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-t9nzd"] Nov 21 14:29:55 crc kubenswrapper[4897]: I1121 14:29:55.227334 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6d1f75d9-2eaf-44c1-b2d0-dae3a8547433-scripts\") pod \"swift-ring-rebalance-t9nzd\" (UID: \"6d1f75d9-2eaf-44c1-b2d0-dae3a8547433\") " pod="openstack/swift-ring-rebalance-t9nzd" Nov 21 14:29:55 crc kubenswrapper[4897]: I1121 14:29:55.227455 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/6d1f75d9-2eaf-44c1-b2d0-dae3a8547433-dispersionconf\") pod \"swift-ring-rebalance-t9nzd\" (UID: \"6d1f75d9-2eaf-44c1-b2d0-dae3a8547433\") " pod="openstack/swift-ring-rebalance-t9nzd" Nov 21 14:29:55 crc kubenswrapper[4897]: I1121 14:29:55.227493 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/6d1f75d9-2eaf-44c1-b2d0-dae3a8547433-ring-data-devices\") pod \"swift-ring-rebalance-t9nzd\" (UID: \"6d1f75d9-2eaf-44c1-b2d0-dae3a8547433\") " pod="openstack/swift-ring-rebalance-t9nzd" Nov 21 14:29:55 crc kubenswrapper[4897]: I1121 14:29:55.227916 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d1f75d9-2eaf-44c1-b2d0-dae3a8547433-combined-ca-bundle\") pod \"swift-ring-rebalance-t9nzd\" (UID: \"6d1f75d9-2eaf-44c1-b2d0-dae3a8547433\") " pod="openstack/swift-ring-rebalance-t9nzd" Nov 21 14:29:55 crc kubenswrapper[4897]: I1121 14:29:55.227983 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z9584\" (UniqueName: \"kubernetes.io/projected/6d1f75d9-2eaf-44c1-b2d0-dae3a8547433-kube-api-access-z9584\") pod \"swift-ring-rebalance-t9nzd\" (UID: \"6d1f75d9-2eaf-44c1-b2d0-dae3a8547433\") " pod="openstack/swift-ring-rebalance-t9nzd" Nov 21 14:29:55 crc kubenswrapper[4897]: I1121 14:29:55.228055 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/6d1f75d9-2eaf-44c1-b2d0-dae3a8547433-swiftconf\") pod \"swift-ring-rebalance-t9nzd\" (UID: \"6d1f75d9-2eaf-44c1-b2d0-dae3a8547433\") " pod="openstack/swift-ring-rebalance-t9nzd" Nov 21 14:29:55 crc kubenswrapper[4897]: I1121 14:29:55.228166 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/6d1f75d9-2eaf-44c1-b2d0-dae3a8547433-etc-swift\") pod \"swift-ring-rebalance-t9nzd\" (UID: \"6d1f75d9-2eaf-44c1-b2d0-dae3a8547433\") " pod="openstack/swift-ring-rebalance-t9nzd" Nov 21 14:29:55 crc kubenswrapper[4897]: I1121 14:29:55.330976 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d1f75d9-2eaf-44c1-b2d0-dae3a8547433-combined-ca-bundle\") pod \"swift-ring-rebalance-t9nzd\" (UID: \"6d1f75d9-2eaf-44c1-b2d0-dae3a8547433\") " pod="openstack/swift-ring-rebalance-t9nzd" Nov 21 14:29:55 crc kubenswrapper[4897]: I1121 14:29:55.331048 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z9584\" (UniqueName: \"kubernetes.io/projected/6d1f75d9-2eaf-44c1-b2d0-dae3a8547433-kube-api-access-z9584\") pod \"swift-ring-rebalance-t9nzd\" (UID: \"6d1f75d9-2eaf-44c1-b2d0-dae3a8547433\") " pod="openstack/swift-ring-rebalance-t9nzd" Nov 21 14:29:55 crc kubenswrapper[4897]: I1121 14:29:55.331124 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/6d1f75d9-2eaf-44c1-b2d0-dae3a8547433-swiftconf\") pod \"swift-ring-rebalance-t9nzd\" (UID: \"6d1f75d9-2eaf-44c1-b2d0-dae3a8547433\") " pod="openstack/swift-ring-rebalance-t9nzd" Nov 21 14:29:55 crc kubenswrapper[4897]: I1121 14:29:55.331169 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/6d1f75d9-2eaf-44c1-b2d0-dae3a8547433-etc-swift\") pod \"swift-ring-rebalance-t9nzd\" (UID: \"6d1f75d9-2eaf-44c1-b2d0-dae3a8547433\") " pod="openstack/swift-ring-rebalance-t9nzd" Nov 21 14:29:55 crc kubenswrapper[4897]: I1121 14:29:55.331272 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6d1f75d9-2eaf-44c1-b2d0-dae3a8547433-scripts\") pod \"swift-ring-rebalance-t9nzd\" (UID: \"6d1f75d9-2eaf-44c1-b2d0-dae3a8547433\") " pod="openstack/swift-ring-rebalance-t9nzd" Nov 21 14:29:55 crc kubenswrapper[4897]: I1121 14:29:55.331372 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/6d1f75d9-2eaf-44c1-b2d0-dae3a8547433-dispersionconf\") pod \"swift-ring-rebalance-t9nzd\" (UID: \"6d1f75d9-2eaf-44c1-b2d0-dae3a8547433\") " pod="openstack/swift-ring-rebalance-t9nzd" Nov 21 14:29:55 crc kubenswrapper[4897]: I1121 14:29:55.331420 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/6d1f75d9-2eaf-44c1-b2d0-dae3a8547433-ring-data-devices\") pod \"swift-ring-rebalance-t9nzd\" (UID: \"6d1f75d9-2eaf-44c1-b2d0-dae3a8547433\") " pod="openstack/swift-ring-rebalance-t9nzd" Nov 21 14:29:55 crc kubenswrapper[4897]: I1121 14:29:55.332329 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/6d1f75d9-2eaf-44c1-b2d0-dae3a8547433-ring-data-devices\") pod \"swift-ring-rebalance-t9nzd\" (UID: \"6d1f75d9-2eaf-44c1-b2d0-dae3a8547433\") " pod="openstack/swift-ring-rebalance-t9nzd" Nov 21 14:29:55 crc kubenswrapper[4897]: I1121 14:29:55.333630 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6d1f75d9-2eaf-44c1-b2d0-dae3a8547433-scripts\") pod \"swift-ring-rebalance-t9nzd\" (UID: \"6d1f75d9-2eaf-44c1-b2d0-dae3a8547433\") " pod="openstack/swift-ring-rebalance-t9nzd" Nov 21 14:29:55 crc kubenswrapper[4897]: I1121 14:29:55.333888 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/6d1f75d9-2eaf-44c1-b2d0-dae3a8547433-etc-swift\") pod \"swift-ring-rebalance-t9nzd\" (UID: \"6d1f75d9-2eaf-44c1-b2d0-dae3a8547433\") " pod="openstack/swift-ring-rebalance-t9nzd" Nov 21 14:29:55 crc kubenswrapper[4897]: I1121 14:29:55.341718 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/6d1f75d9-2eaf-44c1-b2d0-dae3a8547433-swiftconf\") pod \"swift-ring-rebalance-t9nzd\" (UID: \"6d1f75d9-2eaf-44c1-b2d0-dae3a8547433\") " pod="openstack/swift-ring-rebalance-t9nzd" Nov 21 14:29:55 crc kubenswrapper[4897]: I1121 14:29:55.343108 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d1f75d9-2eaf-44c1-b2d0-dae3a8547433-combined-ca-bundle\") pod \"swift-ring-rebalance-t9nzd\" (UID: \"6d1f75d9-2eaf-44c1-b2d0-dae3a8547433\") " pod="openstack/swift-ring-rebalance-t9nzd" Nov 21 14:29:55 crc kubenswrapper[4897]: I1121 14:29:55.343336 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/6d1f75d9-2eaf-44c1-b2d0-dae3a8547433-dispersionconf\") pod \"swift-ring-rebalance-t9nzd\" (UID: \"6d1f75d9-2eaf-44c1-b2d0-dae3a8547433\") " pod="openstack/swift-ring-rebalance-t9nzd" Nov 21 14:29:55 crc kubenswrapper[4897]: I1121 14:29:55.356526 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z9584\" (UniqueName: \"kubernetes.io/projected/6d1f75d9-2eaf-44c1-b2d0-dae3a8547433-kube-api-access-z9584\") pod \"swift-ring-rebalance-t9nzd\" (UID: \"6d1f75d9-2eaf-44c1-b2d0-dae3a8547433\") " pod="openstack/swift-ring-rebalance-t9nzd" Nov 21 14:29:55 crc kubenswrapper[4897]: I1121 14:29:55.415687 4897 generic.go:334] "Generic (PLEG): container finished" podID="6bf04b77-0ef9-42f0-8a02-8c4c7568e390" containerID="6b2b520616824e276cba653c73af285909212743e4ced43cb4bd3f06fdcd21a9" exitCode=0 Nov 21 14:29:55 crc kubenswrapper[4897]: I1121 14:29:55.415873 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-khkvq" event={"ID":"6bf04b77-0ef9-42f0-8a02-8c4c7568e390","Type":"ContainerDied","Data":"6b2b520616824e276cba653c73af285909212743e4ced43cb4bd3f06fdcd21a9"} Nov 21 14:29:55 crc kubenswrapper[4897]: I1121 14:29:55.419295 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-k28mx" event={"ID":"ae71c0a8-809d-4e15-af09-b3a8b326b5a2","Type":"ContainerStarted","Data":"6dd012f44c162b5980c2608be4d8079d2fd30ea07f41a4d6c43745a95ea04b7b"} Nov 21 14:29:55 crc kubenswrapper[4897]: I1121 14:29:55.419355 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-k28mx" Nov 21 14:29:55 crc kubenswrapper[4897]: I1121 14:29:55.419372 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-k28mx" Nov 21 14:29:55 crc kubenswrapper[4897]: I1121 14:29:55.423722 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-b42cn" event={"ID":"9306e5a3-58c4-462b-bad5-0e1f876bf832","Type":"ContainerStarted","Data":"0eca3681977bafa554467e45117ee7088cec805d162c8912d1c4a66c201f62f4"} Nov 21 14:29:55 crc kubenswrapper[4897]: I1121 14:29:55.424366 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7cb5889db5-b42cn" Nov 21 14:29:55 crc kubenswrapper[4897]: I1121 14:29:55.466373 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-k28mx" podStartSLOduration=37.090913107 podStartE2EDuration="42.466352653s" podCreationTimestamp="2025-11-21 14:29:13 +0000 UTC" firstStartedPulling="2025-11-21 14:29:44.447294186 +0000 UTC m=+1261.731887671" lastFinishedPulling="2025-11-21 14:29:49.822733742 +0000 UTC m=+1267.107327217" observedRunningTime="2025-11-21 14:29:55.46622952 +0000 UTC m=+1272.750823015" watchObservedRunningTime="2025-11-21 14:29:55.466352653 +0000 UTC m=+1272.750946128" Nov 21 14:29:55 crc kubenswrapper[4897]: I1121 14:29:55.468197 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-t9nzd" Nov 21 14:29:55 crc kubenswrapper[4897]: I1121 14:29:55.492930 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7cb5889db5-b42cn" podStartSLOduration=5.064270938 podStartE2EDuration="5.492913895s" podCreationTimestamp="2025-11-21 14:29:50 +0000 UTC" firstStartedPulling="2025-11-21 14:29:51.093039756 +0000 UTC m=+1268.377633221" lastFinishedPulling="2025-11-21 14:29:51.521682703 +0000 UTC m=+1268.806276178" observedRunningTime="2025-11-21 14:29:55.483972375 +0000 UTC m=+1272.768565850" watchObservedRunningTime="2025-11-21 14:29:55.492913895 +0000 UTC m=+1272.777507370" Nov 21 14:29:55 crc kubenswrapper[4897]: I1121 14:29:55.991206 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-t9nzd"] Nov 21 14:29:56 crc kubenswrapper[4897]: I1121 14:29:56.442734 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-khkvq" event={"ID":"6bf04b77-0ef9-42f0-8a02-8c4c7568e390","Type":"ContainerStarted","Data":"b06105b5b07e06d646b4c304d1f59865380586cfc30e6b169b23c6dfa177f20a"} Nov 21 14:29:56 crc kubenswrapper[4897]: I1121 14:29:56.442948 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-666b6646f7-khkvq" Nov 21 14:29:56 crc kubenswrapper[4897]: I1121 14:29:56.453358 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"8c8fd3ef-d89c-429e-a82b-454894334b2d","Type":"ContainerStarted","Data":"86f6bdbf58f427ec3ae137327f9c5b8fa5901f30962e3487d359321adc9c18bc"} Nov 21 14:29:56 crc kubenswrapper[4897]: I1121 14:29:56.477467 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-666b6646f7-khkvq" podStartSLOduration=-9223371983.37733 podStartE2EDuration="53.477444333s" podCreationTimestamp="2025-11-21 14:29:03 +0000 UTC" firstStartedPulling="2025-11-21 14:29:04.548640171 +0000 UTC m=+1221.833233636" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:29:56.458051493 +0000 UTC m=+1273.742644998" watchObservedRunningTime="2025-11-21 14:29:56.477444333 +0000 UTC m=+1273.762037828" Nov 21 14:29:56 crc kubenswrapper[4897]: I1121 14:29:56.481150 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"dabe882c-8166-4930-86ad-e2296a7dbf04","Type":"ContainerStarted","Data":"72472a68a5b88dc6f62941d6c3ddfbf1c09f7831ca7bf44bd32ecdc34d9d6076"} Nov 21 14:29:56 crc kubenswrapper[4897]: I1121 14:29:56.483220 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Nov 21 14:29:56 crc kubenswrapper[4897]: I1121 14:29:56.485743 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Nov 21 14:29:56 crc kubenswrapper[4897]: I1121 14:29:56.486200 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=28.536708783 podStartE2EDuration="40.486173117s" podCreationTimestamp="2025-11-21 14:29:16 +0000 UTC" firstStartedPulling="2025-11-21 14:29:43.496150952 +0000 UTC m=+1260.780744427" lastFinishedPulling="2025-11-21 14:29:55.445615296 +0000 UTC m=+1272.730208761" observedRunningTime="2025-11-21 14:29:56.482721885 +0000 UTC m=+1273.767315370" watchObservedRunningTime="2025-11-21 14:29:56.486173117 +0000 UTC m=+1273.770766602" Nov 21 14:29:56 crc kubenswrapper[4897]: I1121 14:29:56.511966 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"bd692188-6da6-4387-a46d-003a2da0d0c8","Type":"ContainerStarted","Data":"5e97248a8dc790521af481c0714596ab753484b35fa256f425d03e7bac6e870a"} Nov 21 14:29:56 crc kubenswrapper[4897]: I1121 14:29:56.520953 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-t9nzd" event={"ID":"6d1f75d9-2eaf-44c1-b2d0-dae3a8547433","Type":"ContainerStarted","Data":"a73917e008195dc7f69097b0ca104e56ee86b44a4ae10413b142f977531470f8"} Nov 21 14:29:56 crc kubenswrapper[4897]: I1121 14:29:56.527862 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=30.990151949 podStartE2EDuration="41.527840925s" podCreationTimestamp="2025-11-21 14:29:15 +0000 UTC" firstStartedPulling="2025-11-21 14:29:44.949417254 +0000 UTC m=+1262.234010729" lastFinishedPulling="2025-11-21 14:29:55.48710623 +0000 UTC m=+1272.771699705" observedRunningTime="2025-11-21 14:29:56.52205129 +0000 UTC m=+1273.806644765" watchObservedRunningTime="2025-11-21 14:29:56.527840925 +0000 UTC m=+1273.812434400" Nov 21 14:29:56 crc kubenswrapper[4897]: I1121 14:29:56.582781 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Nov 21 14:29:56 crc kubenswrapper[4897]: I1121 14:29:56.804565 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-p7hgh"] Nov 21 14:29:56 crc kubenswrapper[4897]: I1121 14:29:56.806340 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-p7hgh" Nov 21 14:29:56 crc kubenswrapper[4897]: I1121 14:29:56.808463 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Nov 21 14:29:56 crc kubenswrapper[4897]: I1121 14:29:56.817311 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-p7hgh"] Nov 21 14:29:56 crc kubenswrapper[4897]: I1121 14:29:56.876831 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5724k\" (UniqueName: \"kubernetes.io/projected/15ab9acd-00ef-4b25-8c82-b86e95ea046f-kube-api-access-5724k\") pod \"ovn-controller-metrics-p7hgh\" (UID: \"15ab9acd-00ef-4b25-8c82-b86e95ea046f\") " pod="openstack/ovn-controller-metrics-p7hgh" Nov 21 14:29:56 crc kubenswrapper[4897]: I1121 14:29:56.877049 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/15ab9acd-00ef-4b25-8c82-b86e95ea046f-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-p7hgh\" (UID: \"15ab9acd-00ef-4b25-8c82-b86e95ea046f\") " pod="openstack/ovn-controller-metrics-p7hgh" Nov 21 14:29:56 crc kubenswrapper[4897]: I1121 14:29:56.877088 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/15ab9acd-00ef-4b25-8c82-b86e95ea046f-config\") pod \"ovn-controller-metrics-p7hgh\" (UID: \"15ab9acd-00ef-4b25-8c82-b86e95ea046f\") " pod="openstack/ovn-controller-metrics-p7hgh" Nov 21 14:29:56 crc kubenswrapper[4897]: I1121 14:29:56.877148 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/15ab9acd-00ef-4b25-8c82-b86e95ea046f-ovn-rundir\") pod \"ovn-controller-metrics-p7hgh\" (UID: \"15ab9acd-00ef-4b25-8c82-b86e95ea046f\") " pod="openstack/ovn-controller-metrics-p7hgh" Nov 21 14:29:56 crc kubenswrapper[4897]: I1121 14:29:56.877174 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15ab9acd-00ef-4b25-8c82-b86e95ea046f-combined-ca-bundle\") pod \"ovn-controller-metrics-p7hgh\" (UID: \"15ab9acd-00ef-4b25-8c82-b86e95ea046f\") " pod="openstack/ovn-controller-metrics-p7hgh" Nov 21 14:29:56 crc kubenswrapper[4897]: I1121 14:29:56.877229 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/15ab9acd-00ef-4b25-8c82-b86e95ea046f-ovs-rundir\") pod \"ovn-controller-metrics-p7hgh\" (UID: \"15ab9acd-00ef-4b25-8c82-b86e95ea046f\") " pod="openstack/ovn-controller-metrics-p7hgh" Nov 21 14:29:56 crc kubenswrapper[4897]: I1121 14:29:56.951425 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-khkvq"] Nov 21 14:29:56 crc kubenswrapper[4897]: I1121 14:29:56.980610 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d65f699f-cf7rm"] Nov 21 14:29:56 crc kubenswrapper[4897]: I1121 14:29:56.981718 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5724k\" (UniqueName: \"kubernetes.io/projected/15ab9acd-00ef-4b25-8c82-b86e95ea046f-kube-api-access-5724k\") pod \"ovn-controller-metrics-p7hgh\" (UID: \"15ab9acd-00ef-4b25-8c82-b86e95ea046f\") " pod="openstack/ovn-controller-metrics-p7hgh" Nov 21 14:29:56 crc kubenswrapper[4897]: I1121 14:29:56.982164 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/15ab9acd-00ef-4b25-8c82-b86e95ea046f-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-p7hgh\" (UID: \"15ab9acd-00ef-4b25-8c82-b86e95ea046f\") " pod="openstack/ovn-controller-metrics-p7hgh" Nov 21 14:29:56 crc kubenswrapper[4897]: I1121 14:29:56.982191 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/15ab9acd-00ef-4b25-8c82-b86e95ea046f-config\") pod \"ovn-controller-metrics-p7hgh\" (UID: \"15ab9acd-00ef-4b25-8c82-b86e95ea046f\") " pod="openstack/ovn-controller-metrics-p7hgh" Nov 21 14:29:56 crc kubenswrapper[4897]: I1121 14:29:56.982257 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/15ab9acd-00ef-4b25-8c82-b86e95ea046f-ovn-rundir\") pod \"ovn-controller-metrics-p7hgh\" (UID: \"15ab9acd-00ef-4b25-8c82-b86e95ea046f\") " pod="openstack/ovn-controller-metrics-p7hgh" Nov 21 14:29:56 crc kubenswrapper[4897]: I1121 14:29:56.982276 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15ab9acd-00ef-4b25-8c82-b86e95ea046f-combined-ca-bundle\") pod \"ovn-controller-metrics-p7hgh\" (UID: \"15ab9acd-00ef-4b25-8c82-b86e95ea046f\") " pod="openstack/ovn-controller-metrics-p7hgh" Nov 21 14:29:56 crc kubenswrapper[4897]: I1121 14:29:56.982316 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/15ab9acd-00ef-4b25-8c82-b86e95ea046f-ovs-rundir\") pod \"ovn-controller-metrics-p7hgh\" (UID: \"15ab9acd-00ef-4b25-8c82-b86e95ea046f\") " pod="openstack/ovn-controller-metrics-p7hgh" Nov 21 14:29:56 crc kubenswrapper[4897]: I1121 14:29:56.982656 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/15ab9acd-00ef-4b25-8c82-b86e95ea046f-ovs-rundir\") pod \"ovn-controller-metrics-p7hgh\" (UID: \"15ab9acd-00ef-4b25-8c82-b86e95ea046f\") " pod="openstack/ovn-controller-metrics-p7hgh" Nov 21 14:29:56 crc kubenswrapper[4897]: I1121 14:29:56.982874 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d65f699f-cf7rm" Nov 21 14:29:56 crc kubenswrapper[4897]: I1121 14:29:56.983726 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/15ab9acd-00ef-4b25-8c82-b86e95ea046f-ovn-rundir\") pod \"ovn-controller-metrics-p7hgh\" (UID: \"15ab9acd-00ef-4b25-8c82-b86e95ea046f\") " pod="openstack/ovn-controller-metrics-p7hgh" Nov 21 14:29:56 crc kubenswrapper[4897]: I1121 14:29:56.984396 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/15ab9acd-00ef-4b25-8c82-b86e95ea046f-config\") pod \"ovn-controller-metrics-p7hgh\" (UID: \"15ab9acd-00ef-4b25-8c82-b86e95ea046f\") " pod="openstack/ovn-controller-metrics-p7hgh" Nov 21 14:29:56 crc kubenswrapper[4897]: I1121 14:29:56.985872 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Nov 21 14:29:56 crc kubenswrapper[4897]: I1121 14:29:56.995282 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d65f699f-cf7rm"] Nov 21 14:29:56 crc kubenswrapper[4897]: I1121 14:29:56.996422 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15ab9acd-00ef-4b25-8c82-b86e95ea046f-combined-ca-bundle\") pod \"ovn-controller-metrics-p7hgh\" (UID: \"15ab9acd-00ef-4b25-8c82-b86e95ea046f\") " pod="openstack/ovn-controller-metrics-p7hgh" Nov 21 14:29:57 crc kubenswrapper[4897]: I1121 14:29:57.006491 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/15ab9acd-00ef-4b25-8c82-b86e95ea046f-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-p7hgh\" (UID: \"15ab9acd-00ef-4b25-8c82-b86e95ea046f\") " pod="openstack/ovn-controller-metrics-p7hgh" Nov 21 14:29:57 crc kubenswrapper[4897]: I1121 14:29:57.009883 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5724k\" (UniqueName: \"kubernetes.io/projected/15ab9acd-00ef-4b25-8c82-b86e95ea046f-kube-api-access-5724k\") pod \"ovn-controller-metrics-p7hgh\" (UID: \"15ab9acd-00ef-4b25-8c82-b86e95ea046f\") " pod="openstack/ovn-controller-metrics-p7hgh" Nov 21 14:29:57 crc kubenswrapper[4897]: I1121 14:29:57.083773 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vpm8r\" (UniqueName: \"kubernetes.io/projected/56092cc0-5e4c-47bb-98c1-c1c82d89e0a6-kube-api-access-vpm8r\") pod \"dnsmasq-dns-57d65f699f-cf7rm\" (UID: \"56092cc0-5e4c-47bb-98c1-c1c82d89e0a6\") " pod="openstack/dnsmasq-dns-57d65f699f-cf7rm" Nov 21 14:29:57 crc kubenswrapper[4897]: I1121 14:29:57.083913 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/56092cc0-5e4c-47bb-98c1-c1c82d89e0a6-ovsdbserver-nb\") pod \"dnsmasq-dns-57d65f699f-cf7rm\" (UID: \"56092cc0-5e4c-47bb-98c1-c1c82d89e0a6\") " pod="openstack/dnsmasq-dns-57d65f699f-cf7rm" Nov 21 14:29:57 crc kubenswrapper[4897]: I1121 14:29:57.083951 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/56092cc0-5e4c-47bb-98c1-c1c82d89e0a6-dns-svc\") pod \"dnsmasq-dns-57d65f699f-cf7rm\" (UID: \"56092cc0-5e4c-47bb-98c1-c1c82d89e0a6\") " pod="openstack/dnsmasq-dns-57d65f699f-cf7rm" Nov 21 14:29:57 crc kubenswrapper[4897]: I1121 14:29:57.084131 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/56092cc0-5e4c-47bb-98c1-c1c82d89e0a6-config\") pod \"dnsmasq-dns-57d65f699f-cf7rm\" (UID: \"56092cc0-5e4c-47bb-98c1-c1c82d89e0a6\") " pod="openstack/dnsmasq-dns-57d65f699f-cf7rm" Nov 21 14:29:57 crc kubenswrapper[4897]: I1121 14:29:57.125142 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-b42cn"] Nov 21 14:29:57 crc kubenswrapper[4897]: I1121 14:29:57.129923 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-p7hgh" Nov 21 14:29:57 crc kubenswrapper[4897]: I1121 14:29:57.151093 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-7kmrw"] Nov 21 14:29:57 crc kubenswrapper[4897]: I1121 14:29:57.154253 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-7kmrw" Nov 21 14:29:57 crc kubenswrapper[4897]: I1121 14:29:57.158700 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Nov 21 14:29:57 crc kubenswrapper[4897]: I1121 14:29:57.161161 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-7kmrw"] Nov 21 14:29:57 crc kubenswrapper[4897]: I1121 14:29:57.186728 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/56092cc0-5e4c-47bb-98c1-c1c82d89e0a6-ovsdbserver-nb\") pod \"dnsmasq-dns-57d65f699f-cf7rm\" (UID: \"56092cc0-5e4c-47bb-98c1-c1c82d89e0a6\") " pod="openstack/dnsmasq-dns-57d65f699f-cf7rm" Nov 21 14:29:57 crc kubenswrapper[4897]: I1121 14:29:57.187262 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/56092cc0-5e4c-47bb-98c1-c1c82d89e0a6-dns-svc\") pod \"dnsmasq-dns-57d65f699f-cf7rm\" (UID: \"56092cc0-5e4c-47bb-98c1-c1c82d89e0a6\") " pod="openstack/dnsmasq-dns-57d65f699f-cf7rm" Nov 21 14:29:57 crc kubenswrapper[4897]: I1121 14:29:57.187349 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7a31e75c-6e03-40e3-927f-6b4e04c1efee-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-7kmrw\" (UID: \"7a31e75c-6e03-40e3-927f-6b4e04c1efee\") " pod="openstack/dnsmasq-dns-b8fbc5445-7kmrw" Nov 21 14:29:57 crc kubenswrapper[4897]: I1121 14:29:57.187436 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7a31e75c-6e03-40e3-927f-6b4e04c1efee-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-7kmrw\" (UID: \"7a31e75c-6e03-40e3-927f-6b4e04c1efee\") " pod="openstack/dnsmasq-dns-b8fbc5445-7kmrw" Nov 21 14:29:57 crc kubenswrapper[4897]: I1121 14:29:57.187526 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a31e75c-6e03-40e3-927f-6b4e04c1efee-config\") pod \"dnsmasq-dns-b8fbc5445-7kmrw\" (UID: \"7a31e75c-6e03-40e3-927f-6b4e04c1efee\") " pod="openstack/dnsmasq-dns-b8fbc5445-7kmrw" Nov 21 14:29:57 crc kubenswrapper[4897]: I1121 14:29:57.187577 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5vxc9\" (UniqueName: \"kubernetes.io/projected/7a31e75c-6e03-40e3-927f-6b4e04c1efee-kube-api-access-5vxc9\") pod \"dnsmasq-dns-b8fbc5445-7kmrw\" (UID: \"7a31e75c-6e03-40e3-927f-6b4e04c1efee\") " pod="openstack/dnsmasq-dns-b8fbc5445-7kmrw" Nov 21 14:29:57 crc kubenswrapper[4897]: I1121 14:29:57.187663 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/56092cc0-5e4c-47bb-98c1-c1c82d89e0a6-config\") pod \"dnsmasq-dns-57d65f699f-cf7rm\" (UID: \"56092cc0-5e4c-47bb-98c1-c1c82d89e0a6\") " pod="openstack/dnsmasq-dns-57d65f699f-cf7rm" Nov 21 14:29:57 crc kubenswrapper[4897]: I1121 14:29:57.187730 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7a31e75c-6e03-40e3-927f-6b4e04c1efee-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-7kmrw\" (UID: \"7a31e75c-6e03-40e3-927f-6b4e04c1efee\") " pod="openstack/dnsmasq-dns-b8fbc5445-7kmrw" Nov 21 14:29:57 crc kubenswrapper[4897]: I1121 14:29:57.187788 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vpm8r\" (UniqueName: \"kubernetes.io/projected/56092cc0-5e4c-47bb-98c1-c1c82d89e0a6-kube-api-access-vpm8r\") pod \"dnsmasq-dns-57d65f699f-cf7rm\" (UID: \"56092cc0-5e4c-47bb-98c1-c1c82d89e0a6\") " pod="openstack/dnsmasq-dns-57d65f699f-cf7rm" Nov 21 14:29:57 crc kubenswrapper[4897]: I1121 14:29:57.189022 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/56092cc0-5e4c-47bb-98c1-c1c82d89e0a6-dns-svc\") pod \"dnsmasq-dns-57d65f699f-cf7rm\" (UID: \"56092cc0-5e4c-47bb-98c1-c1c82d89e0a6\") " pod="openstack/dnsmasq-dns-57d65f699f-cf7rm" Nov 21 14:29:57 crc kubenswrapper[4897]: I1121 14:29:57.189611 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/56092cc0-5e4c-47bb-98c1-c1c82d89e0a6-ovsdbserver-nb\") pod \"dnsmasq-dns-57d65f699f-cf7rm\" (UID: \"56092cc0-5e4c-47bb-98c1-c1c82d89e0a6\") " pod="openstack/dnsmasq-dns-57d65f699f-cf7rm" Nov 21 14:29:57 crc kubenswrapper[4897]: I1121 14:29:57.191203 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/56092cc0-5e4c-47bb-98c1-c1c82d89e0a6-config\") pod \"dnsmasq-dns-57d65f699f-cf7rm\" (UID: \"56092cc0-5e4c-47bb-98c1-c1c82d89e0a6\") " pod="openstack/dnsmasq-dns-57d65f699f-cf7rm" Nov 21 14:29:57 crc kubenswrapper[4897]: I1121 14:29:57.217714 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vpm8r\" (UniqueName: \"kubernetes.io/projected/56092cc0-5e4c-47bb-98c1-c1c82d89e0a6-kube-api-access-vpm8r\") pod \"dnsmasq-dns-57d65f699f-cf7rm\" (UID: \"56092cc0-5e4c-47bb-98c1-c1c82d89e0a6\") " pod="openstack/dnsmasq-dns-57d65f699f-cf7rm" Nov 21 14:29:57 crc kubenswrapper[4897]: I1121 14:29:57.289332 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7a31e75c-6e03-40e3-927f-6b4e04c1efee-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-7kmrw\" (UID: \"7a31e75c-6e03-40e3-927f-6b4e04c1efee\") " pod="openstack/dnsmasq-dns-b8fbc5445-7kmrw" Nov 21 14:29:57 crc kubenswrapper[4897]: I1121 14:29:57.289451 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7a31e75c-6e03-40e3-927f-6b4e04c1efee-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-7kmrw\" (UID: \"7a31e75c-6e03-40e3-927f-6b4e04c1efee\") " pod="openstack/dnsmasq-dns-b8fbc5445-7kmrw" Nov 21 14:29:57 crc kubenswrapper[4897]: I1121 14:29:57.289498 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7a31e75c-6e03-40e3-927f-6b4e04c1efee-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-7kmrw\" (UID: \"7a31e75c-6e03-40e3-927f-6b4e04c1efee\") " pod="openstack/dnsmasq-dns-b8fbc5445-7kmrw" Nov 21 14:29:57 crc kubenswrapper[4897]: I1121 14:29:57.289552 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a31e75c-6e03-40e3-927f-6b4e04c1efee-config\") pod \"dnsmasq-dns-b8fbc5445-7kmrw\" (UID: \"7a31e75c-6e03-40e3-927f-6b4e04c1efee\") " pod="openstack/dnsmasq-dns-b8fbc5445-7kmrw" Nov 21 14:29:57 crc kubenswrapper[4897]: I1121 14:29:57.289588 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5vxc9\" (UniqueName: \"kubernetes.io/projected/7a31e75c-6e03-40e3-927f-6b4e04c1efee-kube-api-access-5vxc9\") pod \"dnsmasq-dns-b8fbc5445-7kmrw\" (UID: \"7a31e75c-6e03-40e3-927f-6b4e04c1efee\") " pod="openstack/dnsmasq-dns-b8fbc5445-7kmrw" Nov 21 14:29:57 crc kubenswrapper[4897]: I1121 14:29:57.290409 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7a31e75c-6e03-40e3-927f-6b4e04c1efee-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-7kmrw\" (UID: \"7a31e75c-6e03-40e3-927f-6b4e04c1efee\") " pod="openstack/dnsmasq-dns-b8fbc5445-7kmrw" Nov 21 14:29:57 crc kubenswrapper[4897]: I1121 14:29:57.290437 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7a31e75c-6e03-40e3-927f-6b4e04c1efee-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-7kmrw\" (UID: \"7a31e75c-6e03-40e3-927f-6b4e04c1efee\") " pod="openstack/dnsmasq-dns-b8fbc5445-7kmrw" Nov 21 14:29:57 crc kubenswrapper[4897]: I1121 14:29:57.290610 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7a31e75c-6e03-40e3-927f-6b4e04c1efee-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-7kmrw\" (UID: \"7a31e75c-6e03-40e3-927f-6b4e04c1efee\") " pod="openstack/dnsmasq-dns-b8fbc5445-7kmrw" Nov 21 14:29:57 crc kubenswrapper[4897]: I1121 14:29:57.291285 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a31e75c-6e03-40e3-927f-6b4e04c1efee-config\") pod \"dnsmasq-dns-b8fbc5445-7kmrw\" (UID: \"7a31e75c-6e03-40e3-927f-6b4e04c1efee\") " pod="openstack/dnsmasq-dns-b8fbc5445-7kmrw" Nov 21 14:29:57 crc kubenswrapper[4897]: I1121 14:29:57.310027 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5vxc9\" (UniqueName: \"kubernetes.io/projected/7a31e75c-6e03-40e3-927f-6b4e04c1efee-kube-api-access-5vxc9\") pod \"dnsmasq-dns-b8fbc5445-7kmrw\" (UID: \"7a31e75c-6e03-40e3-927f-6b4e04c1efee\") " pod="openstack/dnsmasq-dns-b8fbc5445-7kmrw" Nov 21 14:29:57 crc kubenswrapper[4897]: I1121 14:29:57.388529 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d65f699f-cf7rm" Nov 21 14:29:57 crc kubenswrapper[4897]: I1121 14:29:57.482659 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Nov 21 14:29:57 crc kubenswrapper[4897]: I1121 14:29:57.491762 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-7kmrw" Nov 21 14:29:57 crc kubenswrapper[4897]: I1121 14:29:57.539554 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Nov 21 14:29:57 crc kubenswrapper[4897]: I1121 14:29:57.540534 4897 generic.go:334] "Generic (PLEG): container finished" podID="fce5a726-bd06-473b-a9d0-f812ffe83693" containerID="8a7ca999bf42855f24b3ac03dac8be32e2bdc1d5214ae873b55de75611e26376" exitCode=0 Nov 21 14:29:57 crc kubenswrapper[4897]: I1121 14:29:57.540650 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"fce5a726-bd06-473b-a9d0-f812ffe83693","Type":"ContainerDied","Data":"8a7ca999bf42855f24b3ac03dac8be32e2bdc1d5214ae873b55de75611e26376"} Nov 21 14:29:57 crc kubenswrapper[4897]: I1121 14:29:57.719461 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-p7hgh"] Nov 21 14:29:57 crc kubenswrapper[4897]: I1121 14:29:57.918410 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Nov 21 14:29:57 crc kubenswrapper[4897]: I1121 14:29:57.919048 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Nov 21 14:29:57 crc kubenswrapper[4897]: I1121 14:29:57.955274 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d65f699f-cf7rm"] Nov 21 14:29:58 crc kubenswrapper[4897]: I1121 14:29:58.105647 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-7kmrw"] Nov 21 14:29:58 crc kubenswrapper[4897]: I1121 14:29:58.209540 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Nov 21 14:29:58 crc kubenswrapper[4897]: I1121 14:29:58.485260 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Nov 21 14:29:58 crc kubenswrapper[4897]: I1121 14:29:58.526800 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Nov 21 14:29:58 crc kubenswrapper[4897]: I1121 14:29:58.552708 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-p7hgh" event={"ID":"15ab9acd-00ef-4b25-8c82-b86e95ea046f","Type":"ContainerStarted","Data":"ddfa243c81a8c192d839fdaad279894c0358e498d5f1e645959f8457900bce0f"} Nov 21 14:29:58 crc kubenswrapper[4897]: I1121 14:29:58.552760 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-p7hgh" event={"ID":"15ab9acd-00ef-4b25-8c82-b86e95ea046f","Type":"ContainerStarted","Data":"ebdea3f4ce03ceac5fd3e87cb5df14289040100cd823084a6bd5290048de26ac"} Nov 21 14:29:58 crc kubenswrapper[4897]: I1121 14:29:58.556040 4897 generic.go:334] "Generic (PLEG): container finished" podID="56092cc0-5e4c-47bb-98c1-c1c82d89e0a6" containerID="f19dbbb5857a286582bed5bb5631a58480e595744701974407d574911f1accf4" exitCode=0 Nov 21 14:29:58 crc kubenswrapper[4897]: I1121 14:29:58.556071 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d65f699f-cf7rm" event={"ID":"56092cc0-5e4c-47bb-98c1-c1c82d89e0a6","Type":"ContainerDied","Data":"f19dbbb5857a286582bed5bb5631a58480e595744701974407d574911f1accf4"} Nov 21 14:29:58 crc kubenswrapper[4897]: I1121 14:29:58.556102 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d65f699f-cf7rm" event={"ID":"56092cc0-5e4c-47bb-98c1-c1c82d89e0a6","Type":"ContainerStarted","Data":"e260d4871100c1508edc71586548bd155eea7f01d400d683d99fc5cf43729b70"} Nov 21 14:29:58 crc kubenswrapper[4897]: I1121 14:29:58.559123 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"fce5a726-bd06-473b-a9d0-f812ffe83693","Type":"ContainerStarted","Data":"52e6b241adf867ce78c2231c1fd888c9a04bea579f474fdae1e94a124523ff5f"} Nov 21 14:29:58 crc kubenswrapper[4897]: I1121 14:29:58.560571 4897 generic.go:334] "Generic (PLEG): container finished" podID="7a31e75c-6e03-40e3-927f-6b4e04c1efee" containerID="1f2d47fd9bfccad73e9b004362d0525c4684960507722d851013a098334c21fd" exitCode=0 Nov 21 14:29:58 crc kubenswrapper[4897]: I1121 14:29:58.560721 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7cb5889db5-b42cn" podUID="9306e5a3-58c4-462b-bad5-0e1f876bf832" containerName="dnsmasq-dns" containerID="cri-o://0eca3681977bafa554467e45117ee7088cec805d162c8912d1c4a66c201f62f4" gracePeriod=10 Nov 21 14:29:58 crc kubenswrapper[4897]: I1121 14:29:58.560781 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-666b6646f7-khkvq" podUID="6bf04b77-0ef9-42f0-8a02-8c4c7568e390" containerName="dnsmasq-dns" containerID="cri-o://b06105b5b07e06d646b4c304d1f59865380586cfc30e6b169b23c6dfa177f20a" gracePeriod=10 Nov 21 14:29:58 crc kubenswrapper[4897]: I1121 14:29:58.560850 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-7kmrw" event={"ID":"7a31e75c-6e03-40e3-927f-6b4e04c1efee","Type":"ContainerDied","Data":"1f2d47fd9bfccad73e9b004362d0525c4684960507722d851013a098334c21fd"} Nov 21 14:29:58 crc kubenswrapper[4897]: I1121 14:29:58.560871 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-7kmrw" event={"ID":"7a31e75c-6e03-40e3-927f-6b4e04c1efee","Type":"ContainerStarted","Data":"fe1b324f77ac15b5c51888eefead22f6b6c637ac576efe5d64e306f14244f92e"} Nov 21 14:29:58 crc kubenswrapper[4897]: I1121 14:29:58.577297 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-p7hgh" podStartSLOduration=2.5772732879999998 podStartE2EDuration="2.577273288s" podCreationTimestamp="2025-11-21 14:29:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:29:58.571699058 +0000 UTC m=+1275.856292553" watchObservedRunningTime="2025-11-21 14:29:58.577273288 +0000 UTC m=+1275.861866773" Nov 21 14:29:58 crc kubenswrapper[4897]: I1121 14:29:58.672776 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Nov 21 14:29:58 crc kubenswrapper[4897]: I1121 14:29:58.706680 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=-9223371983.148117 podStartE2EDuration="53.706658488s" podCreationTimestamp="2025-11-21 14:29:05 +0000 UTC" firstStartedPulling="2025-11-21 14:29:07.466213995 +0000 UTC m=+1224.750807470" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:29:58.697464341 +0000 UTC m=+1275.982057826" watchObservedRunningTime="2025-11-21 14:29:58.706658488 +0000 UTC m=+1275.991251963" Nov 21 14:29:58 crc kubenswrapper[4897]: I1121 14:29:58.755703 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Nov 21 14:29:58 crc kubenswrapper[4897]: I1121 14:29:58.874337 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Nov 21 14:29:58 crc kubenswrapper[4897]: I1121 14:29:58.876234 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Nov 21 14:29:58 crc kubenswrapper[4897]: I1121 14:29:58.881175 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Nov 21 14:29:58 crc kubenswrapper[4897]: I1121 14:29:58.882330 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-99d49" Nov 21 14:29:58 crc kubenswrapper[4897]: I1121 14:29:58.886492 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Nov 21 14:29:58 crc kubenswrapper[4897]: I1121 14:29:58.888743 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Nov 21 14:29:58 crc kubenswrapper[4897]: I1121 14:29:58.888983 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Nov 21 14:29:58 crc kubenswrapper[4897]: I1121 14:29:58.948739 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/63134f01-8622-4fa2-80a3-516572217d0f-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"63134f01-8622-4fa2-80a3-516572217d0f\") " pod="openstack/ovn-northd-0" Nov 21 14:29:58 crc kubenswrapper[4897]: I1121 14:29:58.949064 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/63134f01-8622-4fa2-80a3-516572217d0f-scripts\") pod \"ovn-northd-0\" (UID: \"63134f01-8622-4fa2-80a3-516572217d0f\") " pod="openstack/ovn-northd-0" Nov 21 14:29:58 crc kubenswrapper[4897]: I1121 14:29:58.949176 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6k5vt\" (UniqueName: \"kubernetes.io/projected/63134f01-8622-4fa2-80a3-516572217d0f-kube-api-access-6k5vt\") pod \"ovn-northd-0\" (UID: \"63134f01-8622-4fa2-80a3-516572217d0f\") " pod="openstack/ovn-northd-0" Nov 21 14:29:58 crc kubenswrapper[4897]: I1121 14:29:58.949226 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/63134f01-8622-4fa2-80a3-516572217d0f-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"63134f01-8622-4fa2-80a3-516572217d0f\") " pod="openstack/ovn-northd-0" Nov 21 14:29:58 crc kubenswrapper[4897]: I1121 14:29:58.949323 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63134f01-8622-4fa2-80a3-516572217d0f-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"63134f01-8622-4fa2-80a3-516572217d0f\") " pod="openstack/ovn-northd-0" Nov 21 14:29:58 crc kubenswrapper[4897]: I1121 14:29:58.949376 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/63134f01-8622-4fa2-80a3-516572217d0f-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"63134f01-8622-4fa2-80a3-516572217d0f\") " pod="openstack/ovn-northd-0" Nov 21 14:29:58 crc kubenswrapper[4897]: I1121 14:29:58.949419 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/63134f01-8622-4fa2-80a3-516572217d0f-config\") pod \"ovn-northd-0\" (UID: \"63134f01-8622-4fa2-80a3-516572217d0f\") " pod="openstack/ovn-northd-0" Nov 21 14:29:59 crc kubenswrapper[4897]: I1121 14:29:59.051309 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/63134f01-8622-4fa2-80a3-516572217d0f-scripts\") pod \"ovn-northd-0\" (UID: \"63134f01-8622-4fa2-80a3-516572217d0f\") " pod="openstack/ovn-northd-0" Nov 21 14:29:59 crc kubenswrapper[4897]: I1121 14:29:59.051386 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6k5vt\" (UniqueName: \"kubernetes.io/projected/63134f01-8622-4fa2-80a3-516572217d0f-kube-api-access-6k5vt\") pod \"ovn-northd-0\" (UID: \"63134f01-8622-4fa2-80a3-516572217d0f\") " pod="openstack/ovn-northd-0" Nov 21 14:29:59 crc kubenswrapper[4897]: I1121 14:29:59.051421 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/63134f01-8622-4fa2-80a3-516572217d0f-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"63134f01-8622-4fa2-80a3-516572217d0f\") " pod="openstack/ovn-northd-0" Nov 21 14:29:59 crc kubenswrapper[4897]: I1121 14:29:59.051482 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63134f01-8622-4fa2-80a3-516572217d0f-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"63134f01-8622-4fa2-80a3-516572217d0f\") " pod="openstack/ovn-northd-0" Nov 21 14:29:59 crc kubenswrapper[4897]: I1121 14:29:59.051524 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/63134f01-8622-4fa2-80a3-516572217d0f-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"63134f01-8622-4fa2-80a3-516572217d0f\") " pod="openstack/ovn-northd-0" Nov 21 14:29:59 crc kubenswrapper[4897]: I1121 14:29:59.051554 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/5f9bf5d1-dcbd-4d19-b46d-22810ebf17a2-etc-swift\") pod \"swift-storage-0\" (UID: \"5f9bf5d1-dcbd-4d19-b46d-22810ebf17a2\") " pod="openstack/swift-storage-0" Nov 21 14:29:59 crc kubenswrapper[4897]: I1121 14:29:59.051578 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/63134f01-8622-4fa2-80a3-516572217d0f-config\") pod \"ovn-northd-0\" (UID: \"63134f01-8622-4fa2-80a3-516572217d0f\") " pod="openstack/ovn-northd-0" Nov 21 14:29:59 crc kubenswrapper[4897]: E1121 14:29:59.051792 4897 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 21 14:29:59 crc kubenswrapper[4897]: E1121 14:29:59.051814 4897 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 21 14:29:59 crc kubenswrapper[4897]: E1121 14:29:59.051858 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/5f9bf5d1-dcbd-4d19-b46d-22810ebf17a2-etc-swift podName:5f9bf5d1-dcbd-4d19-b46d-22810ebf17a2 nodeName:}" failed. No retries permitted until 2025-11-21 14:30:07.051838466 +0000 UTC m=+1284.336431951 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/5f9bf5d1-dcbd-4d19-b46d-22810ebf17a2-etc-swift") pod "swift-storage-0" (UID: "5f9bf5d1-dcbd-4d19-b46d-22810ebf17a2") : configmap "swift-ring-files" not found Nov 21 14:29:59 crc kubenswrapper[4897]: I1121 14:29:59.051973 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/63134f01-8622-4fa2-80a3-516572217d0f-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"63134f01-8622-4fa2-80a3-516572217d0f\") " pod="openstack/ovn-northd-0" Nov 21 14:29:59 crc kubenswrapper[4897]: I1121 14:29:59.055359 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/63134f01-8622-4fa2-80a3-516572217d0f-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"63134f01-8622-4fa2-80a3-516572217d0f\") " pod="openstack/ovn-northd-0" Nov 21 14:29:59 crc kubenswrapper[4897]: I1121 14:29:59.055580 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/63134f01-8622-4fa2-80a3-516572217d0f-config\") pod \"ovn-northd-0\" (UID: \"63134f01-8622-4fa2-80a3-516572217d0f\") " pod="openstack/ovn-northd-0" Nov 21 14:29:59 crc kubenswrapper[4897]: I1121 14:29:59.056681 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/63134f01-8622-4fa2-80a3-516572217d0f-scripts\") pod \"ovn-northd-0\" (UID: \"63134f01-8622-4fa2-80a3-516572217d0f\") " pod="openstack/ovn-northd-0" Nov 21 14:29:59 crc kubenswrapper[4897]: I1121 14:29:59.059348 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63134f01-8622-4fa2-80a3-516572217d0f-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"63134f01-8622-4fa2-80a3-516572217d0f\") " pod="openstack/ovn-northd-0" Nov 21 14:29:59 crc kubenswrapper[4897]: I1121 14:29:59.060737 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/63134f01-8622-4fa2-80a3-516572217d0f-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"63134f01-8622-4fa2-80a3-516572217d0f\") " pod="openstack/ovn-northd-0" Nov 21 14:29:59 crc kubenswrapper[4897]: I1121 14:29:59.060799 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/63134f01-8622-4fa2-80a3-516572217d0f-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"63134f01-8622-4fa2-80a3-516572217d0f\") " pod="openstack/ovn-northd-0" Nov 21 14:29:59 crc kubenswrapper[4897]: I1121 14:29:59.069671 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6k5vt\" (UniqueName: \"kubernetes.io/projected/63134f01-8622-4fa2-80a3-516572217d0f-kube-api-access-6k5vt\") pod \"ovn-northd-0\" (UID: \"63134f01-8622-4fa2-80a3-516572217d0f\") " pod="openstack/ovn-northd-0" Nov 21 14:29:59 crc kubenswrapper[4897]: I1121 14:29:59.196767 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Nov 21 14:29:59 crc kubenswrapper[4897]: I1121 14:29:59.571090 4897 generic.go:334] "Generic (PLEG): container finished" podID="9306e5a3-58c4-462b-bad5-0e1f876bf832" containerID="0eca3681977bafa554467e45117ee7088cec805d162c8912d1c4a66c201f62f4" exitCode=0 Nov 21 14:29:59 crc kubenswrapper[4897]: I1121 14:29:59.571165 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-b42cn" event={"ID":"9306e5a3-58c4-462b-bad5-0e1f876bf832","Type":"ContainerDied","Data":"0eca3681977bafa554467e45117ee7088cec805d162c8912d1c4a66c201f62f4"} Nov 21 14:29:59 crc kubenswrapper[4897]: I1121 14:29:59.575578 4897 generic.go:334] "Generic (PLEG): container finished" podID="6bf04b77-0ef9-42f0-8a02-8c4c7568e390" containerID="b06105b5b07e06d646b4c304d1f59865380586cfc30e6b169b23c6dfa177f20a" exitCode=0 Nov 21 14:29:59 crc kubenswrapper[4897]: I1121 14:29:59.575677 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-khkvq" event={"ID":"6bf04b77-0ef9-42f0-8a02-8c4c7568e390","Type":"ContainerDied","Data":"b06105b5b07e06d646b4c304d1f59865380586cfc30e6b169b23c6dfa177f20a"} Nov 21 14:29:59 crc kubenswrapper[4897]: I1121 14:29:59.577975 4897 generic.go:334] "Generic (PLEG): container finished" podID="91aa2270-92a7-4af4-9a92-141ea2b24ee1" containerID="ae9d2f28e206da418bad9d39e2d6c5076e3d741d413ec6b6cb6c3b13dfa61274" exitCode=0 Nov 21 14:29:59 crc kubenswrapper[4897]: I1121 14:29:59.578083 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"91aa2270-92a7-4af4-9a92-141ea2b24ee1","Type":"ContainerDied","Data":"ae9d2f28e206da418bad9d39e2d6c5076e3d741d413ec6b6cb6c3b13dfa61274"} Nov 21 14:30:00 crc kubenswrapper[4897]: I1121 14:30:00.144264 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29395590-t8htq"] Nov 21 14:30:00 crc kubenswrapper[4897]: I1121 14:30:00.145995 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29395590-t8htq" Nov 21 14:30:00 crc kubenswrapper[4897]: I1121 14:30:00.148345 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 21 14:30:00 crc kubenswrapper[4897]: I1121 14:30:00.150481 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 21 14:30:00 crc kubenswrapper[4897]: I1121 14:30:00.156821 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29395590-t8htq"] Nov 21 14:30:00 crc kubenswrapper[4897]: I1121 14:30:00.191259 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/99d4fc7b-7c46-47e7-8bf6-906dfd87af3a-config-volume\") pod \"collect-profiles-29395590-t8htq\" (UID: \"99d4fc7b-7c46-47e7-8bf6-906dfd87af3a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29395590-t8htq" Nov 21 14:30:00 crc kubenswrapper[4897]: I1121 14:30:00.191366 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hznxg\" (UniqueName: \"kubernetes.io/projected/99d4fc7b-7c46-47e7-8bf6-906dfd87af3a-kube-api-access-hznxg\") pod \"collect-profiles-29395590-t8htq\" (UID: \"99d4fc7b-7c46-47e7-8bf6-906dfd87af3a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29395590-t8htq" Nov 21 14:30:00 crc kubenswrapper[4897]: I1121 14:30:00.191813 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/99d4fc7b-7c46-47e7-8bf6-906dfd87af3a-secret-volume\") pod \"collect-profiles-29395590-t8htq\" (UID: \"99d4fc7b-7c46-47e7-8bf6-906dfd87af3a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29395590-t8htq" Nov 21 14:30:00 crc kubenswrapper[4897]: I1121 14:30:00.294427 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/99d4fc7b-7c46-47e7-8bf6-906dfd87af3a-secret-volume\") pod \"collect-profiles-29395590-t8htq\" (UID: \"99d4fc7b-7c46-47e7-8bf6-906dfd87af3a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29395590-t8htq" Nov 21 14:30:00 crc kubenswrapper[4897]: I1121 14:30:00.294939 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/99d4fc7b-7c46-47e7-8bf6-906dfd87af3a-config-volume\") pod \"collect-profiles-29395590-t8htq\" (UID: \"99d4fc7b-7c46-47e7-8bf6-906dfd87af3a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29395590-t8htq" Nov 21 14:30:00 crc kubenswrapper[4897]: I1121 14:30:00.295127 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hznxg\" (UniqueName: \"kubernetes.io/projected/99d4fc7b-7c46-47e7-8bf6-906dfd87af3a-kube-api-access-hznxg\") pod \"collect-profiles-29395590-t8htq\" (UID: \"99d4fc7b-7c46-47e7-8bf6-906dfd87af3a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29395590-t8htq" Nov 21 14:30:00 crc kubenswrapper[4897]: I1121 14:30:00.295815 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/99d4fc7b-7c46-47e7-8bf6-906dfd87af3a-config-volume\") pod \"collect-profiles-29395590-t8htq\" (UID: \"99d4fc7b-7c46-47e7-8bf6-906dfd87af3a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29395590-t8htq" Nov 21 14:30:00 crc kubenswrapper[4897]: I1121 14:30:00.305583 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/99d4fc7b-7c46-47e7-8bf6-906dfd87af3a-secret-volume\") pod \"collect-profiles-29395590-t8htq\" (UID: \"99d4fc7b-7c46-47e7-8bf6-906dfd87af3a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29395590-t8htq" Nov 21 14:30:00 crc kubenswrapper[4897]: I1121 14:30:00.318577 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hznxg\" (UniqueName: \"kubernetes.io/projected/99d4fc7b-7c46-47e7-8bf6-906dfd87af3a-kube-api-access-hznxg\") pod \"collect-profiles-29395590-t8htq\" (UID: \"99d4fc7b-7c46-47e7-8bf6-906dfd87af3a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29395590-t8htq" Nov 21 14:30:00 crc kubenswrapper[4897]: I1121 14:30:00.368481 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Nov 21 14:30:00 crc kubenswrapper[4897]: I1121 14:30:00.477068 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29395590-t8htq" Nov 21 14:30:00 crc kubenswrapper[4897]: I1121 14:30:00.512989 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-7cb5889db5-b42cn" podUID="9306e5a3-58c4-462b-bad5-0e1f876bf832" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.143:5353: connect: connection refused" Nov 21 14:30:01 crc kubenswrapper[4897]: I1121 14:30:01.867267 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-khkvq" Nov 21 14:30:01 crc kubenswrapper[4897]: I1121 14:30:01.946214 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6b5qf\" (UniqueName: \"kubernetes.io/projected/6bf04b77-0ef9-42f0-8a02-8c4c7568e390-kube-api-access-6b5qf\") pod \"6bf04b77-0ef9-42f0-8a02-8c4c7568e390\" (UID: \"6bf04b77-0ef9-42f0-8a02-8c4c7568e390\") " Nov 21 14:30:01 crc kubenswrapper[4897]: I1121 14:30:01.946313 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6bf04b77-0ef9-42f0-8a02-8c4c7568e390-dns-svc\") pod \"6bf04b77-0ef9-42f0-8a02-8c4c7568e390\" (UID: \"6bf04b77-0ef9-42f0-8a02-8c4c7568e390\") " Nov 21 14:30:01 crc kubenswrapper[4897]: I1121 14:30:01.946609 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6bf04b77-0ef9-42f0-8a02-8c4c7568e390-config\") pod \"6bf04b77-0ef9-42f0-8a02-8c4c7568e390\" (UID: \"6bf04b77-0ef9-42f0-8a02-8c4c7568e390\") " Nov 21 14:30:01 crc kubenswrapper[4897]: I1121 14:30:01.954925 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6bf04b77-0ef9-42f0-8a02-8c4c7568e390-kube-api-access-6b5qf" (OuterVolumeSpecName: "kube-api-access-6b5qf") pod "6bf04b77-0ef9-42f0-8a02-8c4c7568e390" (UID: "6bf04b77-0ef9-42f0-8a02-8c4c7568e390"). InnerVolumeSpecName "kube-api-access-6b5qf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:30:02 crc kubenswrapper[4897]: I1121 14:30:02.018344 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6bf04b77-0ef9-42f0-8a02-8c4c7568e390-config" (OuterVolumeSpecName: "config") pod "6bf04b77-0ef9-42f0-8a02-8c4c7568e390" (UID: "6bf04b77-0ef9-42f0-8a02-8c4c7568e390"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:30:02 crc kubenswrapper[4897]: I1121 14:30:02.026524 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6bf04b77-0ef9-42f0-8a02-8c4c7568e390-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6bf04b77-0ef9-42f0-8a02-8c4c7568e390" (UID: "6bf04b77-0ef9-42f0-8a02-8c4c7568e390"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:30:02 crc kubenswrapper[4897]: I1121 14:30:02.053971 4897 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6bf04b77-0ef9-42f0-8a02-8c4c7568e390-config\") on node \"crc\" DevicePath \"\"" Nov 21 14:30:02 crc kubenswrapper[4897]: I1121 14:30:02.054005 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6b5qf\" (UniqueName: \"kubernetes.io/projected/6bf04b77-0ef9-42f0-8a02-8c4c7568e390-kube-api-access-6b5qf\") on node \"crc\" DevicePath \"\"" Nov 21 14:30:02 crc kubenswrapper[4897]: I1121 14:30:02.054015 4897 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6bf04b77-0ef9-42f0-8a02-8c4c7568e390-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 21 14:30:02 crc kubenswrapper[4897]: I1121 14:30:02.092553 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb5889db5-b42cn" Nov 21 14:30:02 crc kubenswrapper[4897]: I1121 14:30:02.155765 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9306e5a3-58c4-462b-bad5-0e1f876bf832-config\") pod \"9306e5a3-58c4-462b-bad5-0e1f876bf832\" (UID: \"9306e5a3-58c4-462b-bad5-0e1f876bf832\") " Nov 21 14:30:02 crc kubenswrapper[4897]: I1121 14:30:02.155826 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9306e5a3-58c4-462b-bad5-0e1f876bf832-dns-svc\") pod \"9306e5a3-58c4-462b-bad5-0e1f876bf832\" (UID: \"9306e5a3-58c4-462b-bad5-0e1f876bf832\") " Nov 21 14:30:02 crc kubenswrapper[4897]: I1121 14:30:02.155963 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m9ttz\" (UniqueName: \"kubernetes.io/projected/9306e5a3-58c4-462b-bad5-0e1f876bf832-kube-api-access-m9ttz\") pod \"9306e5a3-58c4-462b-bad5-0e1f876bf832\" (UID: \"9306e5a3-58c4-462b-bad5-0e1f876bf832\") " Nov 21 14:30:02 crc kubenswrapper[4897]: I1121 14:30:02.201752 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9306e5a3-58c4-462b-bad5-0e1f876bf832-kube-api-access-m9ttz" (OuterVolumeSpecName: "kube-api-access-m9ttz") pod "9306e5a3-58c4-462b-bad5-0e1f876bf832" (UID: "9306e5a3-58c4-462b-bad5-0e1f876bf832"). InnerVolumeSpecName "kube-api-access-m9ttz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:30:02 crc kubenswrapper[4897]: I1121 14:30:02.218131 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9306e5a3-58c4-462b-bad5-0e1f876bf832-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9306e5a3-58c4-462b-bad5-0e1f876bf832" (UID: "9306e5a3-58c4-462b-bad5-0e1f876bf832"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:30:02 crc kubenswrapper[4897]: I1121 14:30:02.228805 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9306e5a3-58c4-462b-bad5-0e1f876bf832-config" (OuterVolumeSpecName: "config") pod "9306e5a3-58c4-462b-bad5-0e1f876bf832" (UID: "9306e5a3-58c4-462b-bad5-0e1f876bf832"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:30:02 crc kubenswrapper[4897]: I1121 14:30:02.258646 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m9ttz\" (UniqueName: \"kubernetes.io/projected/9306e5a3-58c4-462b-bad5-0e1f876bf832-kube-api-access-m9ttz\") on node \"crc\" DevicePath \"\"" Nov 21 14:30:02 crc kubenswrapper[4897]: I1121 14:30:02.258681 4897 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9306e5a3-58c4-462b-bad5-0e1f876bf832-config\") on node \"crc\" DevicePath \"\"" Nov 21 14:30:02 crc kubenswrapper[4897]: I1121 14:30:02.258695 4897 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9306e5a3-58c4-462b-bad5-0e1f876bf832-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 21 14:30:02 crc kubenswrapper[4897]: I1121 14:30:02.310800 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29395590-t8htq"] Nov 21 14:30:02 crc kubenswrapper[4897]: W1121 14:30:02.317323 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod99d4fc7b_7c46_47e7_8bf6_906dfd87af3a.slice/crio-e6140710ef46809c71ee984176cd1e0216051635074e59e0dd5c4389e85a52c0 WatchSource:0}: Error finding container e6140710ef46809c71ee984176cd1e0216051635074e59e0dd5c4389e85a52c0: Status 404 returned error can't find the container with id e6140710ef46809c71ee984176cd1e0216051635074e59e0dd5c4389e85a52c0 Nov 21 14:30:02 crc kubenswrapper[4897]: W1121 14:30:02.330870 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod63134f01_8622_4fa2_80a3_516572217d0f.slice/crio-d481064b3897ae8c70fd433ff9f0f7e46141f4f1ce87d3c1d3f4e7980af9ec10 WatchSource:0}: Error finding container d481064b3897ae8c70fd433ff9f0f7e46141f4f1ce87d3c1d3f4e7980af9ec10: Status 404 returned error can't find the container with id d481064b3897ae8c70fd433ff9f0f7e46141f4f1ce87d3c1d3f4e7980af9ec10 Nov 21 14:30:02 crc kubenswrapper[4897]: I1121 14:30:02.334369 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Nov 21 14:30:02 crc kubenswrapper[4897]: I1121 14:30:02.613869 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29395590-t8htq" event={"ID":"99d4fc7b-7c46-47e7-8bf6-906dfd87af3a","Type":"ContainerStarted","Data":"e6140710ef46809c71ee984176cd1e0216051635074e59e0dd5c4389e85a52c0"} Nov 21 14:30:02 crc kubenswrapper[4897]: I1121 14:30:02.616053 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d65f699f-cf7rm" event={"ID":"56092cc0-5e4c-47bb-98c1-c1c82d89e0a6","Type":"ContainerStarted","Data":"45c47b725cdc3b5aa9ec4a77953186f3bd2971e1387a0ceaa67234ada9133531"} Nov 21 14:30:02 crc kubenswrapper[4897]: I1121 14:30:02.616337 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57d65f699f-cf7rm" Nov 21 14:30:02 crc kubenswrapper[4897]: I1121 14:30:02.617717 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb5889db5-b42cn" Nov 21 14:30:02 crc kubenswrapper[4897]: I1121 14:30:02.617717 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-b42cn" event={"ID":"9306e5a3-58c4-462b-bad5-0e1f876bf832","Type":"ContainerDied","Data":"5ed5f13738252f6e1b569777b5470c4550fd6da8fc267f2a110323041455fc7d"} Nov 21 14:30:02 crc kubenswrapper[4897]: I1121 14:30:02.617880 4897 scope.go:117] "RemoveContainer" containerID="0eca3681977bafa554467e45117ee7088cec805d162c8912d1c4a66c201f62f4" Nov 21 14:30:02 crc kubenswrapper[4897]: I1121 14:30:02.619375 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-7kmrw" event={"ID":"7a31e75c-6e03-40e3-927f-6b4e04c1efee","Type":"ContainerStarted","Data":"3e1bff1bbd0769c15d6aa8951ca8e1ab60f58ea1d63e68a36263aec05a1a078e"} Nov 21 14:30:02 crc kubenswrapper[4897]: I1121 14:30:02.619487 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-b8fbc5445-7kmrw" Nov 21 14:30:02 crc kubenswrapper[4897]: I1121 14:30:02.622965 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"63134f01-8622-4fa2-80a3-516572217d0f","Type":"ContainerStarted","Data":"d481064b3897ae8c70fd433ff9f0f7e46141f4f1ce87d3c1d3f4e7980af9ec10"} Nov 21 14:30:02 crc kubenswrapper[4897]: I1121 14:30:02.627540 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-khkvq" event={"ID":"6bf04b77-0ef9-42f0-8a02-8c4c7568e390","Type":"ContainerDied","Data":"a72fb2283c9a1f64135fea02b1846c00268b9c8f14cf5808328ead7e0788d548"} Nov 21 14:30:02 crc kubenswrapper[4897]: I1121 14:30:02.627587 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-khkvq" Nov 21 14:30:02 crc kubenswrapper[4897]: I1121 14:30:02.637958 4897 scope.go:117] "RemoveContainer" containerID="206ce3c58802a46d150063b061c85011963c120a26ec12245f2b3374b87a68ab" Nov 21 14:30:02 crc kubenswrapper[4897]: I1121 14:30:02.640463 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57d65f699f-cf7rm" podStartSLOduration=6.640447475 podStartE2EDuration="6.640447475s" podCreationTimestamp="2025-11-21 14:29:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:30:02.633966951 +0000 UTC m=+1279.918560426" watchObservedRunningTime="2025-11-21 14:30:02.640447475 +0000 UTC m=+1279.925040950" Nov 21 14:30:02 crc kubenswrapper[4897]: I1121 14:30:02.661668 4897 scope.go:117] "RemoveContainer" containerID="b06105b5b07e06d646b4c304d1f59865380586cfc30e6b169b23c6dfa177f20a" Nov 21 14:30:02 crc kubenswrapper[4897]: I1121 14:30:02.671027 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-b8fbc5445-7kmrw" podStartSLOduration=5.671010484 podStartE2EDuration="5.671010484s" podCreationTimestamp="2025-11-21 14:29:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:30:02.65218021 +0000 UTC m=+1279.936773685" watchObservedRunningTime="2025-11-21 14:30:02.671010484 +0000 UTC m=+1279.955603959" Nov 21 14:30:02 crc kubenswrapper[4897]: I1121 14:30:02.679844 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-b42cn"] Nov 21 14:30:02 crc kubenswrapper[4897]: I1121 14:30:02.687291 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-b42cn"] Nov 21 14:30:02 crc kubenswrapper[4897]: I1121 14:30:02.693920 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-khkvq"] Nov 21 14:30:02 crc kubenswrapper[4897]: I1121 14:30:02.704753 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-khkvq"] Nov 21 14:30:02 crc kubenswrapper[4897]: I1121 14:30:02.791212 4897 scope.go:117] "RemoveContainer" containerID="6b2b520616824e276cba653c73af285909212743e4ced43cb4bd3f06fdcd21a9" Nov 21 14:30:03 crc kubenswrapper[4897]: I1121 14:30:03.641954 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29395590-t8htq" event={"ID":"99d4fc7b-7c46-47e7-8bf6-906dfd87af3a","Type":"ContainerStarted","Data":"695d365282e23d79ab96d852c2a92d6418c185f004a384bf60851c2e6f2d8673"} Nov 21 14:30:04 crc kubenswrapper[4897]: I1121 14:30:04.103751 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6bf04b77-0ef9-42f0-8a02-8c4c7568e390" path="/var/lib/kubelet/pods/6bf04b77-0ef9-42f0-8a02-8c4c7568e390/volumes" Nov 21 14:30:04 crc kubenswrapper[4897]: I1121 14:30:04.104693 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9306e5a3-58c4-462b-bad5-0e1f876bf832" path="/var/lib/kubelet/pods/9306e5a3-58c4-462b-bad5-0e1f876bf832/volumes" Nov 21 14:30:06 crc kubenswrapper[4897]: I1121 14:30:06.596213 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Nov 21 14:30:06 crc kubenswrapper[4897]: I1121 14:30:06.596859 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Nov 21 14:30:07 crc kubenswrapper[4897]: I1121 14:30:07.073091 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/5f9bf5d1-dcbd-4d19-b46d-22810ebf17a2-etc-swift\") pod \"swift-storage-0\" (UID: \"5f9bf5d1-dcbd-4d19-b46d-22810ebf17a2\") " pod="openstack/swift-storage-0" Nov 21 14:30:07 crc kubenswrapper[4897]: E1121 14:30:07.073467 4897 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 21 14:30:07 crc kubenswrapper[4897]: E1121 14:30:07.073499 4897 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 21 14:30:07 crc kubenswrapper[4897]: E1121 14:30:07.073593 4897 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/5f9bf5d1-dcbd-4d19-b46d-22810ebf17a2-etc-swift podName:5f9bf5d1-dcbd-4d19-b46d-22810ebf17a2 nodeName:}" failed. No retries permitted until 2025-11-21 14:30:23.073569716 +0000 UTC m=+1300.358163201 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/5f9bf5d1-dcbd-4d19-b46d-22810ebf17a2-etc-swift") pod "swift-storage-0" (UID: "5f9bf5d1-dcbd-4d19-b46d-22810ebf17a2") : configmap "swift-ring-files" not found Nov 21 14:30:07 crc kubenswrapper[4897]: I1121 14:30:07.390709 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-57d65f699f-cf7rm" Nov 21 14:30:07 crc kubenswrapper[4897]: I1121 14:30:07.493685 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-b8fbc5445-7kmrw" Nov 21 14:30:07 crc kubenswrapper[4897]: I1121 14:30:07.551523 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d65f699f-cf7rm"] Nov 21 14:30:07 crc kubenswrapper[4897]: I1121 14:30:07.685217 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57d65f699f-cf7rm" podUID="56092cc0-5e4c-47bb-98c1-c1c82d89e0a6" containerName="dnsmasq-dns" containerID="cri-o://45c47b725cdc3b5aa9ec4a77953186f3bd2971e1387a0ceaa67234ada9133531" gracePeriod=10 Nov 21 14:30:11 crc kubenswrapper[4897]: I1121 14:30:11.747025 4897 generic.go:334] "Generic (PLEG): container finished" podID="56092cc0-5e4c-47bb-98c1-c1c82d89e0a6" containerID="45c47b725cdc3b5aa9ec4a77953186f3bd2971e1387a0ceaa67234ada9133531" exitCode=0 Nov 21 14:30:11 crc kubenswrapper[4897]: I1121 14:30:11.747279 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d65f699f-cf7rm" event={"ID":"56092cc0-5e4c-47bb-98c1-c1c82d89e0a6","Type":"ContainerDied","Data":"45c47b725cdc3b5aa9ec4a77953186f3bd2971e1387a0ceaa67234ada9133531"} Nov 21 14:30:11 crc kubenswrapper[4897]: I1121 14:30:11.749981 4897 generic.go:334] "Generic (PLEG): container finished" podID="99d4fc7b-7c46-47e7-8bf6-906dfd87af3a" containerID="695d365282e23d79ab96d852c2a92d6418c185f004a384bf60851c2e6f2d8673" exitCode=0 Nov 21 14:30:11 crc kubenswrapper[4897]: I1121 14:30:11.750035 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29395590-t8htq" event={"ID":"99d4fc7b-7c46-47e7-8bf6-906dfd87af3a","Type":"ContainerDied","Data":"695d365282e23d79ab96d852c2a92d6418c185f004a384bf60851c2e6f2d8673"} Nov 21 14:30:12 crc kubenswrapper[4897]: I1121 14:30:12.767238 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d65f699f-cf7rm" event={"ID":"56092cc0-5e4c-47bb-98c1-c1c82d89e0a6","Type":"ContainerDied","Data":"e260d4871100c1508edc71586548bd155eea7f01d400d683d99fc5cf43729b70"} Nov 21 14:30:12 crc kubenswrapper[4897]: I1121 14:30:12.767794 4897 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e260d4871100c1508edc71586548bd155eea7f01d400d683d99fc5cf43729b70" Nov 21 14:30:12 crc kubenswrapper[4897]: I1121 14:30:12.855386 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d65f699f-cf7rm" Nov 21 14:30:12 crc kubenswrapper[4897]: I1121 14:30:12.902215 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vpm8r\" (UniqueName: \"kubernetes.io/projected/56092cc0-5e4c-47bb-98c1-c1c82d89e0a6-kube-api-access-vpm8r\") pod \"56092cc0-5e4c-47bb-98c1-c1c82d89e0a6\" (UID: \"56092cc0-5e4c-47bb-98c1-c1c82d89e0a6\") " Nov 21 14:30:12 crc kubenswrapper[4897]: I1121 14:30:12.902378 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/56092cc0-5e4c-47bb-98c1-c1c82d89e0a6-dns-svc\") pod \"56092cc0-5e4c-47bb-98c1-c1c82d89e0a6\" (UID: \"56092cc0-5e4c-47bb-98c1-c1c82d89e0a6\") " Nov 21 14:30:12 crc kubenswrapper[4897]: I1121 14:30:12.902490 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/56092cc0-5e4c-47bb-98c1-c1c82d89e0a6-config\") pod \"56092cc0-5e4c-47bb-98c1-c1c82d89e0a6\" (UID: \"56092cc0-5e4c-47bb-98c1-c1c82d89e0a6\") " Nov 21 14:30:12 crc kubenswrapper[4897]: I1121 14:30:12.902597 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/56092cc0-5e4c-47bb-98c1-c1c82d89e0a6-ovsdbserver-nb\") pod \"56092cc0-5e4c-47bb-98c1-c1c82d89e0a6\" (UID: \"56092cc0-5e4c-47bb-98c1-c1c82d89e0a6\") " Nov 21 14:30:12 crc kubenswrapper[4897]: I1121 14:30:12.921779 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56092cc0-5e4c-47bb-98c1-c1c82d89e0a6-kube-api-access-vpm8r" (OuterVolumeSpecName: "kube-api-access-vpm8r") pod "56092cc0-5e4c-47bb-98c1-c1c82d89e0a6" (UID: "56092cc0-5e4c-47bb-98c1-c1c82d89e0a6"). InnerVolumeSpecName "kube-api-access-vpm8r". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:30:13 crc kubenswrapper[4897]: I1121 14:30:13.005935 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vpm8r\" (UniqueName: \"kubernetes.io/projected/56092cc0-5e4c-47bb-98c1-c1c82d89e0a6-kube-api-access-vpm8r\") on node \"crc\" DevicePath \"\"" Nov 21 14:30:13 crc kubenswrapper[4897]: I1121 14:30:13.040003 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/56092cc0-5e4c-47bb-98c1-c1c82d89e0a6-config" (OuterVolumeSpecName: "config") pod "56092cc0-5e4c-47bb-98c1-c1c82d89e0a6" (UID: "56092cc0-5e4c-47bb-98c1-c1c82d89e0a6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:30:13 crc kubenswrapper[4897]: I1121 14:30:13.047920 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/56092cc0-5e4c-47bb-98c1-c1c82d89e0a6-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "56092cc0-5e4c-47bb-98c1-c1c82d89e0a6" (UID: "56092cc0-5e4c-47bb-98c1-c1c82d89e0a6"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:30:13 crc kubenswrapper[4897]: I1121 14:30:13.062588 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/56092cc0-5e4c-47bb-98c1-c1c82d89e0a6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "56092cc0-5e4c-47bb-98c1-c1c82d89e0a6" (UID: "56092cc0-5e4c-47bb-98c1-c1c82d89e0a6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:30:13 crc kubenswrapper[4897]: I1121 14:30:13.093313 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29395590-t8htq" Nov 21 14:30:13 crc kubenswrapper[4897]: I1121 14:30:13.109994 4897 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/56092cc0-5e4c-47bb-98c1-c1c82d89e0a6-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 21 14:30:13 crc kubenswrapper[4897]: I1121 14:30:13.110266 4897 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/56092cc0-5e4c-47bb-98c1-c1c82d89e0a6-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 21 14:30:13 crc kubenswrapper[4897]: I1121 14:30:13.110362 4897 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/56092cc0-5e4c-47bb-98c1-c1c82d89e0a6-config\") on node \"crc\" DevicePath \"\"" Nov 21 14:30:13 crc kubenswrapper[4897]: I1121 14:30:13.211198 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/99d4fc7b-7c46-47e7-8bf6-906dfd87af3a-config-volume\") pod \"99d4fc7b-7c46-47e7-8bf6-906dfd87af3a\" (UID: \"99d4fc7b-7c46-47e7-8bf6-906dfd87af3a\") " Nov 21 14:30:13 crc kubenswrapper[4897]: I1121 14:30:13.211253 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hznxg\" (UniqueName: \"kubernetes.io/projected/99d4fc7b-7c46-47e7-8bf6-906dfd87af3a-kube-api-access-hznxg\") pod \"99d4fc7b-7c46-47e7-8bf6-906dfd87af3a\" (UID: \"99d4fc7b-7c46-47e7-8bf6-906dfd87af3a\") " Nov 21 14:30:13 crc kubenswrapper[4897]: I1121 14:30:13.211354 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/99d4fc7b-7c46-47e7-8bf6-906dfd87af3a-secret-volume\") pod \"99d4fc7b-7c46-47e7-8bf6-906dfd87af3a\" (UID: \"99d4fc7b-7c46-47e7-8bf6-906dfd87af3a\") " Nov 21 14:30:13 crc kubenswrapper[4897]: I1121 14:30:13.212027 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/99d4fc7b-7c46-47e7-8bf6-906dfd87af3a-config-volume" (OuterVolumeSpecName: "config-volume") pod "99d4fc7b-7c46-47e7-8bf6-906dfd87af3a" (UID: "99d4fc7b-7c46-47e7-8bf6-906dfd87af3a"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:30:13 crc kubenswrapper[4897]: I1121 14:30:13.214956 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99d4fc7b-7c46-47e7-8bf6-906dfd87af3a-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "99d4fc7b-7c46-47e7-8bf6-906dfd87af3a" (UID: "99d4fc7b-7c46-47e7-8bf6-906dfd87af3a"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:30:13 crc kubenswrapper[4897]: I1121 14:30:13.215041 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99d4fc7b-7c46-47e7-8bf6-906dfd87af3a-kube-api-access-hznxg" (OuterVolumeSpecName: "kube-api-access-hznxg") pod "99d4fc7b-7c46-47e7-8bf6-906dfd87af3a" (UID: "99d4fc7b-7c46-47e7-8bf6-906dfd87af3a"). InnerVolumeSpecName "kube-api-access-hznxg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:30:13 crc kubenswrapper[4897]: I1121 14:30:13.313678 4897 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/99d4fc7b-7c46-47e7-8bf6-906dfd87af3a-config-volume\") on node \"crc\" DevicePath \"\"" Nov 21 14:30:13 crc kubenswrapper[4897]: I1121 14:30:13.313717 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hznxg\" (UniqueName: \"kubernetes.io/projected/99d4fc7b-7c46-47e7-8bf6-906dfd87af3a-kube-api-access-hznxg\") on node \"crc\" DevicePath \"\"" Nov 21 14:30:13 crc kubenswrapper[4897]: I1121 14:30:13.313731 4897 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/99d4fc7b-7c46-47e7-8bf6-906dfd87af3a-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 21 14:30:13 crc kubenswrapper[4897]: I1121 14:30:13.778657 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29395590-t8htq" event={"ID":"99d4fc7b-7c46-47e7-8bf6-906dfd87af3a","Type":"ContainerDied","Data":"e6140710ef46809c71ee984176cd1e0216051635074e59e0dd5c4389e85a52c0"} Nov 21 14:30:13 crc kubenswrapper[4897]: I1121 14:30:13.778976 4897 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e6140710ef46809c71ee984176cd1e0216051635074e59e0dd5c4389e85a52c0" Nov 21 14:30:13 crc kubenswrapper[4897]: I1121 14:30:13.779028 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29395590-t8htq" Nov 21 14:30:13 crc kubenswrapper[4897]: I1121 14:30:13.790562 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-t9nzd" event={"ID":"6d1f75d9-2eaf-44c1-b2d0-dae3a8547433","Type":"ContainerStarted","Data":"470e1f993391940c03765e4498a36ab375ab427bb7aaf8afafebcb169881092c"} Nov 21 14:30:13 crc kubenswrapper[4897]: I1121 14:30:13.796062 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d65f699f-cf7rm" Nov 21 14:30:13 crc kubenswrapper[4897]: I1121 14:30:13.800550 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"63134f01-8622-4fa2-80a3-516572217d0f","Type":"ContainerStarted","Data":"1c5a7fa2c9734b82dcb9d6720ab7bbf083737a44fbfb91e575989ddb05731f24"} Nov 21 14:30:13 crc kubenswrapper[4897]: I1121 14:30:13.800623 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"63134f01-8622-4fa2-80a3-516572217d0f","Type":"ContainerStarted","Data":"3af62fdc3ec62135cadc73335843ac6fdeeda5f5ad8333cc44d548375e602d0e"} Nov 21 14:30:13 crc kubenswrapper[4897]: I1121 14:30:13.800667 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Nov 21 14:30:13 crc kubenswrapper[4897]: I1121 14:30:13.812634 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-t9nzd" podStartSLOduration=2.271636996 podStartE2EDuration="18.812615169s" podCreationTimestamp="2025-11-21 14:29:55 +0000 UTC" firstStartedPulling="2025-11-21 14:29:55.993765869 +0000 UTC m=+1273.278359354" lastFinishedPulling="2025-11-21 14:30:12.534744052 +0000 UTC m=+1289.819337527" observedRunningTime="2025-11-21 14:30:13.804334697 +0000 UTC m=+1291.088928212" watchObservedRunningTime="2025-11-21 14:30:13.812615169 +0000 UTC m=+1291.097208654" Nov 21 14:30:13 crc kubenswrapper[4897]: I1121 14:30:13.837455 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=5.650760645 podStartE2EDuration="15.837435915s" podCreationTimestamp="2025-11-21 14:29:58 +0000 UTC" firstStartedPulling="2025-11-21 14:30:02.339592175 +0000 UTC m=+1279.624185660" lastFinishedPulling="2025-11-21 14:30:12.526267455 +0000 UTC m=+1289.810860930" observedRunningTime="2025-11-21 14:30:13.829922183 +0000 UTC m=+1291.114515648" watchObservedRunningTime="2025-11-21 14:30:13.837435915 +0000 UTC m=+1291.122029390" Nov 21 14:30:13 crc kubenswrapper[4897]: I1121 14:30:13.851789 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d65f699f-cf7rm"] Nov 21 14:30:13 crc kubenswrapper[4897]: I1121 14:30:13.858778 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d65f699f-cf7rm"] Nov 21 14:30:14 crc kubenswrapper[4897]: I1121 14:30:14.119412 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="56092cc0-5e4c-47bb-98c1-c1c82d89e0a6" path="/var/lib/kubelet/pods/56092cc0-5e4c-47bb-98c1-c1c82d89e0a6/volumes" Nov 21 14:30:15 crc kubenswrapper[4897]: I1121 14:30:15.123379 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Nov 21 14:30:15 crc kubenswrapper[4897]: I1121 14:30:15.228561 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Nov 21 14:30:17 crc kubenswrapper[4897]: I1121 14:30:17.390811 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-57d65f699f-cf7rm" podUID="56092cc0-5e4c-47bb-98c1-c1c82d89e0a6" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.147:5353: i/o timeout" Nov 21 14:30:17 crc kubenswrapper[4897]: I1121 14:30:17.622173 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f58d5cd65-2pc2p" podUID="ab48e37a-819f-4f63-8e67-9a2cbeaaeb82" containerName="console" containerID="cri-o://8219d3138b4d11b7091ac5ca0fff7e8fe08b2afd7ab553ce6e30d332a3d5d333" gracePeriod=15 Nov 21 14:30:17 crc kubenswrapper[4897]: I1121 14:30:17.793586 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-2snxf"] Nov 21 14:30:17 crc kubenswrapper[4897]: E1121 14:30:17.794067 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6bf04b77-0ef9-42f0-8a02-8c4c7568e390" containerName="dnsmasq-dns" Nov 21 14:30:17 crc kubenswrapper[4897]: I1121 14:30:17.794087 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="6bf04b77-0ef9-42f0-8a02-8c4c7568e390" containerName="dnsmasq-dns" Nov 21 14:30:17 crc kubenswrapper[4897]: E1121 14:30:17.794101 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9306e5a3-58c4-462b-bad5-0e1f876bf832" containerName="init" Nov 21 14:30:17 crc kubenswrapper[4897]: I1121 14:30:17.794109 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="9306e5a3-58c4-462b-bad5-0e1f876bf832" containerName="init" Nov 21 14:30:17 crc kubenswrapper[4897]: E1121 14:30:17.794127 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56092cc0-5e4c-47bb-98c1-c1c82d89e0a6" containerName="init" Nov 21 14:30:17 crc kubenswrapper[4897]: I1121 14:30:17.794135 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="56092cc0-5e4c-47bb-98c1-c1c82d89e0a6" containerName="init" Nov 21 14:30:17 crc kubenswrapper[4897]: E1121 14:30:17.794152 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56092cc0-5e4c-47bb-98c1-c1c82d89e0a6" containerName="dnsmasq-dns" Nov 21 14:30:17 crc kubenswrapper[4897]: I1121 14:30:17.794161 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="56092cc0-5e4c-47bb-98c1-c1c82d89e0a6" containerName="dnsmasq-dns" Nov 21 14:30:17 crc kubenswrapper[4897]: E1121 14:30:17.794190 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99d4fc7b-7c46-47e7-8bf6-906dfd87af3a" containerName="collect-profiles" Nov 21 14:30:17 crc kubenswrapper[4897]: I1121 14:30:17.794197 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="99d4fc7b-7c46-47e7-8bf6-906dfd87af3a" containerName="collect-profiles" Nov 21 14:30:17 crc kubenswrapper[4897]: E1121 14:30:17.794215 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6bf04b77-0ef9-42f0-8a02-8c4c7568e390" containerName="init" Nov 21 14:30:17 crc kubenswrapper[4897]: I1121 14:30:17.794223 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="6bf04b77-0ef9-42f0-8a02-8c4c7568e390" containerName="init" Nov 21 14:30:17 crc kubenswrapper[4897]: E1121 14:30:17.794242 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9306e5a3-58c4-462b-bad5-0e1f876bf832" containerName="dnsmasq-dns" Nov 21 14:30:17 crc kubenswrapper[4897]: I1121 14:30:17.794250 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="9306e5a3-58c4-462b-bad5-0e1f876bf832" containerName="dnsmasq-dns" Nov 21 14:30:17 crc kubenswrapper[4897]: I1121 14:30:17.794497 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="9306e5a3-58c4-462b-bad5-0e1f876bf832" containerName="dnsmasq-dns" Nov 21 14:30:17 crc kubenswrapper[4897]: I1121 14:30:17.794532 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="6bf04b77-0ef9-42f0-8a02-8c4c7568e390" containerName="dnsmasq-dns" Nov 21 14:30:17 crc kubenswrapper[4897]: I1121 14:30:17.794551 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="99d4fc7b-7c46-47e7-8bf6-906dfd87af3a" containerName="collect-profiles" Nov 21 14:30:17 crc kubenswrapper[4897]: I1121 14:30:17.794562 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="56092cc0-5e4c-47bb-98c1-c1c82d89e0a6" containerName="dnsmasq-dns" Nov 21 14:30:17 crc kubenswrapper[4897]: I1121 14:30:17.795367 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-2snxf" Nov 21 14:30:17 crc kubenswrapper[4897]: I1121 14:30:17.816194 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gwdzt\" (UniqueName: \"kubernetes.io/projected/d7a5e302-fb61-41b2-a7d2-ff02160c6679-kube-api-access-gwdzt\") pod \"keystone-db-create-2snxf\" (UID: \"d7a5e302-fb61-41b2-a7d2-ff02160c6679\") " pod="openstack/keystone-db-create-2snxf" Nov 21 14:30:17 crc kubenswrapper[4897]: I1121 14:30:17.816388 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d7a5e302-fb61-41b2-a7d2-ff02160c6679-operator-scripts\") pod \"keystone-db-create-2snxf\" (UID: \"d7a5e302-fb61-41b2-a7d2-ff02160c6679\") " pod="openstack/keystone-db-create-2snxf" Nov 21 14:30:17 crc kubenswrapper[4897]: I1121 14:30:17.837072 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-2snxf"] Nov 21 14:30:17 crc kubenswrapper[4897]: I1121 14:30:17.862431 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-b8bf-account-create-x2vdt"] Nov 21 14:30:17 crc kubenswrapper[4897]: I1121 14:30:17.866564 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-b8bf-account-create-x2vdt" Nov 21 14:30:17 crc kubenswrapper[4897]: I1121 14:30:17.870611 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Nov 21 14:30:17 crc kubenswrapper[4897]: I1121 14:30:17.898860 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-b8bf-account-create-x2vdt"] Nov 21 14:30:17 crc kubenswrapper[4897]: I1121 14:30:17.918460 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/980c0de3-d435-439f-8477-0dc68fbaf7f0-operator-scripts\") pod \"keystone-b8bf-account-create-x2vdt\" (UID: \"980c0de3-d435-439f-8477-0dc68fbaf7f0\") " pod="openstack/keystone-b8bf-account-create-x2vdt" Nov 21 14:30:17 crc kubenswrapper[4897]: I1121 14:30:17.918585 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pdplw\" (UniqueName: \"kubernetes.io/projected/980c0de3-d435-439f-8477-0dc68fbaf7f0-kube-api-access-pdplw\") pod \"keystone-b8bf-account-create-x2vdt\" (UID: \"980c0de3-d435-439f-8477-0dc68fbaf7f0\") " pod="openstack/keystone-b8bf-account-create-x2vdt" Nov 21 14:30:17 crc kubenswrapper[4897]: I1121 14:30:17.918701 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d7a5e302-fb61-41b2-a7d2-ff02160c6679-operator-scripts\") pod \"keystone-db-create-2snxf\" (UID: \"d7a5e302-fb61-41b2-a7d2-ff02160c6679\") " pod="openstack/keystone-db-create-2snxf" Nov 21 14:30:17 crc kubenswrapper[4897]: I1121 14:30:17.918882 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gwdzt\" (UniqueName: \"kubernetes.io/projected/d7a5e302-fb61-41b2-a7d2-ff02160c6679-kube-api-access-gwdzt\") pod \"keystone-db-create-2snxf\" (UID: \"d7a5e302-fb61-41b2-a7d2-ff02160c6679\") " pod="openstack/keystone-db-create-2snxf" Nov 21 14:30:17 crc kubenswrapper[4897]: I1121 14:30:17.924482 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d7a5e302-fb61-41b2-a7d2-ff02160c6679-operator-scripts\") pod \"keystone-db-create-2snxf\" (UID: \"d7a5e302-fb61-41b2-a7d2-ff02160c6679\") " pod="openstack/keystone-db-create-2snxf" Nov 21 14:30:17 crc kubenswrapper[4897]: I1121 14:30:17.947085 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gwdzt\" (UniqueName: \"kubernetes.io/projected/d7a5e302-fb61-41b2-a7d2-ff02160c6679-kube-api-access-gwdzt\") pod \"keystone-db-create-2snxf\" (UID: \"d7a5e302-fb61-41b2-a7d2-ff02160c6679\") " pod="openstack/keystone-db-create-2snxf" Nov 21 14:30:18 crc kubenswrapper[4897]: I1121 14:30:18.020913 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/980c0de3-d435-439f-8477-0dc68fbaf7f0-operator-scripts\") pod \"keystone-b8bf-account-create-x2vdt\" (UID: \"980c0de3-d435-439f-8477-0dc68fbaf7f0\") " pod="openstack/keystone-b8bf-account-create-x2vdt" Nov 21 14:30:18 crc kubenswrapper[4897]: I1121 14:30:18.021058 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pdplw\" (UniqueName: \"kubernetes.io/projected/980c0de3-d435-439f-8477-0dc68fbaf7f0-kube-api-access-pdplw\") pod \"keystone-b8bf-account-create-x2vdt\" (UID: \"980c0de3-d435-439f-8477-0dc68fbaf7f0\") " pod="openstack/keystone-b8bf-account-create-x2vdt" Nov 21 14:30:18 crc kubenswrapper[4897]: I1121 14:30:18.022497 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-sbnf2"] Nov 21 14:30:18 crc kubenswrapper[4897]: I1121 14:30:18.022902 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/980c0de3-d435-439f-8477-0dc68fbaf7f0-operator-scripts\") pod \"keystone-b8bf-account-create-x2vdt\" (UID: \"980c0de3-d435-439f-8477-0dc68fbaf7f0\") " pod="openstack/keystone-b8bf-account-create-x2vdt" Nov 21 14:30:18 crc kubenswrapper[4897]: I1121 14:30:18.024089 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-sbnf2" Nov 21 14:30:18 crc kubenswrapper[4897]: I1121 14:30:18.034425 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-sbnf2"] Nov 21 14:30:18 crc kubenswrapper[4897]: I1121 14:30:18.044033 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pdplw\" (UniqueName: \"kubernetes.io/projected/980c0de3-d435-439f-8477-0dc68fbaf7f0-kube-api-access-pdplw\") pod \"keystone-b8bf-account-create-x2vdt\" (UID: \"980c0de3-d435-439f-8477-0dc68fbaf7f0\") " pod="openstack/keystone-b8bf-account-create-x2vdt" Nov 21 14:30:18 crc kubenswrapper[4897]: I1121 14:30:18.047630 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-b2f5-account-create-69ls6"] Nov 21 14:30:18 crc kubenswrapper[4897]: I1121 14:30:18.049198 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-b2f5-account-create-69ls6" Nov 21 14:30:18 crc kubenswrapper[4897]: I1121 14:30:18.051350 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Nov 21 14:30:18 crc kubenswrapper[4897]: I1121 14:30:18.057330 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-b2f5-account-create-69ls6"] Nov 21 14:30:18 crc kubenswrapper[4897]: I1121 14:30:18.120944 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-2snxf" Nov 21 14:30:18 crc kubenswrapper[4897]: I1121 14:30:18.124197 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/34a64a5d-5b9d-4aac-be3e-96edc721fa3b-operator-scripts\") pod \"placement-b2f5-account-create-69ls6\" (UID: \"34a64a5d-5b9d-4aac-be3e-96edc721fa3b\") " pod="openstack/placement-b2f5-account-create-69ls6" Nov 21 14:30:18 crc kubenswrapper[4897]: I1121 14:30:18.124269 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-994hn\" (UniqueName: \"kubernetes.io/projected/d3441939-1c13-4b55-a129-a39f4bed36cd-kube-api-access-994hn\") pod \"placement-db-create-sbnf2\" (UID: \"d3441939-1c13-4b55-a129-a39f4bed36cd\") " pod="openstack/placement-db-create-sbnf2" Nov 21 14:30:18 crc kubenswrapper[4897]: I1121 14:30:18.124299 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d3441939-1c13-4b55-a129-a39f4bed36cd-operator-scripts\") pod \"placement-db-create-sbnf2\" (UID: \"d3441939-1c13-4b55-a129-a39f4bed36cd\") " pod="openstack/placement-db-create-sbnf2" Nov 21 14:30:18 crc kubenswrapper[4897]: I1121 14:30:18.124334 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k9dr2\" (UniqueName: \"kubernetes.io/projected/34a64a5d-5b9d-4aac-be3e-96edc721fa3b-kube-api-access-k9dr2\") pod \"placement-b2f5-account-create-69ls6\" (UID: \"34a64a5d-5b9d-4aac-be3e-96edc721fa3b\") " pod="openstack/placement-b2f5-account-create-69ls6" Nov 21 14:30:18 crc kubenswrapper[4897]: I1121 14:30:18.182925 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-b8bf-account-create-x2vdt" Nov 21 14:30:18 crc kubenswrapper[4897]: I1121 14:30:18.225919 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/34a64a5d-5b9d-4aac-be3e-96edc721fa3b-operator-scripts\") pod \"placement-b2f5-account-create-69ls6\" (UID: \"34a64a5d-5b9d-4aac-be3e-96edc721fa3b\") " pod="openstack/placement-b2f5-account-create-69ls6" Nov 21 14:30:18 crc kubenswrapper[4897]: I1121 14:30:18.225990 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-994hn\" (UniqueName: \"kubernetes.io/projected/d3441939-1c13-4b55-a129-a39f4bed36cd-kube-api-access-994hn\") pod \"placement-db-create-sbnf2\" (UID: \"d3441939-1c13-4b55-a129-a39f4bed36cd\") " pod="openstack/placement-db-create-sbnf2" Nov 21 14:30:18 crc kubenswrapper[4897]: I1121 14:30:18.226015 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d3441939-1c13-4b55-a129-a39f4bed36cd-operator-scripts\") pod \"placement-db-create-sbnf2\" (UID: \"d3441939-1c13-4b55-a129-a39f4bed36cd\") " pod="openstack/placement-db-create-sbnf2" Nov 21 14:30:18 crc kubenswrapper[4897]: I1121 14:30:18.226046 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k9dr2\" (UniqueName: \"kubernetes.io/projected/34a64a5d-5b9d-4aac-be3e-96edc721fa3b-kube-api-access-k9dr2\") pod \"placement-b2f5-account-create-69ls6\" (UID: \"34a64a5d-5b9d-4aac-be3e-96edc721fa3b\") " pod="openstack/placement-b2f5-account-create-69ls6" Nov 21 14:30:18 crc kubenswrapper[4897]: I1121 14:30:18.226965 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/34a64a5d-5b9d-4aac-be3e-96edc721fa3b-operator-scripts\") pod \"placement-b2f5-account-create-69ls6\" (UID: \"34a64a5d-5b9d-4aac-be3e-96edc721fa3b\") " pod="openstack/placement-b2f5-account-create-69ls6" Nov 21 14:30:18 crc kubenswrapper[4897]: I1121 14:30:18.227061 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d3441939-1c13-4b55-a129-a39f4bed36cd-operator-scripts\") pod \"placement-db-create-sbnf2\" (UID: \"d3441939-1c13-4b55-a129-a39f4bed36cd\") " pod="openstack/placement-db-create-sbnf2" Nov 21 14:30:18 crc kubenswrapper[4897]: I1121 14:30:18.242742 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k9dr2\" (UniqueName: \"kubernetes.io/projected/34a64a5d-5b9d-4aac-be3e-96edc721fa3b-kube-api-access-k9dr2\") pod \"placement-b2f5-account-create-69ls6\" (UID: \"34a64a5d-5b9d-4aac-be3e-96edc721fa3b\") " pod="openstack/placement-b2f5-account-create-69ls6" Nov 21 14:30:18 crc kubenswrapper[4897]: I1121 14:30:18.247688 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-994hn\" (UniqueName: \"kubernetes.io/projected/d3441939-1c13-4b55-a129-a39f4bed36cd-kube-api-access-994hn\") pod \"placement-db-create-sbnf2\" (UID: \"d3441939-1c13-4b55-a129-a39f4bed36cd\") " pod="openstack/placement-db-create-sbnf2" Nov 21 14:30:18 crc kubenswrapper[4897]: I1121 14:30:18.373372 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-rkm52"] Nov 21 14:30:18 crc kubenswrapper[4897]: I1121 14:30:18.374675 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-rkm52" Nov 21 14:30:18 crc kubenswrapper[4897]: I1121 14:30:18.388794 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-rkm52"] Nov 21 14:30:18 crc kubenswrapper[4897]: I1121 14:30:18.401474 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-sbnf2" Nov 21 14:30:18 crc kubenswrapper[4897]: I1121 14:30:18.412394 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-b2f5-account-create-69ls6" Nov 21 14:30:18 crc kubenswrapper[4897]: I1121 14:30:18.431936 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xckvk\" (UniqueName: \"kubernetes.io/projected/7501a167-0655-4ad4-bb04-c597a81359ba-kube-api-access-xckvk\") pod \"glance-db-create-rkm52\" (UID: \"7501a167-0655-4ad4-bb04-c597a81359ba\") " pod="openstack/glance-db-create-rkm52" Nov 21 14:30:18 crc kubenswrapper[4897]: I1121 14:30:18.432155 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7501a167-0655-4ad4-bb04-c597a81359ba-operator-scripts\") pod \"glance-db-create-rkm52\" (UID: \"7501a167-0655-4ad4-bb04-c597a81359ba\") " pod="openstack/glance-db-create-rkm52" Nov 21 14:30:18 crc kubenswrapper[4897]: I1121 14:30:18.497096 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-376f-account-create-q6xsf"] Nov 21 14:30:18 crc kubenswrapper[4897]: I1121 14:30:18.498325 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-376f-account-create-q6xsf" Nov 21 14:30:18 crc kubenswrapper[4897]: I1121 14:30:18.503934 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Nov 21 14:30:18 crc kubenswrapper[4897]: I1121 14:30:18.509076 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-376f-account-create-q6xsf"] Nov 21 14:30:18 crc kubenswrapper[4897]: I1121 14:30:18.534390 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b6lg4\" (UniqueName: \"kubernetes.io/projected/994c36f9-336e-49dc-b45d-0eb4bfa0b686-kube-api-access-b6lg4\") pod \"glance-376f-account-create-q6xsf\" (UID: \"994c36f9-336e-49dc-b45d-0eb4bfa0b686\") " pod="openstack/glance-376f-account-create-q6xsf" Nov 21 14:30:18 crc kubenswrapper[4897]: I1121 14:30:18.534531 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xckvk\" (UniqueName: \"kubernetes.io/projected/7501a167-0655-4ad4-bb04-c597a81359ba-kube-api-access-xckvk\") pod \"glance-db-create-rkm52\" (UID: \"7501a167-0655-4ad4-bb04-c597a81359ba\") " pod="openstack/glance-db-create-rkm52" Nov 21 14:30:18 crc kubenswrapper[4897]: I1121 14:30:18.534718 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/994c36f9-336e-49dc-b45d-0eb4bfa0b686-operator-scripts\") pod \"glance-376f-account-create-q6xsf\" (UID: \"994c36f9-336e-49dc-b45d-0eb4bfa0b686\") " pod="openstack/glance-376f-account-create-q6xsf" Nov 21 14:30:18 crc kubenswrapper[4897]: I1121 14:30:18.534748 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7501a167-0655-4ad4-bb04-c597a81359ba-operator-scripts\") pod \"glance-db-create-rkm52\" (UID: \"7501a167-0655-4ad4-bb04-c597a81359ba\") " pod="openstack/glance-db-create-rkm52" Nov 21 14:30:18 crc kubenswrapper[4897]: I1121 14:30:18.535525 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7501a167-0655-4ad4-bb04-c597a81359ba-operator-scripts\") pod \"glance-db-create-rkm52\" (UID: \"7501a167-0655-4ad4-bb04-c597a81359ba\") " pod="openstack/glance-db-create-rkm52" Nov 21 14:30:18 crc kubenswrapper[4897]: I1121 14:30:18.565147 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xckvk\" (UniqueName: \"kubernetes.io/projected/7501a167-0655-4ad4-bb04-c597a81359ba-kube-api-access-xckvk\") pod \"glance-db-create-rkm52\" (UID: \"7501a167-0655-4ad4-bb04-c597a81359ba\") " pod="openstack/glance-db-create-rkm52" Nov 21 14:30:18 crc kubenswrapper[4897]: I1121 14:30:18.637938 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b6lg4\" (UniqueName: \"kubernetes.io/projected/994c36f9-336e-49dc-b45d-0eb4bfa0b686-kube-api-access-b6lg4\") pod \"glance-376f-account-create-q6xsf\" (UID: \"994c36f9-336e-49dc-b45d-0eb4bfa0b686\") " pod="openstack/glance-376f-account-create-q6xsf" Nov 21 14:30:18 crc kubenswrapper[4897]: I1121 14:30:18.638558 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/994c36f9-336e-49dc-b45d-0eb4bfa0b686-operator-scripts\") pod \"glance-376f-account-create-q6xsf\" (UID: \"994c36f9-336e-49dc-b45d-0eb4bfa0b686\") " pod="openstack/glance-376f-account-create-q6xsf" Nov 21 14:30:18 crc kubenswrapper[4897]: I1121 14:30:18.639184 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/994c36f9-336e-49dc-b45d-0eb4bfa0b686-operator-scripts\") pod \"glance-376f-account-create-q6xsf\" (UID: \"994c36f9-336e-49dc-b45d-0eb4bfa0b686\") " pod="openstack/glance-376f-account-create-q6xsf" Nov 21 14:30:18 crc kubenswrapper[4897]: I1121 14:30:18.659399 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b6lg4\" (UniqueName: \"kubernetes.io/projected/994c36f9-336e-49dc-b45d-0eb4bfa0b686-kube-api-access-b6lg4\") pod \"glance-376f-account-create-q6xsf\" (UID: \"994c36f9-336e-49dc-b45d-0eb4bfa0b686\") " pod="openstack/glance-376f-account-create-q6xsf" Nov 21 14:30:18 crc kubenswrapper[4897]: I1121 14:30:18.710436 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-rkm52" Nov 21 14:30:18 crc kubenswrapper[4897]: I1121 14:30:18.816859 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-376f-account-create-q6xsf" Nov 21 14:30:18 crc kubenswrapper[4897]: I1121 14:30:18.875318 4897 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f58d5cd65-2pc2p_ab48e37a-819f-4f63-8e67-9a2cbeaaeb82/console/0.log" Nov 21 14:30:18 crc kubenswrapper[4897]: I1121 14:30:18.875371 4897 generic.go:334] "Generic (PLEG): container finished" podID="ab48e37a-819f-4f63-8e67-9a2cbeaaeb82" containerID="8219d3138b4d11b7091ac5ca0fff7e8fe08b2afd7ab553ce6e30d332a3d5d333" exitCode=2 Nov 21 14:30:18 crc kubenswrapper[4897]: I1121 14:30:18.875401 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f58d5cd65-2pc2p" event={"ID":"ab48e37a-819f-4f63-8e67-9a2cbeaaeb82","Type":"ContainerDied","Data":"8219d3138b4d11b7091ac5ca0fff7e8fe08b2afd7ab553ce6e30d332a3d5d333"} Nov 21 14:30:18 crc kubenswrapper[4897]: I1121 14:30:18.906644 4897 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f58d5cd65-2pc2p_ab48e37a-819f-4f63-8e67-9a2cbeaaeb82/console/0.log" Nov 21 14:30:18 crc kubenswrapper[4897]: I1121 14:30:18.906701 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f58d5cd65-2pc2p" Nov 21 14:30:18 crc kubenswrapper[4897]: I1121 14:30:18.944728 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ab48e37a-819f-4f63-8e67-9a2cbeaaeb82-trusted-ca-bundle\") pod \"ab48e37a-819f-4f63-8e67-9a2cbeaaeb82\" (UID: \"ab48e37a-819f-4f63-8e67-9a2cbeaaeb82\") " Nov 21 14:30:18 crc kubenswrapper[4897]: I1121 14:30:18.944991 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ab48e37a-819f-4f63-8e67-9a2cbeaaeb82-oauth-serving-cert\") pod \"ab48e37a-819f-4f63-8e67-9a2cbeaaeb82\" (UID: \"ab48e37a-819f-4f63-8e67-9a2cbeaaeb82\") " Nov 21 14:30:18 crc kubenswrapper[4897]: I1121 14:30:18.945057 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ab48e37a-819f-4f63-8e67-9a2cbeaaeb82-console-serving-cert\") pod \"ab48e37a-819f-4f63-8e67-9a2cbeaaeb82\" (UID: \"ab48e37a-819f-4f63-8e67-9a2cbeaaeb82\") " Nov 21 14:30:18 crc kubenswrapper[4897]: I1121 14:30:18.945096 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xdtcg\" (UniqueName: \"kubernetes.io/projected/ab48e37a-819f-4f63-8e67-9a2cbeaaeb82-kube-api-access-xdtcg\") pod \"ab48e37a-819f-4f63-8e67-9a2cbeaaeb82\" (UID: \"ab48e37a-819f-4f63-8e67-9a2cbeaaeb82\") " Nov 21 14:30:18 crc kubenswrapper[4897]: I1121 14:30:18.945208 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ab48e37a-819f-4f63-8e67-9a2cbeaaeb82-service-ca\") pod \"ab48e37a-819f-4f63-8e67-9a2cbeaaeb82\" (UID: \"ab48e37a-819f-4f63-8e67-9a2cbeaaeb82\") " Nov 21 14:30:18 crc kubenswrapper[4897]: I1121 14:30:18.945321 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ab48e37a-819f-4f63-8e67-9a2cbeaaeb82-console-config\") pod \"ab48e37a-819f-4f63-8e67-9a2cbeaaeb82\" (UID: \"ab48e37a-819f-4f63-8e67-9a2cbeaaeb82\") " Nov 21 14:30:18 crc kubenswrapper[4897]: I1121 14:30:18.945369 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ab48e37a-819f-4f63-8e67-9a2cbeaaeb82-console-oauth-config\") pod \"ab48e37a-819f-4f63-8e67-9a2cbeaaeb82\" (UID: \"ab48e37a-819f-4f63-8e67-9a2cbeaaeb82\") " Nov 21 14:30:18 crc kubenswrapper[4897]: I1121 14:30:18.947875 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ab48e37a-819f-4f63-8e67-9a2cbeaaeb82-service-ca" (OuterVolumeSpecName: "service-ca") pod "ab48e37a-819f-4f63-8e67-9a2cbeaaeb82" (UID: "ab48e37a-819f-4f63-8e67-9a2cbeaaeb82"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:30:18 crc kubenswrapper[4897]: I1121 14:30:18.948105 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ab48e37a-819f-4f63-8e67-9a2cbeaaeb82-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "ab48e37a-819f-4f63-8e67-9a2cbeaaeb82" (UID: "ab48e37a-819f-4f63-8e67-9a2cbeaaeb82"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:30:18 crc kubenswrapper[4897]: I1121 14:30:18.948398 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ab48e37a-819f-4f63-8e67-9a2cbeaaeb82-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "ab48e37a-819f-4f63-8e67-9a2cbeaaeb82" (UID: "ab48e37a-819f-4f63-8e67-9a2cbeaaeb82"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:30:18 crc kubenswrapper[4897]: I1121 14:30:18.949992 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ab48e37a-819f-4f63-8e67-9a2cbeaaeb82-console-config" (OuterVolumeSpecName: "console-config") pod "ab48e37a-819f-4f63-8e67-9a2cbeaaeb82" (UID: "ab48e37a-819f-4f63-8e67-9a2cbeaaeb82"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:30:18 crc kubenswrapper[4897]: I1121 14:30:18.950242 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab48e37a-819f-4f63-8e67-9a2cbeaaeb82-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "ab48e37a-819f-4f63-8e67-9a2cbeaaeb82" (UID: "ab48e37a-819f-4f63-8e67-9a2cbeaaeb82"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:30:18 crc kubenswrapper[4897]: I1121 14:30:18.956802 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab48e37a-819f-4f63-8e67-9a2cbeaaeb82-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "ab48e37a-819f-4f63-8e67-9a2cbeaaeb82" (UID: "ab48e37a-819f-4f63-8e67-9a2cbeaaeb82"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:30:18 crc kubenswrapper[4897]: I1121 14:30:18.957734 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab48e37a-819f-4f63-8e67-9a2cbeaaeb82-kube-api-access-xdtcg" (OuterVolumeSpecName: "kube-api-access-xdtcg") pod "ab48e37a-819f-4f63-8e67-9a2cbeaaeb82" (UID: "ab48e37a-819f-4f63-8e67-9a2cbeaaeb82"). InnerVolumeSpecName "kube-api-access-xdtcg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:30:19 crc kubenswrapper[4897]: I1121 14:30:19.049616 4897 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ab48e37a-819f-4f63-8e67-9a2cbeaaeb82-console-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 21 14:30:19 crc kubenswrapper[4897]: I1121 14:30:19.049650 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xdtcg\" (UniqueName: \"kubernetes.io/projected/ab48e37a-819f-4f63-8e67-9a2cbeaaeb82-kube-api-access-xdtcg\") on node \"crc\" DevicePath \"\"" Nov 21 14:30:19 crc kubenswrapper[4897]: I1121 14:30:19.049660 4897 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ab48e37a-819f-4f63-8e67-9a2cbeaaeb82-service-ca\") on node \"crc\" DevicePath \"\"" Nov 21 14:30:19 crc kubenswrapper[4897]: I1121 14:30:19.049669 4897 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ab48e37a-819f-4f63-8e67-9a2cbeaaeb82-console-config\") on node \"crc\" DevicePath \"\"" Nov 21 14:30:19 crc kubenswrapper[4897]: I1121 14:30:19.049677 4897 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ab48e37a-819f-4f63-8e67-9a2cbeaaeb82-console-oauth-config\") on node \"crc\" DevicePath \"\"" Nov 21 14:30:19 crc kubenswrapper[4897]: I1121 14:30:19.049685 4897 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ab48e37a-819f-4f63-8e67-9a2cbeaaeb82-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 21 14:30:19 crc kubenswrapper[4897]: I1121 14:30:19.049693 4897 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ab48e37a-819f-4f63-8e67-9a2cbeaaeb82-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 21 14:30:19 crc kubenswrapper[4897]: I1121 14:30:19.107649 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-2snxf"] Nov 21 14:30:19 crc kubenswrapper[4897]: W1121 14:30:19.124728 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd7a5e302_fb61_41b2_a7d2_ff02160c6679.slice/crio-da85ac06dc94cbce05fef034af847bddbae25cc65707caf1f487f6175d4ecae4 WatchSource:0}: Error finding container da85ac06dc94cbce05fef034af847bddbae25cc65707caf1f487f6175d4ecae4: Status 404 returned error can't find the container with id da85ac06dc94cbce05fef034af847bddbae25cc65707caf1f487f6175d4ecae4 Nov 21 14:30:19 crc kubenswrapper[4897]: I1121 14:30:19.366092 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-b2f5-account-create-69ls6"] Nov 21 14:30:19 crc kubenswrapper[4897]: I1121 14:30:19.377570 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-b8bf-account-create-x2vdt"] Nov 21 14:30:20 crc kubenswrapper[4897]: I1121 14:30:19.417071 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-sbnf2"] Nov 21 14:30:20 crc kubenswrapper[4897]: W1121 14:30:19.446907 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd3441939_1c13_4b55_a129_a39f4bed36cd.slice/crio-ec0ce44abdd2ca8c499480896781a88461939aa9a7047d23589a8ddc6e3bb626 WatchSource:0}: Error finding container ec0ce44abdd2ca8c499480896781a88461939aa9a7047d23589a8ddc6e3bb626: Status 404 returned error can't find the container with id ec0ce44abdd2ca8c499480896781a88461939aa9a7047d23589a8ddc6e3bb626 Nov 21 14:30:20 crc kubenswrapper[4897]: I1121 14:30:19.550957 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-rkm52"] Nov 21 14:30:20 crc kubenswrapper[4897]: I1121 14:30:19.562885 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-376f-account-create-q6xsf"] Nov 21 14:30:20 crc kubenswrapper[4897]: E1121 14:30:19.782254 4897 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd7a5e302_fb61_41b2_a7d2_ff02160c6679.slice/crio-conmon-15cd2591cd330627ec6bc88cb3b35efc10d63e1d88ab6a3b6123dcf88f489ca8.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd7a5e302_fb61_41b2_a7d2_ff02160c6679.slice/crio-15cd2591cd330627ec6bc88cb3b35efc10d63e1d88ab6a3b6123dcf88f489ca8.scope\": RecentStats: unable to find data in memory cache]" Nov 21 14:30:20 crc kubenswrapper[4897]: I1121 14:30:19.892953 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-sbnf2" event={"ID":"d3441939-1c13-4b55-a129-a39f4bed36cd","Type":"ContainerStarted","Data":"f92ea24bf1f109608a14fbb251bce2cef26cc268df4c3daa8daeb16e4d2f599d"} Nov 21 14:30:20 crc kubenswrapper[4897]: I1121 14:30:19.892989 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-sbnf2" event={"ID":"d3441939-1c13-4b55-a129-a39f4bed36cd","Type":"ContainerStarted","Data":"ec0ce44abdd2ca8c499480896781a88461939aa9a7047d23589a8ddc6e3bb626"} Nov 21 14:30:20 crc kubenswrapper[4897]: I1121 14:30:19.896316 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-b8bf-account-create-x2vdt" event={"ID":"980c0de3-d435-439f-8477-0dc68fbaf7f0","Type":"ContainerStarted","Data":"0a09a0b4d2bf00b873dd3a26366c3cb500c70dfa74206e2efd43e17e17491e68"} Nov 21 14:30:20 crc kubenswrapper[4897]: I1121 14:30:19.896338 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-b8bf-account-create-x2vdt" event={"ID":"980c0de3-d435-439f-8477-0dc68fbaf7f0","Type":"ContainerStarted","Data":"b1aaf659a72e0f1f9547ceac86f946bc9644b1473e0559ddc3e90dfa18948ceb"} Nov 21 14:30:20 crc kubenswrapper[4897]: I1121 14:30:19.897771 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-rkm52" event={"ID":"7501a167-0655-4ad4-bb04-c597a81359ba","Type":"ContainerStarted","Data":"1675902aaf6ae6196f179d75e509846f5abd6d1664aa148b6578ceea932c4544"} Nov 21 14:30:20 crc kubenswrapper[4897]: I1121 14:30:19.897812 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-rkm52" event={"ID":"7501a167-0655-4ad4-bb04-c597a81359ba","Type":"ContainerStarted","Data":"d10acce71eb6e650686d1b6fd92cb5b47a393460cb71d96141603be290d78802"} Nov 21 14:30:20 crc kubenswrapper[4897]: I1121 14:30:19.898842 4897 generic.go:334] "Generic (PLEG): container finished" podID="d7a5e302-fb61-41b2-a7d2-ff02160c6679" containerID="15cd2591cd330627ec6bc88cb3b35efc10d63e1d88ab6a3b6123dcf88f489ca8" exitCode=0 Nov 21 14:30:20 crc kubenswrapper[4897]: I1121 14:30:19.898889 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-2snxf" event={"ID":"d7a5e302-fb61-41b2-a7d2-ff02160c6679","Type":"ContainerDied","Data":"15cd2591cd330627ec6bc88cb3b35efc10d63e1d88ab6a3b6123dcf88f489ca8"} Nov 21 14:30:20 crc kubenswrapper[4897]: I1121 14:30:19.898911 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-2snxf" event={"ID":"d7a5e302-fb61-41b2-a7d2-ff02160c6679","Type":"ContainerStarted","Data":"da85ac06dc94cbce05fef034af847bddbae25cc65707caf1f487f6175d4ecae4"} Nov 21 14:30:20 crc kubenswrapper[4897]: I1121 14:30:19.903185 4897 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f58d5cd65-2pc2p_ab48e37a-819f-4f63-8e67-9a2cbeaaeb82/console/0.log" Nov 21 14:30:20 crc kubenswrapper[4897]: I1121 14:30:19.903660 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f58d5cd65-2pc2p" Nov 21 14:30:20 crc kubenswrapper[4897]: I1121 14:30:19.904243 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f58d5cd65-2pc2p" event={"ID":"ab48e37a-819f-4f63-8e67-9a2cbeaaeb82","Type":"ContainerDied","Data":"8dfce227bc2bae973871ee709fef9ec9da02fa04602c7a887424930ec0991bb9"} Nov 21 14:30:20 crc kubenswrapper[4897]: I1121 14:30:19.904288 4897 scope.go:117] "RemoveContainer" containerID="8219d3138b4d11b7091ac5ca0fff7e8fe08b2afd7ab553ce6e30d332a3d5d333" Nov 21 14:30:20 crc kubenswrapper[4897]: I1121 14:30:19.911113 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-b2f5-account-create-69ls6" event={"ID":"34a64a5d-5b9d-4aac-be3e-96edc721fa3b","Type":"ContainerStarted","Data":"860f4995b77cd310712c0190183323bddec94b2b3f7a6bf682efd550850a9716"} Nov 21 14:30:20 crc kubenswrapper[4897]: I1121 14:30:19.911149 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-b2f5-account-create-69ls6" event={"ID":"34a64a5d-5b9d-4aac-be3e-96edc721fa3b","Type":"ContainerStarted","Data":"b894e1ce596ce8e4b5c070cebe296697fb3efa4e874a9b64ed45a7dc5cd4f648"} Nov 21 14:30:20 crc kubenswrapper[4897]: I1121 14:30:19.916807 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"91aa2270-92a7-4af4-9a92-141ea2b24ee1","Type":"ContainerStarted","Data":"d4cba41371733c06d0f9282cb6e29395e6f232e62a3ce987fe67b0eb3d37cb98"} Nov 21 14:30:20 crc kubenswrapper[4897]: I1121 14:30:19.920348 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-376f-account-create-q6xsf" event={"ID":"994c36f9-336e-49dc-b45d-0eb4bfa0b686","Type":"ContainerStarted","Data":"9522822e4d3514856f080fe94bb45b833c81336e3fd4d133f5c86ec83ec51d64"} Nov 21 14:30:20 crc kubenswrapper[4897]: I1121 14:30:19.920377 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-376f-account-create-q6xsf" event={"ID":"994c36f9-336e-49dc-b45d-0eb4bfa0b686","Type":"ContainerStarted","Data":"988d66d02d23611bbf0f2e52b8d01eb5dc8312e548ca15e48ee3bd7cc70c4977"} Nov 21 14:30:20 crc kubenswrapper[4897]: I1121 14:30:19.930601 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-create-sbnf2" podStartSLOduration=2.930577902 podStartE2EDuration="2.930577902s" podCreationTimestamp="2025-11-21 14:30:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:30:19.910394431 +0000 UTC m=+1297.194987916" watchObservedRunningTime="2025-11-21 14:30:19.930577902 +0000 UTC m=+1297.215171377" Nov 21 14:30:20 crc kubenswrapper[4897]: I1121 14:30:19.943115 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-create-rkm52" podStartSLOduration=1.943068207 podStartE2EDuration="1.943068207s" podCreationTimestamp="2025-11-21 14:30:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:30:19.924635703 +0000 UTC m=+1297.209229188" watchObservedRunningTime="2025-11-21 14:30:19.943068207 +0000 UTC m=+1297.227661672" Nov 21 14:30:20 crc kubenswrapper[4897]: I1121 14:30:19.990568 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-376f-account-create-q6xsf" podStartSLOduration=1.990488079 podStartE2EDuration="1.990488079s" podCreationTimestamp="2025-11-21 14:30:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:30:19.982866015 +0000 UTC m=+1297.267459510" watchObservedRunningTime="2025-11-21 14:30:19.990488079 +0000 UTC m=+1297.275081564" Nov 21 14:30:20 crc kubenswrapper[4897]: I1121 14:30:20.126630 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mysqld-exporter-openstack-db-create-z8nrd"] Nov 21 14:30:20 crc kubenswrapper[4897]: E1121 14:30:20.127441 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab48e37a-819f-4f63-8e67-9a2cbeaaeb82" containerName="console" Nov 21 14:30:20 crc kubenswrapper[4897]: I1121 14:30:20.127457 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab48e37a-819f-4f63-8e67-9a2cbeaaeb82" containerName="console" Nov 21 14:30:20 crc kubenswrapper[4897]: I1121 14:30:20.127732 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab48e37a-819f-4f63-8e67-9a2cbeaaeb82" containerName="console" Nov 21 14:30:20 crc kubenswrapper[4897]: I1121 14:30:20.128449 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-openstack-db-create-z8nrd"] Nov 21 14:30:20 crc kubenswrapper[4897]: I1121 14:30:20.128567 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-db-create-z8nrd" Nov 21 14:30:20 crc kubenswrapper[4897]: I1121 14:30:20.174009 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eb1688c2-d3d5-4935-b08b-32ef2ce7f6e1-operator-scripts\") pod \"mysqld-exporter-openstack-db-create-z8nrd\" (UID: \"eb1688c2-d3d5-4935-b08b-32ef2ce7f6e1\") " pod="openstack/mysqld-exporter-openstack-db-create-z8nrd" Nov 21 14:30:20 crc kubenswrapper[4897]: I1121 14:30:20.174103 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vn2gx\" (UniqueName: \"kubernetes.io/projected/eb1688c2-d3d5-4935-b08b-32ef2ce7f6e1-kube-api-access-vn2gx\") pod \"mysqld-exporter-openstack-db-create-z8nrd\" (UID: \"eb1688c2-d3d5-4935-b08b-32ef2ce7f6e1\") " pod="openstack/mysqld-exporter-openstack-db-create-z8nrd" Nov 21 14:30:20 crc kubenswrapper[4897]: I1121 14:30:20.251212 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f58d5cd65-2pc2p"] Nov 21 14:30:20 crc kubenswrapper[4897]: I1121 14:30:20.257751 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f58d5cd65-2pc2p"] Nov 21 14:30:20 crc kubenswrapper[4897]: I1121 14:30:20.276206 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eb1688c2-d3d5-4935-b08b-32ef2ce7f6e1-operator-scripts\") pod \"mysqld-exporter-openstack-db-create-z8nrd\" (UID: \"eb1688c2-d3d5-4935-b08b-32ef2ce7f6e1\") " pod="openstack/mysqld-exporter-openstack-db-create-z8nrd" Nov 21 14:30:20 crc kubenswrapper[4897]: I1121 14:30:20.276296 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vn2gx\" (UniqueName: \"kubernetes.io/projected/eb1688c2-d3d5-4935-b08b-32ef2ce7f6e1-kube-api-access-vn2gx\") pod \"mysqld-exporter-openstack-db-create-z8nrd\" (UID: \"eb1688c2-d3d5-4935-b08b-32ef2ce7f6e1\") " pod="openstack/mysqld-exporter-openstack-db-create-z8nrd" Nov 21 14:30:20 crc kubenswrapper[4897]: I1121 14:30:20.276897 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eb1688c2-d3d5-4935-b08b-32ef2ce7f6e1-operator-scripts\") pod \"mysqld-exporter-openstack-db-create-z8nrd\" (UID: \"eb1688c2-d3d5-4935-b08b-32ef2ce7f6e1\") " pod="openstack/mysqld-exporter-openstack-db-create-z8nrd" Nov 21 14:30:20 crc kubenswrapper[4897]: I1121 14:30:20.302462 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vn2gx\" (UniqueName: \"kubernetes.io/projected/eb1688c2-d3d5-4935-b08b-32ef2ce7f6e1-kube-api-access-vn2gx\") pod \"mysqld-exporter-openstack-db-create-z8nrd\" (UID: \"eb1688c2-d3d5-4935-b08b-32ef2ce7f6e1\") " pod="openstack/mysqld-exporter-openstack-db-create-z8nrd" Nov 21 14:30:20 crc kubenswrapper[4897]: I1121 14:30:20.317880 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mysqld-exporter-95d1-account-create-7vtq6"] Nov 21 14:30:20 crc kubenswrapper[4897]: I1121 14:30:20.319571 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-95d1-account-create-7vtq6" Nov 21 14:30:20 crc kubenswrapper[4897]: I1121 14:30:20.321431 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"mysqld-exporter-openstack-db-secret" Nov 21 14:30:20 crc kubenswrapper[4897]: I1121 14:30:20.325777 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-95d1-account-create-7vtq6"] Nov 21 14:30:20 crc kubenswrapper[4897]: I1121 14:30:20.377814 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5d1468bb-2545-4906-a10c-c95ba9ef8693-operator-scripts\") pod \"mysqld-exporter-95d1-account-create-7vtq6\" (UID: \"5d1468bb-2545-4906-a10c-c95ba9ef8693\") " pod="openstack/mysqld-exporter-95d1-account-create-7vtq6" Nov 21 14:30:20 crc kubenswrapper[4897]: I1121 14:30:20.378092 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-824zh\" (UniqueName: \"kubernetes.io/projected/5d1468bb-2545-4906-a10c-c95ba9ef8693-kube-api-access-824zh\") pod \"mysqld-exporter-95d1-account-create-7vtq6\" (UID: \"5d1468bb-2545-4906-a10c-c95ba9ef8693\") " pod="openstack/mysqld-exporter-95d1-account-create-7vtq6" Nov 21 14:30:20 crc kubenswrapper[4897]: I1121 14:30:20.479686 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-824zh\" (UniqueName: \"kubernetes.io/projected/5d1468bb-2545-4906-a10c-c95ba9ef8693-kube-api-access-824zh\") pod \"mysqld-exporter-95d1-account-create-7vtq6\" (UID: \"5d1468bb-2545-4906-a10c-c95ba9ef8693\") " pod="openstack/mysqld-exporter-95d1-account-create-7vtq6" Nov 21 14:30:20 crc kubenswrapper[4897]: I1121 14:30:20.479768 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5d1468bb-2545-4906-a10c-c95ba9ef8693-operator-scripts\") pod \"mysqld-exporter-95d1-account-create-7vtq6\" (UID: \"5d1468bb-2545-4906-a10c-c95ba9ef8693\") " pod="openstack/mysqld-exporter-95d1-account-create-7vtq6" Nov 21 14:30:20 crc kubenswrapper[4897]: I1121 14:30:20.480376 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5d1468bb-2545-4906-a10c-c95ba9ef8693-operator-scripts\") pod \"mysqld-exporter-95d1-account-create-7vtq6\" (UID: \"5d1468bb-2545-4906-a10c-c95ba9ef8693\") " pod="openstack/mysqld-exporter-95d1-account-create-7vtq6" Nov 21 14:30:20 crc kubenswrapper[4897]: I1121 14:30:20.497963 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-824zh\" (UniqueName: \"kubernetes.io/projected/5d1468bb-2545-4906-a10c-c95ba9ef8693-kube-api-access-824zh\") pod \"mysqld-exporter-95d1-account-create-7vtq6\" (UID: \"5d1468bb-2545-4906-a10c-c95ba9ef8693\") " pod="openstack/mysqld-exporter-95d1-account-create-7vtq6" Nov 21 14:30:20 crc kubenswrapper[4897]: I1121 14:30:20.546936 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-db-create-z8nrd" Nov 21 14:30:20 crc kubenswrapper[4897]: I1121 14:30:20.660635 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-95d1-account-create-7vtq6" Nov 21 14:30:20 crc kubenswrapper[4897]: I1121 14:30:20.931494 4897 generic.go:334] "Generic (PLEG): container finished" podID="7501a167-0655-4ad4-bb04-c597a81359ba" containerID="1675902aaf6ae6196f179d75e509846f5abd6d1664aa148b6578ceea932c4544" exitCode=0 Nov 21 14:30:20 crc kubenswrapper[4897]: I1121 14:30:20.931608 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-rkm52" event={"ID":"7501a167-0655-4ad4-bb04-c597a81359ba","Type":"ContainerDied","Data":"1675902aaf6ae6196f179d75e509846f5abd6d1664aa148b6578ceea932c4544"} Nov 21 14:30:20 crc kubenswrapper[4897]: I1121 14:30:20.937009 4897 generic.go:334] "Generic (PLEG): container finished" podID="34a64a5d-5b9d-4aac-be3e-96edc721fa3b" containerID="860f4995b77cd310712c0190183323bddec94b2b3f7a6bf682efd550850a9716" exitCode=0 Nov 21 14:30:20 crc kubenswrapper[4897]: I1121 14:30:20.937116 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-b2f5-account-create-69ls6" event={"ID":"34a64a5d-5b9d-4aac-be3e-96edc721fa3b","Type":"ContainerDied","Data":"860f4995b77cd310712c0190183323bddec94b2b3f7a6bf682efd550850a9716"} Nov 21 14:30:20 crc kubenswrapper[4897]: I1121 14:30:20.939273 4897 generic.go:334] "Generic (PLEG): container finished" podID="994c36f9-336e-49dc-b45d-0eb4bfa0b686" containerID="9522822e4d3514856f080fe94bb45b833c81336e3fd4d133f5c86ec83ec51d64" exitCode=0 Nov 21 14:30:20 crc kubenswrapper[4897]: I1121 14:30:20.939343 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-376f-account-create-q6xsf" event={"ID":"994c36f9-336e-49dc-b45d-0eb4bfa0b686","Type":"ContainerDied","Data":"9522822e4d3514856f080fe94bb45b833c81336e3fd4d133f5c86ec83ec51d64"} Nov 21 14:30:20 crc kubenswrapper[4897]: I1121 14:30:20.941141 4897 generic.go:334] "Generic (PLEG): container finished" podID="d3441939-1c13-4b55-a129-a39f4bed36cd" containerID="f92ea24bf1f109608a14fbb251bce2cef26cc268df4c3daa8daeb16e4d2f599d" exitCode=0 Nov 21 14:30:20 crc kubenswrapper[4897]: I1121 14:30:20.941244 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-sbnf2" event={"ID":"d3441939-1c13-4b55-a129-a39f4bed36cd","Type":"ContainerDied","Data":"f92ea24bf1f109608a14fbb251bce2cef26cc268df4c3daa8daeb16e4d2f599d"} Nov 21 14:30:20 crc kubenswrapper[4897]: I1121 14:30:20.943300 4897 generic.go:334] "Generic (PLEG): container finished" podID="980c0de3-d435-439f-8477-0dc68fbaf7f0" containerID="0a09a0b4d2bf00b873dd3a26366c3cb500c70dfa74206e2efd43e17e17491e68" exitCode=0 Nov 21 14:30:20 crc kubenswrapper[4897]: I1121 14:30:20.943379 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-b8bf-account-create-x2vdt" event={"ID":"980c0de3-d435-439f-8477-0dc68fbaf7f0","Type":"ContainerDied","Data":"0a09a0b4d2bf00b873dd3a26366c3cb500c70dfa74206e2efd43e17e17491e68"} Nov 21 14:30:21 crc kubenswrapper[4897]: I1121 14:30:21.037212 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-openstack-db-create-z8nrd"] Nov 21 14:30:21 crc kubenswrapper[4897]: I1121 14:30:21.142585 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-95d1-account-create-7vtq6"] Nov 21 14:30:21 crc kubenswrapper[4897]: I1121 14:30:21.794434 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-2snxf" Nov 21 14:30:21 crc kubenswrapper[4897]: I1121 14:30:21.926337 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d7a5e302-fb61-41b2-a7d2-ff02160c6679-operator-scripts\") pod \"d7a5e302-fb61-41b2-a7d2-ff02160c6679\" (UID: \"d7a5e302-fb61-41b2-a7d2-ff02160c6679\") " Nov 21 14:30:21 crc kubenswrapper[4897]: I1121 14:30:21.926566 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gwdzt\" (UniqueName: \"kubernetes.io/projected/d7a5e302-fb61-41b2-a7d2-ff02160c6679-kube-api-access-gwdzt\") pod \"d7a5e302-fb61-41b2-a7d2-ff02160c6679\" (UID: \"d7a5e302-fb61-41b2-a7d2-ff02160c6679\") " Nov 21 14:30:21 crc kubenswrapper[4897]: I1121 14:30:21.928704 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7a5e302-fb61-41b2-a7d2-ff02160c6679-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d7a5e302-fb61-41b2-a7d2-ff02160c6679" (UID: "d7a5e302-fb61-41b2-a7d2-ff02160c6679"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:30:21 crc kubenswrapper[4897]: I1121 14:30:21.957881 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-b8bf-account-create-x2vdt" event={"ID":"980c0de3-d435-439f-8477-0dc68fbaf7f0","Type":"ContainerDied","Data":"b1aaf659a72e0f1f9547ceac86f946bc9644b1473e0559ddc3e90dfa18948ceb"} Nov 21 14:30:21 crc kubenswrapper[4897]: I1121 14:30:21.957927 4897 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b1aaf659a72e0f1f9547ceac86f946bc9644b1473e0559ddc3e90dfa18948ceb" Nov 21 14:30:21 crc kubenswrapper[4897]: I1121 14:30:21.959290 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-95d1-account-create-7vtq6" event={"ID":"5d1468bb-2545-4906-a10c-c95ba9ef8693","Type":"ContainerStarted","Data":"59e1f2a197b68bfefe8b2456c8c419d86f697a9a44431a02e3e85fd09bbfe751"} Nov 21 14:30:21 crc kubenswrapper[4897]: I1121 14:30:21.961942 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-db-create-z8nrd" event={"ID":"eb1688c2-d3d5-4935-b08b-32ef2ce7f6e1","Type":"ContainerStarted","Data":"58c7b6743fa1ab72410a2524ce920b80013ab6e28a954772e19585e08b8cfc0c"} Nov 21 14:30:21 crc kubenswrapper[4897]: I1121 14:30:21.963898 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-2snxf" event={"ID":"d7a5e302-fb61-41b2-a7d2-ff02160c6679","Type":"ContainerDied","Data":"da85ac06dc94cbce05fef034af847bddbae25cc65707caf1f487f6175d4ecae4"} Nov 21 14:30:21 crc kubenswrapper[4897]: I1121 14:30:21.964147 4897 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="da85ac06dc94cbce05fef034af847bddbae25cc65707caf1f487f6175d4ecae4" Nov 21 14:30:21 crc kubenswrapper[4897]: I1121 14:30:21.964166 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-2snxf" Nov 21 14:30:21 crc kubenswrapper[4897]: I1121 14:30:21.969865 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-b2f5-account-create-69ls6" event={"ID":"34a64a5d-5b9d-4aac-be3e-96edc721fa3b","Type":"ContainerDied","Data":"b894e1ce596ce8e4b5c070cebe296697fb3efa4e874a9b64ed45a7dc5cd4f648"} Nov 21 14:30:21 crc kubenswrapper[4897]: I1121 14:30:21.969899 4897 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b894e1ce596ce8e4b5c070cebe296697fb3efa4e874a9b64ed45a7dc5cd4f648" Nov 21 14:30:22 crc kubenswrapper[4897]: I1121 14:30:22.028785 4897 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d7a5e302-fb61-41b2-a7d2-ff02160c6679-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 21 14:30:22 crc kubenswrapper[4897]: I1121 14:30:22.034543 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7a5e302-fb61-41b2-a7d2-ff02160c6679-kube-api-access-gwdzt" (OuterVolumeSpecName: "kube-api-access-gwdzt") pod "d7a5e302-fb61-41b2-a7d2-ff02160c6679" (UID: "d7a5e302-fb61-41b2-a7d2-ff02160c6679"). InnerVolumeSpecName "kube-api-access-gwdzt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:30:22 crc kubenswrapper[4897]: I1121 14:30:22.113191 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-b8bf-account-create-x2vdt" Nov 21 14:30:22 crc kubenswrapper[4897]: I1121 14:30:22.116662 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ab48e37a-819f-4f63-8e67-9a2cbeaaeb82" path="/var/lib/kubelet/pods/ab48e37a-819f-4f63-8e67-9a2cbeaaeb82/volumes" Nov 21 14:30:22 crc kubenswrapper[4897]: I1121 14:30:22.142755 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gwdzt\" (UniqueName: \"kubernetes.io/projected/d7a5e302-fb61-41b2-a7d2-ff02160c6679-kube-api-access-gwdzt\") on node \"crc\" DevicePath \"\"" Nov 21 14:30:22 crc kubenswrapper[4897]: I1121 14:30:22.146062 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-b2f5-account-create-69ls6" Nov 21 14:30:22 crc kubenswrapper[4897]: I1121 14:30:22.244277 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pdplw\" (UniqueName: \"kubernetes.io/projected/980c0de3-d435-439f-8477-0dc68fbaf7f0-kube-api-access-pdplw\") pod \"980c0de3-d435-439f-8477-0dc68fbaf7f0\" (UID: \"980c0de3-d435-439f-8477-0dc68fbaf7f0\") " Nov 21 14:30:22 crc kubenswrapper[4897]: I1121 14:30:22.244360 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/34a64a5d-5b9d-4aac-be3e-96edc721fa3b-operator-scripts\") pod \"34a64a5d-5b9d-4aac-be3e-96edc721fa3b\" (UID: \"34a64a5d-5b9d-4aac-be3e-96edc721fa3b\") " Nov 21 14:30:22 crc kubenswrapper[4897]: I1121 14:30:22.244415 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k9dr2\" (UniqueName: \"kubernetes.io/projected/34a64a5d-5b9d-4aac-be3e-96edc721fa3b-kube-api-access-k9dr2\") pod \"34a64a5d-5b9d-4aac-be3e-96edc721fa3b\" (UID: \"34a64a5d-5b9d-4aac-be3e-96edc721fa3b\") " Nov 21 14:30:22 crc kubenswrapper[4897]: I1121 14:30:22.244444 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/980c0de3-d435-439f-8477-0dc68fbaf7f0-operator-scripts\") pod \"980c0de3-d435-439f-8477-0dc68fbaf7f0\" (UID: \"980c0de3-d435-439f-8477-0dc68fbaf7f0\") " Nov 21 14:30:22 crc kubenswrapper[4897]: I1121 14:30:22.245081 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/34a64a5d-5b9d-4aac-be3e-96edc721fa3b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "34a64a5d-5b9d-4aac-be3e-96edc721fa3b" (UID: "34a64a5d-5b9d-4aac-be3e-96edc721fa3b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:30:22 crc kubenswrapper[4897]: I1121 14:30:22.245225 4897 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/34a64a5d-5b9d-4aac-be3e-96edc721fa3b-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 21 14:30:22 crc kubenswrapper[4897]: I1121 14:30:22.245827 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/980c0de3-d435-439f-8477-0dc68fbaf7f0-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "980c0de3-d435-439f-8477-0dc68fbaf7f0" (UID: "980c0de3-d435-439f-8477-0dc68fbaf7f0"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:30:22 crc kubenswrapper[4897]: I1121 14:30:22.256359 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34a64a5d-5b9d-4aac-be3e-96edc721fa3b-kube-api-access-k9dr2" (OuterVolumeSpecName: "kube-api-access-k9dr2") pod "34a64a5d-5b9d-4aac-be3e-96edc721fa3b" (UID: "34a64a5d-5b9d-4aac-be3e-96edc721fa3b"). InnerVolumeSpecName "kube-api-access-k9dr2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:30:22 crc kubenswrapper[4897]: I1121 14:30:22.256447 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/980c0de3-d435-439f-8477-0dc68fbaf7f0-kube-api-access-pdplw" (OuterVolumeSpecName: "kube-api-access-pdplw") pod "980c0de3-d435-439f-8477-0dc68fbaf7f0" (UID: "980c0de3-d435-439f-8477-0dc68fbaf7f0"). InnerVolumeSpecName "kube-api-access-pdplw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:30:22 crc kubenswrapper[4897]: I1121 14:30:22.346636 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-sbnf2" Nov 21 14:30:22 crc kubenswrapper[4897]: I1121 14:30:22.353634 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pdplw\" (UniqueName: \"kubernetes.io/projected/980c0de3-d435-439f-8477-0dc68fbaf7f0-kube-api-access-pdplw\") on node \"crc\" DevicePath \"\"" Nov 21 14:30:22 crc kubenswrapper[4897]: I1121 14:30:22.353713 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k9dr2\" (UniqueName: \"kubernetes.io/projected/34a64a5d-5b9d-4aac-be3e-96edc721fa3b-kube-api-access-k9dr2\") on node \"crc\" DevicePath \"\"" Nov 21 14:30:22 crc kubenswrapper[4897]: I1121 14:30:22.353733 4897 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/980c0de3-d435-439f-8477-0dc68fbaf7f0-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 21 14:30:22 crc kubenswrapper[4897]: I1121 14:30:22.454905 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-994hn\" (UniqueName: \"kubernetes.io/projected/d3441939-1c13-4b55-a129-a39f4bed36cd-kube-api-access-994hn\") pod \"d3441939-1c13-4b55-a129-a39f4bed36cd\" (UID: \"d3441939-1c13-4b55-a129-a39f4bed36cd\") " Nov 21 14:30:22 crc kubenswrapper[4897]: I1121 14:30:22.455080 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d3441939-1c13-4b55-a129-a39f4bed36cd-operator-scripts\") pod \"d3441939-1c13-4b55-a129-a39f4bed36cd\" (UID: \"d3441939-1c13-4b55-a129-a39f4bed36cd\") " Nov 21 14:30:22 crc kubenswrapper[4897]: I1121 14:30:22.455662 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d3441939-1c13-4b55-a129-a39f4bed36cd-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d3441939-1c13-4b55-a129-a39f4bed36cd" (UID: "d3441939-1c13-4b55-a129-a39f4bed36cd"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:30:22 crc kubenswrapper[4897]: I1121 14:30:22.460852 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3441939-1c13-4b55-a129-a39f4bed36cd-kube-api-access-994hn" (OuterVolumeSpecName: "kube-api-access-994hn") pod "d3441939-1c13-4b55-a129-a39f4bed36cd" (UID: "d3441939-1c13-4b55-a129-a39f4bed36cd"). InnerVolumeSpecName "kube-api-access-994hn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:30:22 crc kubenswrapper[4897]: I1121 14:30:22.557439 4897 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d3441939-1c13-4b55-a129-a39f4bed36cd-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 21 14:30:22 crc kubenswrapper[4897]: I1121 14:30:22.557470 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-994hn\" (UniqueName: \"kubernetes.io/projected/d3441939-1c13-4b55-a129-a39f4bed36cd-kube-api-access-994hn\") on node \"crc\" DevicePath \"\"" Nov 21 14:30:22 crc kubenswrapper[4897]: I1121 14:30:22.756524 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-376f-account-create-q6xsf" Nov 21 14:30:22 crc kubenswrapper[4897]: I1121 14:30:22.763165 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-rkm52" Nov 21 14:30:22 crc kubenswrapper[4897]: I1121 14:30:22.863304 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xckvk\" (UniqueName: \"kubernetes.io/projected/7501a167-0655-4ad4-bb04-c597a81359ba-kube-api-access-xckvk\") pod \"7501a167-0655-4ad4-bb04-c597a81359ba\" (UID: \"7501a167-0655-4ad4-bb04-c597a81359ba\") " Nov 21 14:30:22 crc kubenswrapper[4897]: I1121 14:30:22.863669 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/994c36f9-336e-49dc-b45d-0eb4bfa0b686-operator-scripts\") pod \"994c36f9-336e-49dc-b45d-0eb4bfa0b686\" (UID: \"994c36f9-336e-49dc-b45d-0eb4bfa0b686\") " Nov 21 14:30:22 crc kubenswrapper[4897]: I1121 14:30:22.863752 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7501a167-0655-4ad4-bb04-c597a81359ba-operator-scripts\") pod \"7501a167-0655-4ad4-bb04-c597a81359ba\" (UID: \"7501a167-0655-4ad4-bb04-c597a81359ba\") " Nov 21 14:30:22 crc kubenswrapper[4897]: I1121 14:30:22.863807 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b6lg4\" (UniqueName: \"kubernetes.io/projected/994c36f9-336e-49dc-b45d-0eb4bfa0b686-kube-api-access-b6lg4\") pod \"994c36f9-336e-49dc-b45d-0eb4bfa0b686\" (UID: \"994c36f9-336e-49dc-b45d-0eb4bfa0b686\") " Nov 21 14:30:22 crc kubenswrapper[4897]: I1121 14:30:22.864345 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/994c36f9-336e-49dc-b45d-0eb4bfa0b686-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "994c36f9-336e-49dc-b45d-0eb4bfa0b686" (UID: "994c36f9-336e-49dc-b45d-0eb4bfa0b686"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:30:22 crc kubenswrapper[4897]: I1121 14:30:22.864627 4897 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/994c36f9-336e-49dc-b45d-0eb4bfa0b686-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 21 14:30:22 crc kubenswrapper[4897]: I1121 14:30:22.864617 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7501a167-0655-4ad4-bb04-c597a81359ba-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7501a167-0655-4ad4-bb04-c597a81359ba" (UID: "7501a167-0655-4ad4-bb04-c597a81359ba"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:30:22 crc kubenswrapper[4897]: I1121 14:30:22.868474 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/994c36f9-336e-49dc-b45d-0eb4bfa0b686-kube-api-access-b6lg4" (OuterVolumeSpecName: "kube-api-access-b6lg4") pod "994c36f9-336e-49dc-b45d-0eb4bfa0b686" (UID: "994c36f9-336e-49dc-b45d-0eb4bfa0b686"). InnerVolumeSpecName "kube-api-access-b6lg4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:30:22 crc kubenswrapper[4897]: I1121 14:30:22.870126 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7501a167-0655-4ad4-bb04-c597a81359ba-kube-api-access-xckvk" (OuterVolumeSpecName: "kube-api-access-xckvk") pod "7501a167-0655-4ad4-bb04-c597a81359ba" (UID: "7501a167-0655-4ad4-bb04-c597a81359ba"). InnerVolumeSpecName "kube-api-access-xckvk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:30:22 crc kubenswrapper[4897]: I1121 14:30:22.966262 4897 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7501a167-0655-4ad4-bb04-c597a81359ba-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 21 14:30:22 crc kubenswrapper[4897]: I1121 14:30:22.966307 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b6lg4\" (UniqueName: \"kubernetes.io/projected/994c36f9-336e-49dc-b45d-0eb4bfa0b686-kube-api-access-b6lg4\") on node \"crc\" DevicePath \"\"" Nov 21 14:30:22 crc kubenswrapper[4897]: I1121 14:30:22.966323 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xckvk\" (UniqueName: \"kubernetes.io/projected/7501a167-0655-4ad4-bb04-c597a81359ba-kube-api-access-xckvk\") on node \"crc\" DevicePath \"\"" Nov 21 14:30:22 crc kubenswrapper[4897]: I1121 14:30:22.983912 4897 generic.go:334] "Generic (PLEG): container finished" podID="6d1f75d9-2eaf-44c1-b2d0-dae3a8547433" containerID="470e1f993391940c03765e4498a36ab375ab427bb7aaf8afafebcb169881092c" exitCode=0 Nov 21 14:30:22 crc kubenswrapper[4897]: I1121 14:30:22.984005 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-t9nzd" event={"ID":"6d1f75d9-2eaf-44c1-b2d0-dae3a8547433","Type":"ContainerDied","Data":"470e1f993391940c03765e4498a36ab375ab427bb7aaf8afafebcb169881092c"} Nov 21 14:30:22 crc kubenswrapper[4897]: I1121 14:30:22.988488 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"91aa2270-92a7-4af4-9a92-141ea2b24ee1","Type":"ContainerStarted","Data":"1ae73b0cd6fdca18483acc4c5180aa46a081811e918b88fc1fb45977279fc1f2"} Nov 21 14:30:22 crc kubenswrapper[4897]: I1121 14:30:22.991347 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-376f-account-create-q6xsf" event={"ID":"994c36f9-336e-49dc-b45d-0eb4bfa0b686","Type":"ContainerDied","Data":"988d66d02d23611bbf0f2e52b8d01eb5dc8312e548ca15e48ee3bd7cc70c4977"} Nov 21 14:30:22 crc kubenswrapper[4897]: I1121 14:30:22.991416 4897 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="988d66d02d23611bbf0f2e52b8d01eb5dc8312e548ca15e48ee3bd7cc70c4977" Nov 21 14:30:22 crc kubenswrapper[4897]: I1121 14:30:22.991545 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-376f-account-create-q6xsf" Nov 21 14:30:22 crc kubenswrapper[4897]: I1121 14:30:22.993604 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-sbnf2" event={"ID":"d3441939-1c13-4b55-a129-a39f4bed36cd","Type":"ContainerDied","Data":"ec0ce44abdd2ca8c499480896781a88461939aa9a7047d23589a8ddc6e3bb626"} Nov 21 14:30:22 crc kubenswrapper[4897]: I1121 14:30:22.993647 4897 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ec0ce44abdd2ca8c499480896781a88461939aa9a7047d23589a8ddc6e3bb626" Nov 21 14:30:22 crc kubenswrapper[4897]: I1121 14:30:22.993614 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-sbnf2" Nov 21 14:30:23 crc kubenswrapper[4897]: I1121 14:30:22.999959 4897 generic.go:334] "Generic (PLEG): container finished" podID="5d1468bb-2545-4906-a10c-c95ba9ef8693" containerID="a593f235df64892b4938a5931d426f32f36f7bee61dd522b3168c4919dce75a6" exitCode=0 Nov 21 14:30:23 crc kubenswrapper[4897]: I1121 14:30:23.000075 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-95d1-account-create-7vtq6" event={"ID":"5d1468bb-2545-4906-a10c-c95ba9ef8693","Type":"ContainerDied","Data":"a593f235df64892b4938a5931d426f32f36f7bee61dd522b3168c4919dce75a6"} Nov 21 14:30:23 crc kubenswrapper[4897]: I1121 14:30:23.014074 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-rkm52" Nov 21 14:30:23 crc kubenswrapper[4897]: I1121 14:30:23.014078 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-rkm52" event={"ID":"7501a167-0655-4ad4-bb04-c597a81359ba","Type":"ContainerDied","Data":"d10acce71eb6e650686d1b6fd92cb5b47a393460cb71d96141603be290d78802"} Nov 21 14:30:23 crc kubenswrapper[4897]: I1121 14:30:23.014146 4897 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d10acce71eb6e650686d1b6fd92cb5b47a393460cb71d96141603be290d78802" Nov 21 14:30:23 crc kubenswrapper[4897]: I1121 14:30:23.017992 4897 generic.go:334] "Generic (PLEG): container finished" podID="eb1688c2-d3d5-4935-b08b-32ef2ce7f6e1" containerID="587c7c12071ac71050fe033554b6b7a03bb00d1e6b792ed3be8c7fe6049d2e25" exitCode=0 Nov 21 14:30:23 crc kubenswrapper[4897]: I1121 14:30:23.018075 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-b8bf-account-create-x2vdt" Nov 21 14:30:23 crc kubenswrapper[4897]: I1121 14:30:23.018723 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-db-create-z8nrd" event={"ID":"eb1688c2-d3d5-4935-b08b-32ef2ce7f6e1","Type":"ContainerDied","Data":"587c7c12071ac71050fe033554b6b7a03bb00d1e6b792ed3be8c7fe6049d2e25"} Nov 21 14:30:23 crc kubenswrapper[4897]: I1121 14:30:23.018861 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-b2f5-account-create-69ls6" Nov 21 14:30:23 crc kubenswrapper[4897]: I1121 14:30:23.170331 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/5f9bf5d1-dcbd-4d19-b46d-22810ebf17a2-etc-swift\") pod \"swift-storage-0\" (UID: \"5f9bf5d1-dcbd-4d19-b46d-22810ebf17a2\") " pod="openstack/swift-storage-0" Nov 21 14:30:23 crc kubenswrapper[4897]: I1121 14:30:23.181748 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/5f9bf5d1-dcbd-4d19-b46d-22810ebf17a2-etc-swift\") pod \"swift-storage-0\" (UID: \"5f9bf5d1-dcbd-4d19-b46d-22810ebf17a2\") " pod="openstack/swift-storage-0" Nov 21 14:30:23 crc kubenswrapper[4897]: I1121 14:30:23.324805 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Nov 21 14:30:23 crc kubenswrapper[4897]: I1121 14:30:23.488165 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-mm6tz" podUID="e0c27889-eea7-4a06-9195-3401929cf6b2" containerName="ovn-controller" probeResult="failure" output=< Nov 21 14:30:23 crc kubenswrapper[4897]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Nov 21 14:30:23 crc kubenswrapper[4897]: > Nov 21 14:30:23 crc kubenswrapper[4897]: I1121 14:30:23.495843 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-k28mx" Nov 21 14:30:23 crc kubenswrapper[4897]: I1121 14:30:23.938288 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Nov 21 14:30:23 crc kubenswrapper[4897]: W1121 14:30:23.949596 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5f9bf5d1_dcbd_4d19_b46d_22810ebf17a2.slice/crio-d9a634fbbcfc47c79b1937c635688c665c239cdfe61d56fd03d3cae0c7da10e3 WatchSource:0}: Error finding container d9a634fbbcfc47c79b1937c635688c665c239cdfe61d56fd03d3cae0c7da10e3: Status 404 returned error can't find the container with id d9a634fbbcfc47c79b1937c635688c665c239cdfe61d56fd03d3cae0c7da10e3 Nov 21 14:30:24 crc kubenswrapper[4897]: I1121 14:30:24.029483 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"5f9bf5d1-dcbd-4d19-b46d-22810ebf17a2","Type":"ContainerStarted","Data":"d9a634fbbcfc47c79b1937c635688c665c239cdfe61d56fd03d3cae0c7da10e3"} Nov 21 14:30:24 crc kubenswrapper[4897]: I1121 14:30:24.277841 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Nov 21 14:30:24 crc kubenswrapper[4897]: I1121 14:30:24.498257 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-t9nzd" Nov 21 14:30:24 crc kubenswrapper[4897]: I1121 14:30:24.610450 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6d1f75d9-2eaf-44c1-b2d0-dae3a8547433-scripts\") pod \"6d1f75d9-2eaf-44c1-b2d0-dae3a8547433\" (UID: \"6d1f75d9-2eaf-44c1-b2d0-dae3a8547433\") " Nov 21 14:30:24 crc kubenswrapper[4897]: I1121 14:30:24.610604 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/6d1f75d9-2eaf-44c1-b2d0-dae3a8547433-dispersionconf\") pod \"6d1f75d9-2eaf-44c1-b2d0-dae3a8547433\" (UID: \"6d1f75d9-2eaf-44c1-b2d0-dae3a8547433\") " Nov 21 14:30:24 crc kubenswrapper[4897]: I1121 14:30:24.610689 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/6d1f75d9-2eaf-44c1-b2d0-dae3a8547433-swiftconf\") pod \"6d1f75d9-2eaf-44c1-b2d0-dae3a8547433\" (UID: \"6d1f75d9-2eaf-44c1-b2d0-dae3a8547433\") " Nov 21 14:30:24 crc kubenswrapper[4897]: I1121 14:30:24.610755 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/6d1f75d9-2eaf-44c1-b2d0-dae3a8547433-etc-swift\") pod \"6d1f75d9-2eaf-44c1-b2d0-dae3a8547433\" (UID: \"6d1f75d9-2eaf-44c1-b2d0-dae3a8547433\") " Nov 21 14:30:24 crc kubenswrapper[4897]: I1121 14:30:24.610780 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z9584\" (UniqueName: \"kubernetes.io/projected/6d1f75d9-2eaf-44c1-b2d0-dae3a8547433-kube-api-access-z9584\") pod \"6d1f75d9-2eaf-44c1-b2d0-dae3a8547433\" (UID: \"6d1f75d9-2eaf-44c1-b2d0-dae3a8547433\") " Nov 21 14:30:24 crc kubenswrapper[4897]: I1121 14:30:24.610871 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/6d1f75d9-2eaf-44c1-b2d0-dae3a8547433-ring-data-devices\") pod \"6d1f75d9-2eaf-44c1-b2d0-dae3a8547433\" (UID: \"6d1f75d9-2eaf-44c1-b2d0-dae3a8547433\") " Nov 21 14:30:24 crc kubenswrapper[4897]: I1121 14:30:24.610894 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d1f75d9-2eaf-44c1-b2d0-dae3a8547433-combined-ca-bundle\") pod \"6d1f75d9-2eaf-44c1-b2d0-dae3a8547433\" (UID: \"6d1f75d9-2eaf-44c1-b2d0-dae3a8547433\") " Nov 21 14:30:24 crc kubenswrapper[4897]: I1121 14:30:24.612438 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6d1f75d9-2eaf-44c1-b2d0-dae3a8547433-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "6d1f75d9-2eaf-44c1-b2d0-dae3a8547433" (UID: "6d1f75d9-2eaf-44c1-b2d0-dae3a8547433"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:30:24 crc kubenswrapper[4897]: I1121 14:30:24.612886 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6d1f75d9-2eaf-44c1-b2d0-dae3a8547433-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "6d1f75d9-2eaf-44c1-b2d0-dae3a8547433" (UID: "6d1f75d9-2eaf-44c1-b2d0-dae3a8547433"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:30:24 crc kubenswrapper[4897]: I1121 14:30:24.616519 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d1f75d9-2eaf-44c1-b2d0-dae3a8547433-kube-api-access-z9584" (OuterVolumeSpecName: "kube-api-access-z9584") pod "6d1f75d9-2eaf-44c1-b2d0-dae3a8547433" (UID: "6d1f75d9-2eaf-44c1-b2d0-dae3a8547433"). InnerVolumeSpecName "kube-api-access-z9584". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:30:24 crc kubenswrapper[4897]: I1121 14:30:24.621552 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d1f75d9-2eaf-44c1-b2d0-dae3a8547433-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "6d1f75d9-2eaf-44c1-b2d0-dae3a8547433" (UID: "6d1f75d9-2eaf-44c1-b2d0-dae3a8547433"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:30:24 crc kubenswrapper[4897]: I1121 14:30:24.641865 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6d1f75d9-2eaf-44c1-b2d0-dae3a8547433-scripts" (OuterVolumeSpecName: "scripts") pod "6d1f75d9-2eaf-44c1-b2d0-dae3a8547433" (UID: "6d1f75d9-2eaf-44c1-b2d0-dae3a8547433"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:30:24 crc kubenswrapper[4897]: I1121 14:30:24.649754 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d1f75d9-2eaf-44c1-b2d0-dae3a8547433-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6d1f75d9-2eaf-44c1-b2d0-dae3a8547433" (UID: "6d1f75d9-2eaf-44c1-b2d0-dae3a8547433"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:30:24 crc kubenswrapper[4897]: I1121 14:30:24.652413 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d1f75d9-2eaf-44c1-b2d0-dae3a8547433-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "6d1f75d9-2eaf-44c1-b2d0-dae3a8547433" (UID: "6d1f75d9-2eaf-44c1-b2d0-dae3a8547433"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:30:24 crc kubenswrapper[4897]: I1121 14:30:24.695306 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-db-create-z8nrd" Nov 21 14:30:24 crc kubenswrapper[4897]: I1121 14:30:24.714368 4897 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/6d1f75d9-2eaf-44c1-b2d0-dae3a8547433-etc-swift\") on node \"crc\" DevicePath \"\"" Nov 21 14:30:24 crc kubenswrapper[4897]: I1121 14:30:24.714403 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z9584\" (UniqueName: \"kubernetes.io/projected/6d1f75d9-2eaf-44c1-b2d0-dae3a8547433-kube-api-access-z9584\") on node \"crc\" DevicePath \"\"" Nov 21 14:30:24 crc kubenswrapper[4897]: I1121 14:30:24.716028 4897 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/6d1f75d9-2eaf-44c1-b2d0-dae3a8547433-ring-data-devices\") on node \"crc\" DevicePath \"\"" Nov 21 14:30:24 crc kubenswrapper[4897]: I1121 14:30:24.716047 4897 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d1f75d9-2eaf-44c1-b2d0-dae3a8547433-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 21 14:30:24 crc kubenswrapper[4897]: I1121 14:30:24.716084 4897 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6d1f75d9-2eaf-44c1-b2d0-dae3a8547433-scripts\") on node \"crc\" DevicePath \"\"" Nov 21 14:30:24 crc kubenswrapper[4897]: I1121 14:30:24.716095 4897 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/6d1f75d9-2eaf-44c1-b2d0-dae3a8547433-dispersionconf\") on node \"crc\" DevicePath \"\"" Nov 21 14:30:24 crc kubenswrapper[4897]: I1121 14:30:24.716112 4897 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/6d1f75d9-2eaf-44c1-b2d0-dae3a8547433-swiftconf\") on node \"crc\" DevicePath \"\"" Nov 21 14:30:24 crc kubenswrapper[4897]: I1121 14:30:24.716472 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-95d1-account-create-7vtq6" Nov 21 14:30:24 crc kubenswrapper[4897]: I1121 14:30:24.817397 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-824zh\" (UniqueName: \"kubernetes.io/projected/5d1468bb-2545-4906-a10c-c95ba9ef8693-kube-api-access-824zh\") pod \"5d1468bb-2545-4906-a10c-c95ba9ef8693\" (UID: \"5d1468bb-2545-4906-a10c-c95ba9ef8693\") " Nov 21 14:30:24 crc kubenswrapper[4897]: I1121 14:30:24.817493 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5d1468bb-2545-4906-a10c-c95ba9ef8693-operator-scripts\") pod \"5d1468bb-2545-4906-a10c-c95ba9ef8693\" (UID: \"5d1468bb-2545-4906-a10c-c95ba9ef8693\") " Nov 21 14:30:24 crc kubenswrapper[4897]: I1121 14:30:24.817699 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vn2gx\" (UniqueName: \"kubernetes.io/projected/eb1688c2-d3d5-4935-b08b-32ef2ce7f6e1-kube-api-access-vn2gx\") pod \"eb1688c2-d3d5-4935-b08b-32ef2ce7f6e1\" (UID: \"eb1688c2-d3d5-4935-b08b-32ef2ce7f6e1\") " Nov 21 14:30:24 crc kubenswrapper[4897]: I1121 14:30:24.817747 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eb1688c2-d3d5-4935-b08b-32ef2ce7f6e1-operator-scripts\") pod \"eb1688c2-d3d5-4935-b08b-32ef2ce7f6e1\" (UID: \"eb1688c2-d3d5-4935-b08b-32ef2ce7f6e1\") " Nov 21 14:30:24 crc kubenswrapper[4897]: I1121 14:30:24.818957 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb1688c2-d3d5-4935-b08b-32ef2ce7f6e1-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "eb1688c2-d3d5-4935-b08b-32ef2ce7f6e1" (UID: "eb1688c2-d3d5-4935-b08b-32ef2ce7f6e1"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:30:24 crc kubenswrapper[4897]: I1121 14:30:24.819028 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5d1468bb-2545-4906-a10c-c95ba9ef8693-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5d1468bb-2545-4906-a10c-c95ba9ef8693" (UID: "5d1468bb-2545-4906-a10c-c95ba9ef8693"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:30:24 crc kubenswrapper[4897]: I1121 14:30:24.867835 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb1688c2-d3d5-4935-b08b-32ef2ce7f6e1-kube-api-access-vn2gx" (OuterVolumeSpecName: "kube-api-access-vn2gx") pod "eb1688c2-d3d5-4935-b08b-32ef2ce7f6e1" (UID: "eb1688c2-d3d5-4935-b08b-32ef2ce7f6e1"). InnerVolumeSpecName "kube-api-access-vn2gx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:30:24 crc kubenswrapper[4897]: I1121 14:30:24.885724 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d1468bb-2545-4906-a10c-c95ba9ef8693-kube-api-access-824zh" (OuterVolumeSpecName: "kube-api-access-824zh") pod "5d1468bb-2545-4906-a10c-c95ba9ef8693" (UID: "5d1468bb-2545-4906-a10c-c95ba9ef8693"). InnerVolumeSpecName "kube-api-access-824zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:30:24 crc kubenswrapper[4897]: I1121 14:30:24.919907 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-824zh\" (UniqueName: \"kubernetes.io/projected/5d1468bb-2545-4906-a10c-c95ba9ef8693-kube-api-access-824zh\") on node \"crc\" DevicePath \"\"" Nov 21 14:30:24 crc kubenswrapper[4897]: I1121 14:30:24.919946 4897 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5d1468bb-2545-4906-a10c-c95ba9ef8693-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 21 14:30:24 crc kubenswrapper[4897]: I1121 14:30:24.919962 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vn2gx\" (UniqueName: \"kubernetes.io/projected/eb1688c2-d3d5-4935-b08b-32ef2ce7f6e1-kube-api-access-vn2gx\") on node \"crc\" DevicePath \"\"" Nov 21 14:30:24 crc kubenswrapper[4897]: I1121 14:30:24.919974 4897 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eb1688c2-d3d5-4935-b08b-32ef2ce7f6e1-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 21 14:30:25 crc kubenswrapper[4897]: I1121 14:30:25.046816 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-t9nzd" event={"ID":"6d1f75d9-2eaf-44c1-b2d0-dae3a8547433","Type":"ContainerDied","Data":"a73917e008195dc7f69097b0ca104e56ee86b44a4ae10413b142f977531470f8"} Nov 21 14:30:25 crc kubenswrapper[4897]: I1121 14:30:25.046853 4897 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a73917e008195dc7f69097b0ca104e56ee86b44a4ae10413b142f977531470f8" Nov 21 14:30:25 crc kubenswrapper[4897]: I1121 14:30:25.046861 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-t9nzd" Nov 21 14:30:25 crc kubenswrapper[4897]: I1121 14:30:25.050042 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-95d1-account-create-7vtq6" Nov 21 14:30:25 crc kubenswrapper[4897]: I1121 14:30:25.050622 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-95d1-account-create-7vtq6" event={"ID":"5d1468bb-2545-4906-a10c-c95ba9ef8693","Type":"ContainerDied","Data":"59e1f2a197b68bfefe8b2456c8c419d86f697a9a44431a02e3e85fd09bbfe751"} Nov 21 14:30:25 crc kubenswrapper[4897]: I1121 14:30:25.050652 4897 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="59e1f2a197b68bfefe8b2456c8c419d86f697a9a44431a02e3e85fd09bbfe751" Nov 21 14:30:25 crc kubenswrapper[4897]: I1121 14:30:25.053020 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-db-create-z8nrd" event={"ID":"eb1688c2-d3d5-4935-b08b-32ef2ce7f6e1","Type":"ContainerDied","Data":"58c7b6743fa1ab72410a2524ce920b80013ab6e28a954772e19585e08b8cfc0c"} Nov 21 14:30:25 crc kubenswrapper[4897]: I1121 14:30:25.053038 4897 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="58c7b6743fa1ab72410a2524ce920b80013ab6e28a954772e19585e08b8cfc0c" Nov 21 14:30:25 crc kubenswrapper[4897]: I1121 14:30:25.053086 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-db-create-z8nrd" Nov 21 14:30:25 crc kubenswrapper[4897]: I1121 14:30:25.054539 4897 generic.go:334] "Generic (PLEG): container finished" podID="b513e4ee-7c30-4942-b997-4d52414e1ec0" containerID="3140f08ac469f99b17c2aa56cad1a0ccf3db6d5b4c5110f9454241ae6294b460" exitCode=0 Nov 21 14:30:25 crc kubenswrapper[4897]: I1121 14:30:25.054563 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"b513e4ee-7c30-4942-b997-4d52414e1ec0","Type":"ContainerDied","Data":"3140f08ac469f99b17c2aa56cad1a0ccf3db6d5b4c5110f9454241ae6294b460"} Nov 21 14:30:27 crc kubenswrapper[4897]: I1121 14:30:27.074449 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"5f9bf5d1-dcbd-4d19-b46d-22810ebf17a2","Type":"ContainerStarted","Data":"41a1b96fb01d04c85066a213791dae1d85fa945959a10f489db1a8576c224361"} Nov 21 14:30:27 crc kubenswrapper[4897]: I1121 14:30:27.077225 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"b513e4ee-7c30-4942-b997-4d52414e1ec0","Type":"ContainerStarted","Data":"4ded3f218b2e102f2165dac36295117473edb574550d19babb7b246c41c8316f"} Nov 21 14:30:27 crc kubenswrapper[4897]: I1121 14:30:27.078455 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Nov 21 14:30:27 crc kubenswrapper[4897]: I1121 14:30:27.091464 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"91aa2270-92a7-4af4-9a92-141ea2b24ee1","Type":"ContainerStarted","Data":"d75dc39a201e4f83e8ba5e273e81d6969573634dc7a6223e72bea5261f0bd8db"} Nov 21 14:30:27 crc kubenswrapper[4897]: I1121 14:30:27.100413 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=39.054985092 podStartE2EDuration="1m24.10039119s" podCreationTimestamp="2025-11-21 14:29:03 +0000 UTC" firstStartedPulling="2025-11-21 14:29:05.594713639 +0000 UTC m=+1222.879307114" lastFinishedPulling="2025-11-21 14:29:50.640119737 +0000 UTC m=+1267.924713212" observedRunningTime="2025-11-21 14:30:27.099342972 +0000 UTC m=+1304.383936447" watchObservedRunningTime="2025-11-21 14:30:27.10039119 +0000 UTC m=+1304.384984665" Nov 21 14:30:27 crc kubenswrapper[4897]: I1121 14:30:27.134651 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=32.91182747 podStartE2EDuration="1m17.134634439s" podCreationTimestamp="2025-11-21 14:29:10 +0000 UTC" firstStartedPulling="2025-11-21 14:29:42.399119517 +0000 UTC m=+1259.683713002" lastFinishedPulling="2025-11-21 14:30:26.621926486 +0000 UTC m=+1303.906519971" observedRunningTime="2025-11-21 14:30:27.13020312 +0000 UTC m=+1304.414796605" watchObservedRunningTime="2025-11-21 14:30:27.134634439 +0000 UTC m=+1304.419227914" Nov 21 14:30:28 crc kubenswrapper[4897]: I1121 14:30:28.102827 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"5f9bf5d1-dcbd-4d19-b46d-22810ebf17a2","Type":"ContainerStarted","Data":"aa2095452dcf1604e8b8eaefecbd03f9059bd05ddcfae414985cf8d20666994f"} Nov 21 14:30:28 crc kubenswrapper[4897]: I1121 14:30:28.103164 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"5f9bf5d1-dcbd-4d19-b46d-22810ebf17a2","Type":"ContainerStarted","Data":"50fc34d1598e6c5d576cda9a4d22df7eb34c45a1f9522a4e3c7ed8b076510f3d"} Nov 21 14:30:28 crc kubenswrapper[4897]: I1121 14:30:28.439175 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-mm6tz" podUID="e0c27889-eea7-4a06-9195-3401929cf6b2" containerName="ovn-controller" probeResult="failure" output=< Nov 21 14:30:28 crc kubenswrapper[4897]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Nov 21 14:30:28 crc kubenswrapper[4897]: > Nov 21 14:30:28 crc kubenswrapper[4897]: I1121 14:30:28.452761 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-k28mx" Nov 21 14:30:28 crc kubenswrapper[4897]: I1121 14:30:28.661961 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-mm6tz-config-t2x4t"] Nov 21 14:30:28 crc kubenswrapper[4897]: E1121 14:30:28.662334 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="994c36f9-336e-49dc-b45d-0eb4bfa0b686" containerName="mariadb-account-create" Nov 21 14:30:28 crc kubenswrapper[4897]: I1121 14:30:28.662350 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="994c36f9-336e-49dc-b45d-0eb4bfa0b686" containerName="mariadb-account-create" Nov 21 14:30:28 crc kubenswrapper[4897]: E1121 14:30:28.662362 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb1688c2-d3d5-4935-b08b-32ef2ce7f6e1" containerName="mariadb-database-create" Nov 21 14:30:28 crc kubenswrapper[4897]: I1121 14:30:28.662369 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb1688c2-d3d5-4935-b08b-32ef2ce7f6e1" containerName="mariadb-database-create" Nov 21 14:30:28 crc kubenswrapper[4897]: E1121 14:30:28.662380 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7501a167-0655-4ad4-bb04-c597a81359ba" containerName="mariadb-database-create" Nov 21 14:30:28 crc kubenswrapper[4897]: I1121 14:30:28.662386 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="7501a167-0655-4ad4-bb04-c597a81359ba" containerName="mariadb-database-create" Nov 21 14:30:28 crc kubenswrapper[4897]: E1121 14:30:28.662395 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d1468bb-2545-4906-a10c-c95ba9ef8693" containerName="mariadb-account-create" Nov 21 14:30:28 crc kubenswrapper[4897]: I1121 14:30:28.662401 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d1468bb-2545-4906-a10c-c95ba9ef8693" containerName="mariadb-account-create" Nov 21 14:30:28 crc kubenswrapper[4897]: E1121 14:30:28.662411 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d1f75d9-2eaf-44c1-b2d0-dae3a8547433" containerName="swift-ring-rebalance" Nov 21 14:30:28 crc kubenswrapper[4897]: I1121 14:30:28.662417 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d1f75d9-2eaf-44c1-b2d0-dae3a8547433" containerName="swift-ring-rebalance" Nov 21 14:30:28 crc kubenswrapper[4897]: E1121 14:30:28.662431 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3441939-1c13-4b55-a129-a39f4bed36cd" containerName="mariadb-database-create" Nov 21 14:30:28 crc kubenswrapper[4897]: I1121 14:30:28.662439 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3441939-1c13-4b55-a129-a39f4bed36cd" containerName="mariadb-database-create" Nov 21 14:30:28 crc kubenswrapper[4897]: E1121 14:30:28.662461 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34a64a5d-5b9d-4aac-be3e-96edc721fa3b" containerName="mariadb-account-create" Nov 21 14:30:28 crc kubenswrapper[4897]: I1121 14:30:28.662467 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="34a64a5d-5b9d-4aac-be3e-96edc721fa3b" containerName="mariadb-account-create" Nov 21 14:30:28 crc kubenswrapper[4897]: E1121 14:30:28.662477 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="980c0de3-d435-439f-8477-0dc68fbaf7f0" containerName="mariadb-account-create" Nov 21 14:30:28 crc kubenswrapper[4897]: I1121 14:30:28.662482 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="980c0de3-d435-439f-8477-0dc68fbaf7f0" containerName="mariadb-account-create" Nov 21 14:30:28 crc kubenswrapper[4897]: E1121 14:30:28.662496 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7a5e302-fb61-41b2-a7d2-ff02160c6679" containerName="mariadb-database-create" Nov 21 14:30:28 crc kubenswrapper[4897]: I1121 14:30:28.662517 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7a5e302-fb61-41b2-a7d2-ff02160c6679" containerName="mariadb-database-create" Nov 21 14:30:28 crc kubenswrapper[4897]: I1121 14:30:28.662679 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="34a64a5d-5b9d-4aac-be3e-96edc721fa3b" containerName="mariadb-account-create" Nov 21 14:30:28 crc kubenswrapper[4897]: I1121 14:30:28.662698 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d1468bb-2545-4906-a10c-c95ba9ef8693" containerName="mariadb-account-create" Nov 21 14:30:28 crc kubenswrapper[4897]: I1121 14:30:28.662706 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3441939-1c13-4b55-a129-a39f4bed36cd" containerName="mariadb-database-create" Nov 21 14:30:28 crc kubenswrapper[4897]: I1121 14:30:28.662717 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d1f75d9-2eaf-44c1-b2d0-dae3a8547433" containerName="swift-ring-rebalance" Nov 21 14:30:28 crc kubenswrapper[4897]: I1121 14:30:28.662724 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="994c36f9-336e-49dc-b45d-0eb4bfa0b686" containerName="mariadb-account-create" Nov 21 14:30:28 crc kubenswrapper[4897]: I1121 14:30:28.662733 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="7501a167-0655-4ad4-bb04-c597a81359ba" containerName="mariadb-database-create" Nov 21 14:30:28 crc kubenswrapper[4897]: I1121 14:30:28.662743 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7a5e302-fb61-41b2-a7d2-ff02160c6679" containerName="mariadb-database-create" Nov 21 14:30:28 crc kubenswrapper[4897]: I1121 14:30:28.662754 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="980c0de3-d435-439f-8477-0dc68fbaf7f0" containerName="mariadb-account-create" Nov 21 14:30:28 crc kubenswrapper[4897]: I1121 14:30:28.662763 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb1688c2-d3d5-4935-b08b-32ef2ce7f6e1" containerName="mariadb-database-create" Nov 21 14:30:28 crc kubenswrapper[4897]: I1121 14:30:28.663401 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-mm6tz-config-t2x4t" Nov 21 14:30:28 crc kubenswrapper[4897]: I1121 14:30:28.668637 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Nov 21 14:30:28 crc kubenswrapper[4897]: I1121 14:30:28.677138 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-mm6tz-config-t2x4t"] Nov 21 14:30:28 crc kubenswrapper[4897]: I1121 14:30:28.736183 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-zz6kx"] Nov 21 14:30:28 crc kubenswrapper[4897]: I1121 14:30:28.740635 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-zz6kx" Nov 21 14:30:28 crc kubenswrapper[4897]: I1121 14:30:28.743143 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-wk7cs" Nov 21 14:30:28 crc kubenswrapper[4897]: I1121 14:30:28.747251 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Nov 21 14:30:28 crc kubenswrapper[4897]: I1121 14:30:28.749062 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-zz6kx"] Nov 21 14:30:28 crc kubenswrapper[4897]: I1121 14:30:28.805240 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bfe51641-f509-4880-b8bd-9ce5da1e2291-scripts\") pod \"ovn-controller-mm6tz-config-t2x4t\" (UID: \"bfe51641-f509-4880-b8bd-9ce5da1e2291\") " pod="openstack/ovn-controller-mm6tz-config-t2x4t" Nov 21 14:30:28 crc kubenswrapper[4897]: I1121 14:30:28.805553 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/bfe51641-f509-4880-b8bd-9ce5da1e2291-var-run-ovn\") pod \"ovn-controller-mm6tz-config-t2x4t\" (UID: \"bfe51641-f509-4880-b8bd-9ce5da1e2291\") " pod="openstack/ovn-controller-mm6tz-config-t2x4t" Nov 21 14:30:28 crc kubenswrapper[4897]: I1121 14:30:28.805734 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/bfe51641-f509-4880-b8bd-9ce5da1e2291-additional-scripts\") pod \"ovn-controller-mm6tz-config-t2x4t\" (UID: \"bfe51641-f509-4880-b8bd-9ce5da1e2291\") " pod="openstack/ovn-controller-mm6tz-config-t2x4t" Nov 21 14:30:28 crc kubenswrapper[4897]: I1121 14:30:28.805888 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/bfe51641-f509-4880-b8bd-9ce5da1e2291-var-log-ovn\") pod \"ovn-controller-mm6tz-config-t2x4t\" (UID: \"bfe51641-f509-4880-b8bd-9ce5da1e2291\") " pod="openstack/ovn-controller-mm6tz-config-t2x4t" Nov 21 14:30:28 crc kubenswrapper[4897]: I1121 14:30:28.806109 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/bfe51641-f509-4880-b8bd-9ce5da1e2291-var-run\") pod \"ovn-controller-mm6tz-config-t2x4t\" (UID: \"bfe51641-f509-4880-b8bd-9ce5da1e2291\") " pod="openstack/ovn-controller-mm6tz-config-t2x4t" Nov 21 14:30:28 crc kubenswrapper[4897]: I1121 14:30:28.806296 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6g8tx\" (UniqueName: \"kubernetes.io/projected/bfe51641-f509-4880-b8bd-9ce5da1e2291-kube-api-access-6g8tx\") pod \"ovn-controller-mm6tz-config-t2x4t\" (UID: \"bfe51641-f509-4880-b8bd-9ce5da1e2291\") " pod="openstack/ovn-controller-mm6tz-config-t2x4t" Nov 21 14:30:28 crc kubenswrapper[4897]: I1121 14:30:28.908560 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c08d8b3-d052-47b7-ad42-7b0f9fbe1920-config-data\") pod \"glance-db-sync-zz6kx\" (UID: \"9c08d8b3-d052-47b7-ad42-7b0f9fbe1920\") " pod="openstack/glance-db-sync-zz6kx" Nov 21 14:30:28 crc kubenswrapper[4897]: I1121 14:30:28.908630 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/bfe51641-f509-4880-b8bd-9ce5da1e2291-var-run\") pod \"ovn-controller-mm6tz-config-t2x4t\" (UID: \"bfe51641-f509-4880-b8bd-9ce5da1e2291\") " pod="openstack/ovn-controller-mm6tz-config-t2x4t" Nov 21 14:30:28 crc kubenswrapper[4897]: I1121 14:30:28.908669 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6g8tx\" (UniqueName: \"kubernetes.io/projected/bfe51641-f509-4880-b8bd-9ce5da1e2291-kube-api-access-6g8tx\") pod \"ovn-controller-mm6tz-config-t2x4t\" (UID: \"bfe51641-f509-4880-b8bd-9ce5da1e2291\") " pod="openstack/ovn-controller-mm6tz-config-t2x4t" Nov 21 14:30:28 crc kubenswrapper[4897]: I1121 14:30:28.908699 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gj9bx\" (UniqueName: \"kubernetes.io/projected/9c08d8b3-d052-47b7-ad42-7b0f9fbe1920-kube-api-access-gj9bx\") pod \"glance-db-sync-zz6kx\" (UID: \"9c08d8b3-d052-47b7-ad42-7b0f9fbe1920\") " pod="openstack/glance-db-sync-zz6kx" Nov 21 14:30:28 crc kubenswrapper[4897]: I1121 14:30:28.908761 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/bfe51641-f509-4880-b8bd-9ce5da1e2291-var-run-ovn\") pod \"ovn-controller-mm6tz-config-t2x4t\" (UID: \"bfe51641-f509-4880-b8bd-9ce5da1e2291\") " pod="openstack/ovn-controller-mm6tz-config-t2x4t" Nov 21 14:30:28 crc kubenswrapper[4897]: I1121 14:30:28.908776 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bfe51641-f509-4880-b8bd-9ce5da1e2291-scripts\") pod \"ovn-controller-mm6tz-config-t2x4t\" (UID: \"bfe51641-f509-4880-b8bd-9ce5da1e2291\") " pod="openstack/ovn-controller-mm6tz-config-t2x4t" Nov 21 14:30:28 crc kubenswrapper[4897]: I1121 14:30:28.908818 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/bfe51641-f509-4880-b8bd-9ce5da1e2291-additional-scripts\") pod \"ovn-controller-mm6tz-config-t2x4t\" (UID: \"bfe51641-f509-4880-b8bd-9ce5da1e2291\") " pod="openstack/ovn-controller-mm6tz-config-t2x4t" Nov 21 14:30:28 crc kubenswrapper[4897]: I1121 14:30:28.908854 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/bfe51641-f509-4880-b8bd-9ce5da1e2291-var-log-ovn\") pod \"ovn-controller-mm6tz-config-t2x4t\" (UID: \"bfe51641-f509-4880-b8bd-9ce5da1e2291\") " pod="openstack/ovn-controller-mm6tz-config-t2x4t" Nov 21 14:30:28 crc kubenswrapper[4897]: I1121 14:30:28.908890 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/9c08d8b3-d052-47b7-ad42-7b0f9fbe1920-db-sync-config-data\") pod \"glance-db-sync-zz6kx\" (UID: \"9c08d8b3-d052-47b7-ad42-7b0f9fbe1920\") " pod="openstack/glance-db-sync-zz6kx" Nov 21 14:30:28 crc kubenswrapper[4897]: I1121 14:30:28.908909 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c08d8b3-d052-47b7-ad42-7b0f9fbe1920-combined-ca-bundle\") pod \"glance-db-sync-zz6kx\" (UID: \"9c08d8b3-d052-47b7-ad42-7b0f9fbe1920\") " pod="openstack/glance-db-sync-zz6kx" Nov 21 14:30:28 crc kubenswrapper[4897]: I1121 14:30:28.909108 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/bfe51641-f509-4880-b8bd-9ce5da1e2291-var-run-ovn\") pod \"ovn-controller-mm6tz-config-t2x4t\" (UID: \"bfe51641-f509-4880-b8bd-9ce5da1e2291\") " pod="openstack/ovn-controller-mm6tz-config-t2x4t" Nov 21 14:30:28 crc kubenswrapper[4897]: I1121 14:30:28.909131 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/bfe51641-f509-4880-b8bd-9ce5da1e2291-var-run\") pod \"ovn-controller-mm6tz-config-t2x4t\" (UID: \"bfe51641-f509-4880-b8bd-9ce5da1e2291\") " pod="openstack/ovn-controller-mm6tz-config-t2x4t" Nov 21 14:30:28 crc kubenswrapper[4897]: I1121 14:30:28.909125 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/bfe51641-f509-4880-b8bd-9ce5da1e2291-var-log-ovn\") pod \"ovn-controller-mm6tz-config-t2x4t\" (UID: \"bfe51641-f509-4880-b8bd-9ce5da1e2291\") " pod="openstack/ovn-controller-mm6tz-config-t2x4t" Nov 21 14:30:28 crc kubenswrapper[4897]: I1121 14:30:28.909865 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/bfe51641-f509-4880-b8bd-9ce5da1e2291-additional-scripts\") pod \"ovn-controller-mm6tz-config-t2x4t\" (UID: \"bfe51641-f509-4880-b8bd-9ce5da1e2291\") " pod="openstack/ovn-controller-mm6tz-config-t2x4t" Nov 21 14:30:28 crc kubenswrapper[4897]: I1121 14:30:28.911354 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bfe51641-f509-4880-b8bd-9ce5da1e2291-scripts\") pod \"ovn-controller-mm6tz-config-t2x4t\" (UID: \"bfe51641-f509-4880-b8bd-9ce5da1e2291\") " pod="openstack/ovn-controller-mm6tz-config-t2x4t" Nov 21 14:30:28 crc kubenswrapper[4897]: I1121 14:30:28.940645 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6g8tx\" (UniqueName: \"kubernetes.io/projected/bfe51641-f509-4880-b8bd-9ce5da1e2291-kube-api-access-6g8tx\") pod \"ovn-controller-mm6tz-config-t2x4t\" (UID: \"bfe51641-f509-4880-b8bd-9ce5da1e2291\") " pod="openstack/ovn-controller-mm6tz-config-t2x4t" Nov 21 14:30:28 crc kubenswrapper[4897]: I1121 14:30:28.983125 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-mm6tz-config-t2x4t" Nov 21 14:30:29 crc kubenswrapper[4897]: I1121 14:30:29.010601 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/9c08d8b3-d052-47b7-ad42-7b0f9fbe1920-db-sync-config-data\") pod \"glance-db-sync-zz6kx\" (UID: \"9c08d8b3-d052-47b7-ad42-7b0f9fbe1920\") " pod="openstack/glance-db-sync-zz6kx" Nov 21 14:30:29 crc kubenswrapper[4897]: I1121 14:30:29.010654 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c08d8b3-d052-47b7-ad42-7b0f9fbe1920-combined-ca-bundle\") pod \"glance-db-sync-zz6kx\" (UID: \"9c08d8b3-d052-47b7-ad42-7b0f9fbe1920\") " pod="openstack/glance-db-sync-zz6kx" Nov 21 14:30:29 crc kubenswrapper[4897]: I1121 14:30:29.010717 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c08d8b3-d052-47b7-ad42-7b0f9fbe1920-config-data\") pod \"glance-db-sync-zz6kx\" (UID: \"9c08d8b3-d052-47b7-ad42-7b0f9fbe1920\") " pod="openstack/glance-db-sync-zz6kx" Nov 21 14:30:29 crc kubenswrapper[4897]: I1121 14:30:29.010783 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gj9bx\" (UniqueName: \"kubernetes.io/projected/9c08d8b3-d052-47b7-ad42-7b0f9fbe1920-kube-api-access-gj9bx\") pod \"glance-db-sync-zz6kx\" (UID: \"9c08d8b3-d052-47b7-ad42-7b0f9fbe1920\") " pod="openstack/glance-db-sync-zz6kx" Nov 21 14:30:29 crc kubenswrapper[4897]: I1121 14:30:29.016211 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c08d8b3-d052-47b7-ad42-7b0f9fbe1920-config-data\") pod \"glance-db-sync-zz6kx\" (UID: \"9c08d8b3-d052-47b7-ad42-7b0f9fbe1920\") " pod="openstack/glance-db-sync-zz6kx" Nov 21 14:30:29 crc kubenswrapper[4897]: I1121 14:30:29.016475 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/9c08d8b3-d052-47b7-ad42-7b0f9fbe1920-db-sync-config-data\") pod \"glance-db-sync-zz6kx\" (UID: \"9c08d8b3-d052-47b7-ad42-7b0f9fbe1920\") " pod="openstack/glance-db-sync-zz6kx" Nov 21 14:30:29 crc kubenswrapper[4897]: I1121 14:30:29.026292 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c08d8b3-d052-47b7-ad42-7b0f9fbe1920-combined-ca-bundle\") pod \"glance-db-sync-zz6kx\" (UID: \"9c08d8b3-d052-47b7-ad42-7b0f9fbe1920\") " pod="openstack/glance-db-sync-zz6kx" Nov 21 14:30:29 crc kubenswrapper[4897]: I1121 14:30:29.043450 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gj9bx\" (UniqueName: \"kubernetes.io/projected/9c08d8b3-d052-47b7-ad42-7b0f9fbe1920-kube-api-access-gj9bx\") pod \"glance-db-sync-zz6kx\" (UID: \"9c08d8b3-d052-47b7-ad42-7b0f9fbe1920\") " pod="openstack/glance-db-sync-zz6kx" Nov 21 14:30:29 crc kubenswrapper[4897]: I1121 14:30:29.058196 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-zz6kx" Nov 21 14:30:29 crc kubenswrapper[4897]: I1121 14:30:29.135810 4897 generic.go:334] "Generic (PLEG): container finished" podID="bd692188-6da6-4387-a46d-003a2da0d0c8" containerID="5e97248a8dc790521af481c0714596ab753484b35fa256f425d03e7bac6e870a" exitCode=0 Nov 21 14:30:29 crc kubenswrapper[4897]: I1121 14:30:29.135852 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"bd692188-6da6-4387-a46d-003a2da0d0c8","Type":"ContainerDied","Data":"5e97248a8dc790521af481c0714596ab753484b35fa256f425d03e7bac6e870a"} Nov 21 14:30:29 crc kubenswrapper[4897]: I1121 14:30:29.476837 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-mm6tz-config-t2x4t"] Nov 21 14:30:29 crc kubenswrapper[4897]: W1121 14:30:29.483179 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbfe51641_f509_4880_b8bd_9ce5da1e2291.slice/crio-43f7472e727eb88c81e6a6626e4dbc3452dd03cca036a3b42bfe4060716ebcba WatchSource:0}: Error finding container 43f7472e727eb88c81e6a6626e4dbc3452dd03cca036a3b42bfe4060716ebcba: Status 404 returned error can't find the container with id 43f7472e727eb88c81e6a6626e4dbc3452dd03cca036a3b42bfe4060716ebcba Nov 21 14:30:29 crc kubenswrapper[4897]: I1121 14:30:29.752573 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-zz6kx"] Nov 21 14:30:29 crc kubenswrapper[4897]: W1121 14:30:29.756554 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9c08d8b3_d052_47b7_ad42_7b0f9fbe1920.slice/crio-28c6286341a9a2739d5cbbd8ea2338b86f661821a8bfd60f1c227849a4331e17 WatchSource:0}: Error finding container 28c6286341a9a2739d5cbbd8ea2338b86f661821a8bfd60f1c227849a4331e17: Status 404 returned error can't find the container with id 28c6286341a9a2739d5cbbd8ea2338b86f661821a8bfd60f1c227849a4331e17 Nov 21 14:30:30 crc kubenswrapper[4897]: I1121 14:30:30.146098 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-zz6kx" event={"ID":"9c08d8b3-d052-47b7-ad42-7b0f9fbe1920","Type":"ContainerStarted","Data":"28c6286341a9a2739d5cbbd8ea2338b86f661821a8bfd60f1c227849a4331e17"} Nov 21 14:30:30 crc kubenswrapper[4897]: I1121 14:30:30.148745 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"5f9bf5d1-dcbd-4d19-b46d-22810ebf17a2","Type":"ContainerStarted","Data":"6145b41b2923d75a84011c32ff8930f51bdc0f59959e17ad7c282919c87b212c"} Nov 21 14:30:30 crc kubenswrapper[4897]: I1121 14:30:30.150609 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"bd692188-6da6-4387-a46d-003a2da0d0c8","Type":"ContainerStarted","Data":"99656890d61e67a0d062b6637a03db823a2e86706363f8d471ea687b544eb8ea"} Nov 21 14:30:30 crc kubenswrapper[4897]: I1121 14:30:30.150825 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Nov 21 14:30:30 crc kubenswrapper[4897]: I1121 14:30:30.152239 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-mm6tz-config-t2x4t" event={"ID":"bfe51641-f509-4880-b8bd-9ce5da1e2291","Type":"ContainerStarted","Data":"968d88e3eb6d022e132ec663aa8194b45efff34819cb40fb992e07daba2fd847"} Nov 21 14:30:30 crc kubenswrapper[4897]: I1121 14:30:30.152283 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-mm6tz-config-t2x4t" event={"ID":"bfe51641-f509-4880-b8bd-9ce5da1e2291","Type":"ContainerStarted","Data":"43f7472e727eb88c81e6a6626e4dbc3452dd03cca036a3b42bfe4060716ebcba"} Nov 21 14:30:30 crc kubenswrapper[4897]: I1121 14:30:30.174895 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=-9223371950.679909 podStartE2EDuration="1m26.174866158s" podCreationTimestamp="2025-11-21 14:29:04 +0000 UTC" firstStartedPulling="2025-11-21 14:29:06.004588218 +0000 UTC m=+1223.289181693" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:30:30.173105601 +0000 UTC m=+1307.457699076" watchObservedRunningTime="2025-11-21 14:30:30.174866158 +0000 UTC m=+1307.459459673" Nov 21 14:30:30 crc kubenswrapper[4897]: I1121 14:30:30.538424 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-mm6tz-config-t2x4t" podStartSLOduration=2.53840163 podStartE2EDuration="2.53840163s" podCreationTimestamp="2025-11-21 14:30:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:30:30.19842102 +0000 UTC m=+1307.483014515" watchObservedRunningTime="2025-11-21 14:30:30.53840163 +0000 UTC m=+1307.822995115" Nov 21 14:30:30 crc kubenswrapper[4897]: I1121 14:30:30.548188 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mysqld-exporter-openstack-cell1-db-create-jb9vc"] Nov 21 14:30:30 crc kubenswrapper[4897]: I1121 14:30:30.549952 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-cell1-db-create-jb9vc" Nov 21 14:30:30 crc kubenswrapper[4897]: I1121 14:30:30.574775 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-openstack-cell1-db-create-jb9vc"] Nov 21 14:30:30 crc kubenswrapper[4897]: I1121 14:30:30.641573 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mysqld-exporter-b245-account-create-pprkv"] Nov 21 14:30:30 crc kubenswrapper[4897]: I1121 14:30:30.642946 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-b245-account-create-pprkv" Nov 21 14:30:30 crc kubenswrapper[4897]: I1121 14:30:30.644842 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"mysqld-exporter-openstack-cell1-db-secret" Nov 21 14:30:30 crc kubenswrapper[4897]: I1121 14:30:30.645473 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0722112c-c860-4cb0-ac80-9cf32bd12f39-operator-scripts\") pod \"mysqld-exporter-openstack-cell1-db-create-jb9vc\" (UID: \"0722112c-c860-4cb0-ac80-9cf32bd12f39\") " pod="openstack/mysqld-exporter-openstack-cell1-db-create-jb9vc" Nov 21 14:30:30 crc kubenswrapper[4897]: I1121 14:30:30.645682 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q2kxm\" (UniqueName: \"kubernetes.io/projected/0722112c-c860-4cb0-ac80-9cf32bd12f39-kube-api-access-q2kxm\") pod \"mysqld-exporter-openstack-cell1-db-create-jb9vc\" (UID: \"0722112c-c860-4cb0-ac80-9cf32bd12f39\") " pod="openstack/mysqld-exporter-openstack-cell1-db-create-jb9vc" Nov 21 14:30:30 crc kubenswrapper[4897]: I1121 14:30:30.666694 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-b245-account-create-pprkv"] Nov 21 14:30:30 crc kubenswrapper[4897]: I1121 14:30:30.747223 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d56425f9-2cc4-4936-bda7-64520e84f5ab-operator-scripts\") pod \"mysqld-exporter-b245-account-create-pprkv\" (UID: \"d56425f9-2cc4-4936-bda7-64520e84f5ab\") " pod="openstack/mysqld-exporter-b245-account-create-pprkv" Nov 21 14:30:30 crc kubenswrapper[4897]: I1121 14:30:30.747294 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q2kxm\" (UniqueName: \"kubernetes.io/projected/0722112c-c860-4cb0-ac80-9cf32bd12f39-kube-api-access-q2kxm\") pod \"mysqld-exporter-openstack-cell1-db-create-jb9vc\" (UID: \"0722112c-c860-4cb0-ac80-9cf32bd12f39\") " pod="openstack/mysqld-exporter-openstack-cell1-db-create-jb9vc" Nov 21 14:30:30 crc kubenswrapper[4897]: I1121 14:30:30.747328 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sh4v4\" (UniqueName: \"kubernetes.io/projected/d56425f9-2cc4-4936-bda7-64520e84f5ab-kube-api-access-sh4v4\") pod \"mysqld-exporter-b245-account-create-pprkv\" (UID: \"d56425f9-2cc4-4936-bda7-64520e84f5ab\") " pod="openstack/mysqld-exporter-b245-account-create-pprkv" Nov 21 14:30:30 crc kubenswrapper[4897]: I1121 14:30:30.747350 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0722112c-c860-4cb0-ac80-9cf32bd12f39-operator-scripts\") pod \"mysqld-exporter-openstack-cell1-db-create-jb9vc\" (UID: \"0722112c-c860-4cb0-ac80-9cf32bd12f39\") " pod="openstack/mysqld-exporter-openstack-cell1-db-create-jb9vc" Nov 21 14:30:30 crc kubenswrapper[4897]: I1121 14:30:30.748088 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0722112c-c860-4cb0-ac80-9cf32bd12f39-operator-scripts\") pod \"mysqld-exporter-openstack-cell1-db-create-jb9vc\" (UID: \"0722112c-c860-4cb0-ac80-9cf32bd12f39\") " pod="openstack/mysqld-exporter-openstack-cell1-db-create-jb9vc" Nov 21 14:30:30 crc kubenswrapper[4897]: I1121 14:30:30.782550 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q2kxm\" (UniqueName: \"kubernetes.io/projected/0722112c-c860-4cb0-ac80-9cf32bd12f39-kube-api-access-q2kxm\") pod \"mysqld-exporter-openstack-cell1-db-create-jb9vc\" (UID: \"0722112c-c860-4cb0-ac80-9cf32bd12f39\") " pod="openstack/mysqld-exporter-openstack-cell1-db-create-jb9vc" Nov 21 14:30:30 crc kubenswrapper[4897]: I1121 14:30:30.849204 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sh4v4\" (UniqueName: \"kubernetes.io/projected/d56425f9-2cc4-4936-bda7-64520e84f5ab-kube-api-access-sh4v4\") pod \"mysqld-exporter-b245-account-create-pprkv\" (UID: \"d56425f9-2cc4-4936-bda7-64520e84f5ab\") " pod="openstack/mysqld-exporter-b245-account-create-pprkv" Nov 21 14:30:30 crc kubenswrapper[4897]: I1121 14:30:30.849445 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d56425f9-2cc4-4936-bda7-64520e84f5ab-operator-scripts\") pod \"mysqld-exporter-b245-account-create-pprkv\" (UID: \"d56425f9-2cc4-4936-bda7-64520e84f5ab\") " pod="openstack/mysqld-exporter-b245-account-create-pprkv" Nov 21 14:30:30 crc kubenswrapper[4897]: I1121 14:30:30.850294 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d56425f9-2cc4-4936-bda7-64520e84f5ab-operator-scripts\") pod \"mysqld-exporter-b245-account-create-pprkv\" (UID: \"d56425f9-2cc4-4936-bda7-64520e84f5ab\") " pod="openstack/mysqld-exporter-b245-account-create-pprkv" Nov 21 14:30:30 crc kubenswrapper[4897]: I1121 14:30:30.872124 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sh4v4\" (UniqueName: \"kubernetes.io/projected/d56425f9-2cc4-4936-bda7-64520e84f5ab-kube-api-access-sh4v4\") pod \"mysqld-exporter-b245-account-create-pprkv\" (UID: \"d56425f9-2cc4-4936-bda7-64520e84f5ab\") " pod="openstack/mysqld-exporter-b245-account-create-pprkv" Nov 21 14:30:30 crc kubenswrapper[4897]: I1121 14:30:30.891743 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-cell1-db-create-jb9vc" Nov 21 14:30:30 crc kubenswrapper[4897]: I1121 14:30:30.977389 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-b245-account-create-pprkv" Nov 21 14:30:31 crc kubenswrapper[4897]: I1121 14:30:31.174493 4897 generic.go:334] "Generic (PLEG): container finished" podID="bfe51641-f509-4880-b8bd-9ce5da1e2291" containerID="968d88e3eb6d022e132ec663aa8194b45efff34819cb40fb992e07daba2fd847" exitCode=0 Nov 21 14:30:31 crc kubenswrapper[4897]: I1121 14:30:31.175018 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-mm6tz-config-t2x4t" event={"ID":"bfe51641-f509-4880-b8bd-9ce5da1e2291","Type":"ContainerDied","Data":"968d88e3eb6d022e132ec663aa8194b45efff34819cb40fb992e07daba2fd847"} Nov 21 14:30:31 crc kubenswrapper[4897]: I1121 14:30:31.403456 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-openstack-cell1-db-create-jb9vc"] Nov 21 14:30:31 crc kubenswrapper[4897]: I1121 14:30:31.500058 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Nov 21 14:30:31 crc kubenswrapper[4897]: I1121 14:30:31.581090 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-b245-account-create-pprkv"] Nov 21 14:30:31 crc kubenswrapper[4897]: W1121 14:30:31.581718 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0722112c_c860_4cb0_ac80_9cf32bd12f39.slice/crio-c645093c6b13b49e00b7759c283bd4d3841ff073ac4fc379a3239ad01b579249 WatchSource:0}: Error finding container c645093c6b13b49e00b7759c283bd4d3841ff073ac4fc379a3239ad01b579249: Status 404 returned error can't find the container with id c645093c6b13b49e00b7759c283bd4d3841ff073ac4fc379a3239ad01b579249 Nov 21 14:30:31 crc kubenswrapper[4897]: W1121 14:30:31.587941 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd56425f9_2cc4_4936_bda7_64520e84f5ab.slice/crio-1eb322ca974935a011b50872c735bc68c12fa072252c6669f1dcb40743753a00 WatchSource:0}: Error finding container 1eb322ca974935a011b50872c735bc68c12fa072252c6669f1dcb40743753a00: Status 404 returned error can't find the container with id 1eb322ca974935a011b50872c735bc68c12fa072252c6669f1dcb40743753a00 Nov 21 14:30:32 crc kubenswrapper[4897]: I1121 14:30:32.188924 4897 generic.go:334] "Generic (PLEG): container finished" podID="d56425f9-2cc4-4936-bda7-64520e84f5ab" containerID="980de6ce7e9fd815f36367eb491b1f46149c4541db0944212ac301ea1c5c30c1" exitCode=0 Nov 21 14:30:32 crc kubenswrapper[4897]: I1121 14:30:32.189313 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-b245-account-create-pprkv" event={"ID":"d56425f9-2cc4-4936-bda7-64520e84f5ab","Type":"ContainerDied","Data":"980de6ce7e9fd815f36367eb491b1f46149c4541db0944212ac301ea1c5c30c1"} Nov 21 14:30:32 crc kubenswrapper[4897]: I1121 14:30:32.189339 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-b245-account-create-pprkv" event={"ID":"d56425f9-2cc4-4936-bda7-64520e84f5ab","Type":"ContainerStarted","Data":"1eb322ca974935a011b50872c735bc68c12fa072252c6669f1dcb40743753a00"} Nov 21 14:30:32 crc kubenswrapper[4897]: I1121 14:30:32.197228 4897 generic.go:334] "Generic (PLEG): container finished" podID="0722112c-c860-4cb0-ac80-9cf32bd12f39" containerID="ca081cb9216b8b6726dc8d2233dd7e385f478246a9c0246f412237ec6aa83a6a" exitCode=0 Nov 21 14:30:32 crc kubenswrapper[4897]: I1121 14:30:32.197297 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-cell1-db-create-jb9vc" event={"ID":"0722112c-c860-4cb0-ac80-9cf32bd12f39","Type":"ContainerDied","Data":"ca081cb9216b8b6726dc8d2233dd7e385f478246a9c0246f412237ec6aa83a6a"} Nov 21 14:30:32 crc kubenswrapper[4897]: I1121 14:30:32.197317 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-cell1-db-create-jb9vc" event={"ID":"0722112c-c860-4cb0-ac80-9cf32bd12f39","Type":"ContainerStarted","Data":"c645093c6b13b49e00b7759c283bd4d3841ff073ac4fc379a3239ad01b579249"} Nov 21 14:30:32 crc kubenswrapper[4897]: I1121 14:30:32.214423 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"5f9bf5d1-dcbd-4d19-b46d-22810ebf17a2","Type":"ContainerStarted","Data":"be999347355dd6a758b2b12674e5d1936c39003c80cf99877cef75807d110c0e"} Nov 21 14:30:32 crc kubenswrapper[4897]: I1121 14:30:32.566066 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-mm6tz-config-t2x4t" Nov 21 14:30:32 crc kubenswrapper[4897]: I1121 14:30:32.691727 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/bfe51641-f509-4880-b8bd-9ce5da1e2291-var-run\") pod \"bfe51641-f509-4880-b8bd-9ce5da1e2291\" (UID: \"bfe51641-f509-4880-b8bd-9ce5da1e2291\") " Nov 21 14:30:32 crc kubenswrapper[4897]: I1121 14:30:32.691823 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/bfe51641-f509-4880-b8bd-9ce5da1e2291-additional-scripts\") pod \"bfe51641-f509-4880-b8bd-9ce5da1e2291\" (UID: \"bfe51641-f509-4880-b8bd-9ce5da1e2291\") " Nov 21 14:30:32 crc kubenswrapper[4897]: I1121 14:30:32.691862 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bfe51641-f509-4880-b8bd-9ce5da1e2291-var-run" (OuterVolumeSpecName: "var-run") pod "bfe51641-f509-4880-b8bd-9ce5da1e2291" (UID: "bfe51641-f509-4880-b8bd-9ce5da1e2291"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 21 14:30:32 crc kubenswrapper[4897]: I1121 14:30:32.691884 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bfe51641-f509-4880-b8bd-9ce5da1e2291-scripts\") pod \"bfe51641-f509-4880-b8bd-9ce5da1e2291\" (UID: \"bfe51641-f509-4880-b8bd-9ce5da1e2291\") " Nov 21 14:30:32 crc kubenswrapper[4897]: I1121 14:30:32.692001 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/bfe51641-f509-4880-b8bd-9ce5da1e2291-var-run-ovn\") pod \"bfe51641-f509-4880-b8bd-9ce5da1e2291\" (UID: \"bfe51641-f509-4880-b8bd-9ce5da1e2291\") " Nov 21 14:30:32 crc kubenswrapper[4897]: I1121 14:30:32.692083 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/bfe51641-f509-4880-b8bd-9ce5da1e2291-var-log-ovn\") pod \"bfe51641-f509-4880-b8bd-9ce5da1e2291\" (UID: \"bfe51641-f509-4880-b8bd-9ce5da1e2291\") " Nov 21 14:30:32 crc kubenswrapper[4897]: I1121 14:30:32.692151 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g8tx\" (UniqueName: \"kubernetes.io/projected/bfe51641-f509-4880-b8bd-9ce5da1e2291-kube-api-access-6g8tx\") pod \"bfe51641-f509-4880-b8bd-9ce5da1e2291\" (UID: \"bfe51641-f509-4880-b8bd-9ce5da1e2291\") " Nov 21 14:30:32 crc kubenswrapper[4897]: I1121 14:30:32.693174 4897 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/bfe51641-f509-4880-b8bd-9ce5da1e2291-var-run\") on node \"crc\" DevicePath \"\"" Nov 21 14:30:32 crc kubenswrapper[4897]: I1121 14:30:32.693644 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bfe51641-f509-4880-b8bd-9ce5da1e2291-scripts" (OuterVolumeSpecName: "scripts") pod "bfe51641-f509-4880-b8bd-9ce5da1e2291" (UID: "bfe51641-f509-4880-b8bd-9ce5da1e2291"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:30:32 crc kubenswrapper[4897]: I1121 14:30:32.693889 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bfe51641-f509-4880-b8bd-9ce5da1e2291-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "bfe51641-f509-4880-b8bd-9ce5da1e2291" (UID: "bfe51641-f509-4880-b8bd-9ce5da1e2291"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 21 14:30:32 crc kubenswrapper[4897]: I1121 14:30:32.693909 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bfe51641-f509-4880-b8bd-9ce5da1e2291-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "bfe51641-f509-4880-b8bd-9ce5da1e2291" (UID: "bfe51641-f509-4880-b8bd-9ce5da1e2291"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 21 14:30:32 crc kubenswrapper[4897]: I1121 14:30:32.694120 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bfe51641-f509-4880-b8bd-9ce5da1e2291-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "bfe51641-f509-4880-b8bd-9ce5da1e2291" (UID: "bfe51641-f509-4880-b8bd-9ce5da1e2291"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:30:32 crc kubenswrapper[4897]: I1121 14:30:32.701664 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bfe51641-f509-4880-b8bd-9ce5da1e2291-kube-api-access-6g8tx" (OuterVolumeSpecName: "kube-api-access-6g8tx") pod "bfe51641-f509-4880-b8bd-9ce5da1e2291" (UID: "bfe51641-f509-4880-b8bd-9ce5da1e2291"). InnerVolumeSpecName "kube-api-access-6g8tx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:30:32 crc kubenswrapper[4897]: I1121 14:30:32.795236 4897 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/bfe51641-f509-4880-b8bd-9ce5da1e2291-additional-scripts\") on node \"crc\" DevicePath \"\"" Nov 21 14:30:32 crc kubenswrapper[4897]: I1121 14:30:32.795273 4897 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bfe51641-f509-4880-b8bd-9ce5da1e2291-scripts\") on node \"crc\" DevicePath \"\"" Nov 21 14:30:32 crc kubenswrapper[4897]: I1121 14:30:32.795283 4897 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/bfe51641-f509-4880-b8bd-9ce5da1e2291-var-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 21 14:30:32 crc kubenswrapper[4897]: I1121 14:30:32.795290 4897 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/bfe51641-f509-4880-b8bd-9ce5da1e2291-var-log-ovn\") on node \"crc\" DevicePath \"\"" Nov 21 14:30:32 crc kubenswrapper[4897]: I1121 14:30:32.795299 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g8tx\" (UniqueName: \"kubernetes.io/projected/bfe51641-f509-4880-b8bd-9ce5da1e2291-kube-api-access-6g8tx\") on node \"crc\" DevicePath \"\"" Nov 21 14:30:33 crc kubenswrapper[4897]: I1121 14:30:33.228638 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-mm6tz-config-t2x4t" event={"ID":"bfe51641-f509-4880-b8bd-9ce5da1e2291","Type":"ContainerDied","Data":"43f7472e727eb88c81e6a6626e4dbc3452dd03cca036a3b42bfe4060716ebcba"} Nov 21 14:30:33 crc kubenswrapper[4897]: I1121 14:30:33.228684 4897 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="43f7472e727eb88c81e6a6626e4dbc3452dd03cca036a3b42bfe4060716ebcba" Nov 21 14:30:33 crc kubenswrapper[4897]: I1121 14:30:33.228651 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-mm6tz-config-t2x4t" Nov 21 14:30:33 crc kubenswrapper[4897]: I1121 14:30:33.233397 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"5f9bf5d1-dcbd-4d19-b46d-22810ebf17a2","Type":"ContainerStarted","Data":"5a0dc1b9bb584c50be3988a62cec6d31b430924a2cf04cba3bee6ae436bb99f7"} Nov 21 14:30:33 crc kubenswrapper[4897]: I1121 14:30:33.233457 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"5f9bf5d1-dcbd-4d19-b46d-22810ebf17a2","Type":"ContainerStarted","Data":"f340949cd506f5faf424374d9a22bc16ddb1734c399275de3c42fdf548e4880e"} Nov 21 14:30:33 crc kubenswrapper[4897]: I1121 14:30:33.233479 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"5f9bf5d1-dcbd-4d19-b46d-22810ebf17a2","Type":"ContainerStarted","Data":"fdfe95899e606188cefea08a35b5544f2a65979c693d4fe49de1c19290be8f5e"} Nov 21 14:30:33 crc kubenswrapper[4897]: I1121 14:30:33.476096 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-mm6tz" Nov 21 14:30:33 crc kubenswrapper[4897]: I1121 14:30:33.680777 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-mm6tz-config-t2x4t"] Nov 21 14:30:33 crc kubenswrapper[4897]: I1121 14:30:33.687467 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-b245-account-create-pprkv" Nov 21 14:30:33 crc kubenswrapper[4897]: I1121 14:30:33.688024 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-mm6tz-config-t2x4t"] Nov 21 14:30:33 crc kubenswrapper[4897]: I1121 14:30:33.694372 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-cell1-db-create-jb9vc" Nov 21 14:30:33 crc kubenswrapper[4897]: I1121 14:30:33.815741 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q2kxm\" (UniqueName: \"kubernetes.io/projected/0722112c-c860-4cb0-ac80-9cf32bd12f39-kube-api-access-q2kxm\") pod \"0722112c-c860-4cb0-ac80-9cf32bd12f39\" (UID: \"0722112c-c860-4cb0-ac80-9cf32bd12f39\") " Nov 21 14:30:33 crc kubenswrapper[4897]: I1121 14:30:33.815849 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sh4v4\" (UniqueName: \"kubernetes.io/projected/d56425f9-2cc4-4936-bda7-64520e84f5ab-kube-api-access-sh4v4\") pod \"d56425f9-2cc4-4936-bda7-64520e84f5ab\" (UID: \"d56425f9-2cc4-4936-bda7-64520e84f5ab\") " Nov 21 14:30:33 crc kubenswrapper[4897]: I1121 14:30:33.815972 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d56425f9-2cc4-4936-bda7-64520e84f5ab-operator-scripts\") pod \"d56425f9-2cc4-4936-bda7-64520e84f5ab\" (UID: \"d56425f9-2cc4-4936-bda7-64520e84f5ab\") " Nov 21 14:30:33 crc kubenswrapper[4897]: I1121 14:30:33.816090 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0722112c-c860-4cb0-ac80-9cf32bd12f39-operator-scripts\") pod \"0722112c-c860-4cb0-ac80-9cf32bd12f39\" (UID: \"0722112c-c860-4cb0-ac80-9cf32bd12f39\") " Nov 21 14:30:33 crc kubenswrapper[4897]: I1121 14:30:33.816823 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d56425f9-2cc4-4936-bda7-64520e84f5ab-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d56425f9-2cc4-4936-bda7-64520e84f5ab" (UID: "d56425f9-2cc4-4936-bda7-64520e84f5ab"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:30:33 crc kubenswrapper[4897]: I1121 14:30:33.816902 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0722112c-c860-4cb0-ac80-9cf32bd12f39-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0722112c-c860-4cb0-ac80-9cf32bd12f39" (UID: "0722112c-c860-4cb0-ac80-9cf32bd12f39"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:30:33 crc kubenswrapper[4897]: I1121 14:30:33.821492 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d56425f9-2cc4-4936-bda7-64520e84f5ab-kube-api-access-sh4v4" (OuterVolumeSpecName: "kube-api-access-sh4v4") pod "d56425f9-2cc4-4936-bda7-64520e84f5ab" (UID: "d56425f9-2cc4-4936-bda7-64520e84f5ab"). InnerVolumeSpecName "kube-api-access-sh4v4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:30:33 crc kubenswrapper[4897]: I1121 14:30:33.821998 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0722112c-c860-4cb0-ac80-9cf32bd12f39-kube-api-access-q2kxm" (OuterVolumeSpecName: "kube-api-access-q2kxm") pod "0722112c-c860-4cb0-ac80-9cf32bd12f39" (UID: "0722112c-c860-4cb0-ac80-9cf32bd12f39"). InnerVolumeSpecName "kube-api-access-q2kxm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:30:33 crc kubenswrapper[4897]: I1121 14:30:33.918691 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q2kxm\" (UniqueName: \"kubernetes.io/projected/0722112c-c860-4cb0-ac80-9cf32bd12f39-kube-api-access-q2kxm\") on node \"crc\" DevicePath \"\"" Nov 21 14:30:33 crc kubenswrapper[4897]: I1121 14:30:33.918722 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sh4v4\" (UniqueName: \"kubernetes.io/projected/d56425f9-2cc4-4936-bda7-64520e84f5ab-kube-api-access-sh4v4\") on node \"crc\" DevicePath \"\"" Nov 21 14:30:33 crc kubenswrapper[4897]: I1121 14:30:33.918732 4897 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d56425f9-2cc4-4936-bda7-64520e84f5ab-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 21 14:30:33 crc kubenswrapper[4897]: I1121 14:30:33.918742 4897 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0722112c-c860-4cb0-ac80-9cf32bd12f39-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 21 14:30:34 crc kubenswrapper[4897]: I1121 14:30:34.123154 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bfe51641-f509-4880-b8bd-9ce5da1e2291" path="/var/lib/kubelet/pods/bfe51641-f509-4880-b8bd-9ce5da1e2291/volumes" Nov 21 14:30:34 crc kubenswrapper[4897]: I1121 14:30:34.252238 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-b245-account-create-pprkv" Nov 21 14:30:34 crc kubenswrapper[4897]: I1121 14:30:34.253161 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-b245-account-create-pprkv" event={"ID":"d56425f9-2cc4-4936-bda7-64520e84f5ab","Type":"ContainerDied","Data":"1eb322ca974935a011b50872c735bc68c12fa072252c6669f1dcb40743753a00"} Nov 21 14:30:34 crc kubenswrapper[4897]: I1121 14:30:34.253208 4897 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1eb322ca974935a011b50872c735bc68c12fa072252c6669f1dcb40743753a00" Nov 21 14:30:34 crc kubenswrapper[4897]: I1121 14:30:34.259789 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-cell1-db-create-jb9vc" event={"ID":"0722112c-c860-4cb0-ac80-9cf32bd12f39","Type":"ContainerDied","Data":"c645093c6b13b49e00b7759c283bd4d3841ff073ac4fc379a3239ad01b579249"} Nov 21 14:30:34 crc kubenswrapper[4897]: I1121 14:30:34.259838 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-cell1-db-create-jb9vc" Nov 21 14:30:34 crc kubenswrapper[4897]: I1121 14:30:34.259843 4897 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c645093c6b13b49e00b7759c283bd4d3841ff073ac4fc379a3239ad01b579249" Nov 21 14:30:34 crc kubenswrapper[4897]: I1121 14:30:34.370867 4897 patch_prober.go:28] interesting pod/machine-config-daemon-krv5b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 21 14:30:34 crc kubenswrapper[4897]: I1121 14:30:34.370933 4897 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 21 14:30:35 crc kubenswrapper[4897]: I1121 14:30:35.275317 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"5f9bf5d1-dcbd-4d19-b46d-22810ebf17a2","Type":"ContainerStarted","Data":"74895ca52336c8f4ad92a5209c86b99509e2781fae75dc0dfe6540f068432565"} Nov 21 14:30:35 crc kubenswrapper[4897]: I1121 14:30:35.275677 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"5f9bf5d1-dcbd-4d19-b46d-22810ebf17a2","Type":"ContainerStarted","Data":"63030e2158d375ca7166087fe34d0d80f272277b636dfda3aea43603c732847d"} Nov 21 14:30:35 crc kubenswrapper[4897]: I1121 14:30:35.804096 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mysqld-exporter-0"] Nov 21 14:30:35 crc kubenswrapper[4897]: E1121 14:30:35.804756 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfe51641-f509-4880-b8bd-9ce5da1e2291" containerName="ovn-config" Nov 21 14:30:35 crc kubenswrapper[4897]: I1121 14:30:35.804774 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfe51641-f509-4880-b8bd-9ce5da1e2291" containerName="ovn-config" Nov 21 14:30:35 crc kubenswrapper[4897]: E1121 14:30:35.804790 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d56425f9-2cc4-4936-bda7-64520e84f5ab" containerName="mariadb-account-create" Nov 21 14:30:35 crc kubenswrapper[4897]: I1121 14:30:35.804796 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="d56425f9-2cc4-4936-bda7-64520e84f5ab" containerName="mariadb-account-create" Nov 21 14:30:35 crc kubenswrapper[4897]: E1121 14:30:35.804816 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0722112c-c860-4cb0-ac80-9cf32bd12f39" containerName="mariadb-database-create" Nov 21 14:30:35 crc kubenswrapper[4897]: I1121 14:30:35.804823 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="0722112c-c860-4cb0-ac80-9cf32bd12f39" containerName="mariadb-database-create" Nov 21 14:30:35 crc kubenswrapper[4897]: I1121 14:30:35.805023 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="0722112c-c860-4cb0-ac80-9cf32bd12f39" containerName="mariadb-database-create" Nov 21 14:30:35 crc kubenswrapper[4897]: I1121 14:30:35.805042 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="d56425f9-2cc4-4936-bda7-64520e84f5ab" containerName="mariadb-account-create" Nov 21 14:30:35 crc kubenswrapper[4897]: I1121 14:30:35.805053 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="bfe51641-f509-4880-b8bd-9ce5da1e2291" containerName="ovn-config" Nov 21 14:30:35 crc kubenswrapper[4897]: I1121 14:30:35.805771 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0" Nov 21 14:30:35 crc kubenswrapper[4897]: I1121 14:30:35.807721 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"mysqld-exporter-config-data" Nov 21 14:30:35 crc kubenswrapper[4897]: I1121 14:30:35.814070 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-0"] Nov 21 14:30:35 crc kubenswrapper[4897]: I1121 14:30:35.863725 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/813eae15-fec1-4d3a-bae2-1dc597c528f7-combined-ca-bundle\") pod \"mysqld-exporter-0\" (UID: \"813eae15-fec1-4d3a-bae2-1dc597c528f7\") " pod="openstack/mysqld-exporter-0" Nov 21 14:30:35 crc kubenswrapper[4897]: I1121 14:30:35.863824 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dvt2r\" (UniqueName: \"kubernetes.io/projected/813eae15-fec1-4d3a-bae2-1dc597c528f7-kube-api-access-dvt2r\") pod \"mysqld-exporter-0\" (UID: \"813eae15-fec1-4d3a-bae2-1dc597c528f7\") " pod="openstack/mysqld-exporter-0" Nov 21 14:30:35 crc kubenswrapper[4897]: I1121 14:30:35.863852 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/813eae15-fec1-4d3a-bae2-1dc597c528f7-config-data\") pod \"mysqld-exporter-0\" (UID: \"813eae15-fec1-4d3a-bae2-1dc597c528f7\") " pod="openstack/mysqld-exporter-0" Nov 21 14:30:35 crc kubenswrapper[4897]: I1121 14:30:35.965021 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dvt2r\" (UniqueName: \"kubernetes.io/projected/813eae15-fec1-4d3a-bae2-1dc597c528f7-kube-api-access-dvt2r\") pod \"mysqld-exporter-0\" (UID: \"813eae15-fec1-4d3a-bae2-1dc597c528f7\") " pod="openstack/mysqld-exporter-0" Nov 21 14:30:35 crc kubenswrapper[4897]: I1121 14:30:35.965088 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/813eae15-fec1-4d3a-bae2-1dc597c528f7-config-data\") pod \"mysqld-exporter-0\" (UID: \"813eae15-fec1-4d3a-bae2-1dc597c528f7\") " pod="openstack/mysqld-exporter-0" Nov 21 14:30:35 crc kubenswrapper[4897]: I1121 14:30:35.965230 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/813eae15-fec1-4d3a-bae2-1dc597c528f7-combined-ca-bundle\") pod \"mysqld-exporter-0\" (UID: \"813eae15-fec1-4d3a-bae2-1dc597c528f7\") " pod="openstack/mysqld-exporter-0" Nov 21 14:30:35 crc kubenswrapper[4897]: I1121 14:30:35.971972 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/813eae15-fec1-4d3a-bae2-1dc597c528f7-combined-ca-bundle\") pod \"mysqld-exporter-0\" (UID: \"813eae15-fec1-4d3a-bae2-1dc597c528f7\") " pod="openstack/mysqld-exporter-0" Nov 21 14:30:35 crc kubenswrapper[4897]: I1121 14:30:35.972321 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/813eae15-fec1-4d3a-bae2-1dc597c528f7-config-data\") pod \"mysqld-exporter-0\" (UID: \"813eae15-fec1-4d3a-bae2-1dc597c528f7\") " pod="openstack/mysqld-exporter-0" Nov 21 14:30:35 crc kubenswrapper[4897]: I1121 14:30:35.986378 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dvt2r\" (UniqueName: \"kubernetes.io/projected/813eae15-fec1-4d3a-bae2-1dc597c528f7-kube-api-access-dvt2r\") pod \"mysqld-exporter-0\" (UID: \"813eae15-fec1-4d3a-bae2-1dc597c528f7\") " pod="openstack/mysqld-exporter-0" Nov 21 14:30:36 crc kubenswrapper[4897]: I1121 14:30:36.124336 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0" Nov 21 14:30:36 crc kubenswrapper[4897]: I1121 14:30:36.302528 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"5f9bf5d1-dcbd-4d19-b46d-22810ebf17a2","Type":"ContainerStarted","Data":"c641b48ed422083d00a7ede2b5bafd93872cd0af34eea24adf5157d3af545409"} Nov 21 14:30:36 crc kubenswrapper[4897]: I1121 14:30:36.302810 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"5f9bf5d1-dcbd-4d19-b46d-22810ebf17a2","Type":"ContainerStarted","Data":"227565c6f01bd7e3dc79051ba02148f4621c55778865772d24745bec070c919d"} Nov 21 14:30:36 crc kubenswrapper[4897]: I1121 14:30:36.302828 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"5f9bf5d1-dcbd-4d19-b46d-22810ebf17a2","Type":"ContainerStarted","Data":"2100a9143d6485ea98ad70c2d5abf9f47561a47ff8279bcd27c5df75c17b51e3"} Nov 21 14:30:36 crc kubenswrapper[4897]: I1121 14:30:36.302840 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"5f9bf5d1-dcbd-4d19-b46d-22810ebf17a2","Type":"ContainerStarted","Data":"685b797caaf18c91bbdad93efa7d63ed323b1eabb76d214692ad1f9873fb7e6c"} Nov 21 14:30:36 crc kubenswrapper[4897]: I1121 14:30:36.596047 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-0"] Nov 21 14:30:36 crc kubenswrapper[4897]: W1121 14:30:36.604655 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod813eae15_fec1_4d3a_bae2_1dc597c528f7.slice/crio-90f8d530e55f456883c5f70d3642d86ec952d99a11869feb42d70bf1ee5dcfe9 WatchSource:0}: Error finding container 90f8d530e55f456883c5f70d3642d86ec952d99a11869feb42d70bf1ee5dcfe9: Status 404 returned error can't find the container with id 90f8d530e55f456883c5f70d3642d86ec952d99a11869feb42d70bf1ee5dcfe9 Nov 21 14:30:37 crc kubenswrapper[4897]: I1121 14:30:37.322931 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"5f9bf5d1-dcbd-4d19-b46d-22810ebf17a2","Type":"ContainerStarted","Data":"c94f15b9efee056cd87f2f2baa9bfa7e895af8b3e4165cbcfbb2086982c55dec"} Nov 21 14:30:37 crc kubenswrapper[4897]: I1121 14:30:37.326459 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0" event={"ID":"813eae15-fec1-4d3a-bae2-1dc597c528f7","Type":"ContainerStarted","Data":"90f8d530e55f456883c5f70d3642d86ec952d99a11869feb42d70bf1ee5dcfe9"} Nov 21 14:30:37 crc kubenswrapper[4897]: I1121 14:30:37.368496 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=36.933397052 podStartE2EDuration="47.368475025s" podCreationTimestamp="2025-11-21 14:29:50 +0000 UTC" firstStartedPulling="2025-11-21 14:30:23.952054122 +0000 UTC m=+1301.236647597" lastFinishedPulling="2025-11-21 14:30:34.387132095 +0000 UTC m=+1311.671725570" observedRunningTime="2025-11-21 14:30:37.354954022 +0000 UTC m=+1314.639547517" watchObservedRunningTime="2025-11-21 14:30:37.368475025 +0000 UTC m=+1314.653068520" Nov 21 14:30:37 crc kubenswrapper[4897]: I1121 14:30:37.720497 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6d5b6d6b67-qc7fs"] Nov 21 14:30:37 crc kubenswrapper[4897]: I1121 14:30:37.725439 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d5b6d6b67-qc7fs" Nov 21 14:30:37 crc kubenswrapper[4897]: I1121 14:30:37.727164 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Nov 21 14:30:37 crc kubenswrapper[4897]: I1121 14:30:37.732904 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d5b6d6b67-qc7fs"] Nov 21 14:30:37 crc kubenswrapper[4897]: I1121 14:30:37.804639 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/17a6c4a6-4493-4531-8f5c-814c7333a6ae-dns-svc\") pod \"dnsmasq-dns-6d5b6d6b67-qc7fs\" (UID: \"17a6c4a6-4493-4531-8f5c-814c7333a6ae\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-qc7fs" Nov 21 14:30:37 crc kubenswrapper[4897]: I1121 14:30:37.804711 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/17a6c4a6-4493-4531-8f5c-814c7333a6ae-dns-swift-storage-0\") pod \"dnsmasq-dns-6d5b6d6b67-qc7fs\" (UID: \"17a6c4a6-4493-4531-8f5c-814c7333a6ae\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-qc7fs" Nov 21 14:30:37 crc kubenswrapper[4897]: I1121 14:30:37.804764 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/17a6c4a6-4493-4531-8f5c-814c7333a6ae-ovsdbserver-nb\") pod \"dnsmasq-dns-6d5b6d6b67-qc7fs\" (UID: \"17a6c4a6-4493-4531-8f5c-814c7333a6ae\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-qc7fs" Nov 21 14:30:37 crc kubenswrapper[4897]: I1121 14:30:37.804967 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17a6c4a6-4493-4531-8f5c-814c7333a6ae-config\") pod \"dnsmasq-dns-6d5b6d6b67-qc7fs\" (UID: \"17a6c4a6-4493-4531-8f5c-814c7333a6ae\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-qc7fs" Nov 21 14:30:37 crc kubenswrapper[4897]: I1121 14:30:37.805223 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/17a6c4a6-4493-4531-8f5c-814c7333a6ae-ovsdbserver-sb\") pod \"dnsmasq-dns-6d5b6d6b67-qc7fs\" (UID: \"17a6c4a6-4493-4531-8f5c-814c7333a6ae\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-qc7fs" Nov 21 14:30:37 crc kubenswrapper[4897]: I1121 14:30:37.805632 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2blq5\" (UniqueName: \"kubernetes.io/projected/17a6c4a6-4493-4531-8f5c-814c7333a6ae-kube-api-access-2blq5\") pod \"dnsmasq-dns-6d5b6d6b67-qc7fs\" (UID: \"17a6c4a6-4493-4531-8f5c-814c7333a6ae\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-qc7fs" Nov 21 14:30:37 crc kubenswrapper[4897]: I1121 14:30:37.907948 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2blq5\" (UniqueName: \"kubernetes.io/projected/17a6c4a6-4493-4531-8f5c-814c7333a6ae-kube-api-access-2blq5\") pod \"dnsmasq-dns-6d5b6d6b67-qc7fs\" (UID: \"17a6c4a6-4493-4531-8f5c-814c7333a6ae\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-qc7fs" Nov 21 14:30:37 crc kubenswrapper[4897]: I1121 14:30:37.908014 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/17a6c4a6-4493-4531-8f5c-814c7333a6ae-dns-svc\") pod \"dnsmasq-dns-6d5b6d6b67-qc7fs\" (UID: \"17a6c4a6-4493-4531-8f5c-814c7333a6ae\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-qc7fs" Nov 21 14:30:37 crc kubenswrapper[4897]: I1121 14:30:37.908039 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/17a6c4a6-4493-4531-8f5c-814c7333a6ae-dns-swift-storage-0\") pod \"dnsmasq-dns-6d5b6d6b67-qc7fs\" (UID: \"17a6c4a6-4493-4531-8f5c-814c7333a6ae\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-qc7fs" Nov 21 14:30:37 crc kubenswrapper[4897]: I1121 14:30:37.908073 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/17a6c4a6-4493-4531-8f5c-814c7333a6ae-ovsdbserver-nb\") pod \"dnsmasq-dns-6d5b6d6b67-qc7fs\" (UID: \"17a6c4a6-4493-4531-8f5c-814c7333a6ae\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-qc7fs" Nov 21 14:30:37 crc kubenswrapper[4897]: I1121 14:30:37.908150 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17a6c4a6-4493-4531-8f5c-814c7333a6ae-config\") pod \"dnsmasq-dns-6d5b6d6b67-qc7fs\" (UID: \"17a6c4a6-4493-4531-8f5c-814c7333a6ae\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-qc7fs" Nov 21 14:30:37 crc kubenswrapper[4897]: I1121 14:30:37.908183 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/17a6c4a6-4493-4531-8f5c-814c7333a6ae-ovsdbserver-sb\") pod \"dnsmasq-dns-6d5b6d6b67-qc7fs\" (UID: \"17a6c4a6-4493-4531-8f5c-814c7333a6ae\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-qc7fs" Nov 21 14:30:37 crc kubenswrapper[4897]: I1121 14:30:37.909020 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/17a6c4a6-4493-4531-8f5c-814c7333a6ae-ovsdbserver-sb\") pod \"dnsmasq-dns-6d5b6d6b67-qc7fs\" (UID: \"17a6c4a6-4493-4531-8f5c-814c7333a6ae\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-qc7fs" Nov 21 14:30:37 crc kubenswrapper[4897]: I1121 14:30:37.909742 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17a6c4a6-4493-4531-8f5c-814c7333a6ae-config\") pod \"dnsmasq-dns-6d5b6d6b67-qc7fs\" (UID: \"17a6c4a6-4493-4531-8f5c-814c7333a6ae\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-qc7fs" Nov 21 14:30:37 crc kubenswrapper[4897]: I1121 14:30:37.909961 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/17a6c4a6-4493-4531-8f5c-814c7333a6ae-dns-swift-storage-0\") pod \"dnsmasq-dns-6d5b6d6b67-qc7fs\" (UID: \"17a6c4a6-4493-4531-8f5c-814c7333a6ae\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-qc7fs" Nov 21 14:30:37 crc kubenswrapper[4897]: I1121 14:30:37.910875 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/17a6c4a6-4493-4531-8f5c-814c7333a6ae-ovsdbserver-nb\") pod \"dnsmasq-dns-6d5b6d6b67-qc7fs\" (UID: \"17a6c4a6-4493-4531-8f5c-814c7333a6ae\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-qc7fs" Nov 21 14:30:37 crc kubenswrapper[4897]: I1121 14:30:37.911698 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/17a6c4a6-4493-4531-8f5c-814c7333a6ae-dns-svc\") pod \"dnsmasq-dns-6d5b6d6b67-qc7fs\" (UID: \"17a6c4a6-4493-4531-8f5c-814c7333a6ae\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-qc7fs" Nov 21 14:30:37 crc kubenswrapper[4897]: I1121 14:30:37.930557 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2blq5\" (UniqueName: \"kubernetes.io/projected/17a6c4a6-4493-4531-8f5c-814c7333a6ae-kube-api-access-2blq5\") pod \"dnsmasq-dns-6d5b6d6b67-qc7fs\" (UID: \"17a6c4a6-4493-4531-8f5c-814c7333a6ae\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-qc7fs" Nov 21 14:30:38 crc kubenswrapper[4897]: I1121 14:30:38.057824 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d5b6d6b67-qc7fs" Nov 21 14:30:41 crc kubenswrapper[4897]: I1121 14:30:41.500305 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Nov 21 14:30:41 crc kubenswrapper[4897]: I1121 14:30:41.504815 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Nov 21 14:30:42 crc kubenswrapper[4897]: I1121 14:30:42.381019 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Nov 21 14:30:44 crc kubenswrapper[4897]: I1121 14:30:44.177556 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 21 14:30:44 crc kubenswrapper[4897]: I1121 14:30:44.399948 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="91aa2270-92a7-4af4-9a92-141ea2b24ee1" containerName="prometheus" containerID="cri-o://d4cba41371733c06d0f9282cb6e29395e6f232e62a3ce987fe67b0eb3d37cb98" gracePeriod=600 Nov 21 14:30:44 crc kubenswrapper[4897]: I1121 14:30:44.400012 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="91aa2270-92a7-4af4-9a92-141ea2b24ee1" containerName="thanos-sidecar" containerID="cri-o://d75dc39a201e4f83e8ba5e273e81d6969573634dc7a6223e72bea5261f0bd8db" gracePeriod=600 Nov 21 14:30:44 crc kubenswrapper[4897]: I1121 14:30:44.400065 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="91aa2270-92a7-4af4-9a92-141ea2b24ee1" containerName="config-reloader" containerID="cri-o://1ae73b0cd6fdca18483acc4c5180aa46a081811e918b88fc1fb45977279fc1f2" gracePeriod=600 Nov 21 14:30:44 crc kubenswrapper[4897]: I1121 14:30:44.956764 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Nov 21 14:30:45 crc kubenswrapper[4897]: I1121 14:30:45.376527 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-dpnlm"] Nov 21 14:30:45 crc kubenswrapper[4897]: I1121 14:30:45.377906 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-dpnlm" Nov 21 14:30:45 crc kubenswrapper[4897]: I1121 14:30:45.387536 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-dpnlm"] Nov 21 14:30:45 crc kubenswrapper[4897]: I1121 14:30:45.393711 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Nov 21 14:30:45 crc kubenswrapper[4897]: I1121 14:30:45.402429 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-49wwn\" (UniqueName: \"kubernetes.io/projected/67dd6a02-c8f6-4035-ad18-c6bbea13c688-kube-api-access-49wwn\") pod \"cinder-db-create-dpnlm\" (UID: \"67dd6a02-c8f6-4035-ad18-c6bbea13c688\") " pod="openstack/cinder-db-create-dpnlm" Nov 21 14:30:45 crc kubenswrapper[4897]: I1121 14:30:45.402631 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/67dd6a02-c8f6-4035-ad18-c6bbea13c688-operator-scripts\") pod \"cinder-db-create-dpnlm\" (UID: \"67dd6a02-c8f6-4035-ad18-c6bbea13c688\") " pod="openstack/cinder-db-create-dpnlm" Nov 21 14:30:45 crc kubenswrapper[4897]: I1121 14:30:45.472684 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-create-hg4dh"] Nov 21 14:30:45 crc kubenswrapper[4897]: I1121 14:30:45.474094 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-hg4dh" Nov 21 14:30:45 crc kubenswrapper[4897]: I1121 14:30:45.494841 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-hg4dh"] Nov 21 14:30:45 crc kubenswrapper[4897]: I1121 14:30:45.504497 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/67dd6a02-c8f6-4035-ad18-c6bbea13c688-operator-scripts\") pod \"cinder-db-create-dpnlm\" (UID: \"67dd6a02-c8f6-4035-ad18-c6bbea13c688\") " pod="openstack/cinder-db-create-dpnlm" Nov 21 14:30:45 crc kubenswrapper[4897]: I1121 14:30:45.504569 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ee8a4380-421f-45b1-ba19-9832b69935fe-operator-scripts\") pod \"heat-db-create-hg4dh\" (UID: \"ee8a4380-421f-45b1-ba19-9832b69935fe\") " pod="openstack/heat-db-create-hg4dh" Nov 21 14:30:45 crc kubenswrapper[4897]: I1121 14:30:45.504698 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-49wwn\" (UniqueName: \"kubernetes.io/projected/67dd6a02-c8f6-4035-ad18-c6bbea13c688-kube-api-access-49wwn\") pod \"cinder-db-create-dpnlm\" (UID: \"67dd6a02-c8f6-4035-ad18-c6bbea13c688\") " pod="openstack/cinder-db-create-dpnlm" Nov 21 14:30:45 crc kubenswrapper[4897]: I1121 14:30:45.504738 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pzsg2\" (UniqueName: \"kubernetes.io/projected/ee8a4380-421f-45b1-ba19-9832b69935fe-kube-api-access-pzsg2\") pod \"heat-db-create-hg4dh\" (UID: \"ee8a4380-421f-45b1-ba19-9832b69935fe\") " pod="openstack/heat-db-create-hg4dh" Nov 21 14:30:45 crc kubenswrapper[4897]: I1121 14:30:45.505457 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/67dd6a02-c8f6-4035-ad18-c6bbea13c688-operator-scripts\") pod \"cinder-db-create-dpnlm\" (UID: \"67dd6a02-c8f6-4035-ad18-c6bbea13c688\") " pod="openstack/cinder-db-create-dpnlm" Nov 21 14:30:45 crc kubenswrapper[4897]: I1121 14:30:45.527407 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-49wwn\" (UniqueName: \"kubernetes.io/projected/67dd6a02-c8f6-4035-ad18-c6bbea13c688-kube-api-access-49wwn\") pod \"cinder-db-create-dpnlm\" (UID: \"67dd6a02-c8f6-4035-ad18-c6bbea13c688\") " pod="openstack/cinder-db-create-dpnlm" Nov 21 14:30:45 crc kubenswrapper[4897]: I1121 14:30:45.583828 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-5f6b-account-create-29rlc"] Nov 21 14:30:45 crc kubenswrapper[4897]: I1121 14:30:45.585121 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-5f6b-account-create-29rlc" Nov 21 14:30:45 crc kubenswrapper[4897]: I1121 14:30:45.587113 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-db-secret" Nov 21 14:30:45 crc kubenswrapper[4897]: I1121 14:30:45.594832 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-5f6b-account-create-29rlc"] Nov 21 14:30:45 crc kubenswrapper[4897]: I1121 14:30:45.605425 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pzsg2\" (UniqueName: \"kubernetes.io/projected/ee8a4380-421f-45b1-ba19-9832b69935fe-kube-api-access-pzsg2\") pod \"heat-db-create-hg4dh\" (UID: \"ee8a4380-421f-45b1-ba19-9832b69935fe\") " pod="openstack/heat-db-create-hg4dh" Nov 21 14:30:45 crc kubenswrapper[4897]: I1121 14:30:45.605497 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ksmjv\" (UniqueName: \"kubernetes.io/projected/a69b6579-ce5a-44d7-ab3d-5454dd5f85df-kube-api-access-ksmjv\") pod \"heat-5f6b-account-create-29rlc\" (UID: \"a69b6579-ce5a-44d7-ab3d-5454dd5f85df\") " pod="openstack/heat-5f6b-account-create-29rlc" Nov 21 14:30:45 crc kubenswrapper[4897]: I1121 14:30:45.605656 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a69b6579-ce5a-44d7-ab3d-5454dd5f85df-operator-scripts\") pod \"heat-5f6b-account-create-29rlc\" (UID: \"a69b6579-ce5a-44d7-ab3d-5454dd5f85df\") " pod="openstack/heat-5f6b-account-create-29rlc" Nov 21 14:30:45 crc kubenswrapper[4897]: I1121 14:30:45.605935 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ee8a4380-421f-45b1-ba19-9832b69935fe-operator-scripts\") pod \"heat-db-create-hg4dh\" (UID: \"ee8a4380-421f-45b1-ba19-9832b69935fe\") " pod="openstack/heat-db-create-hg4dh" Nov 21 14:30:45 crc kubenswrapper[4897]: I1121 14:30:45.606647 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ee8a4380-421f-45b1-ba19-9832b69935fe-operator-scripts\") pod \"heat-db-create-hg4dh\" (UID: \"ee8a4380-421f-45b1-ba19-9832b69935fe\") " pod="openstack/heat-db-create-hg4dh" Nov 21 14:30:45 crc kubenswrapper[4897]: I1121 14:30:45.629482 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pzsg2\" (UniqueName: \"kubernetes.io/projected/ee8a4380-421f-45b1-ba19-9832b69935fe-kube-api-access-pzsg2\") pod \"heat-db-create-hg4dh\" (UID: \"ee8a4380-421f-45b1-ba19-9832b69935fe\") " pod="openstack/heat-db-create-hg4dh" Nov 21 14:30:45 crc kubenswrapper[4897]: I1121 14:30:45.678763 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-cgbsk"] Nov 21 14:30:45 crc kubenswrapper[4897]: I1121 14:30:45.679985 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-cgbsk" Nov 21 14:30:45 crc kubenswrapper[4897]: I1121 14:30:45.694927 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-dpnlm" Nov 21 14:30:45 crc kubenswrapper[4897]: I1121 14:30:45.703986 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-cgbsk"] Nov 21 14:30:45 crc kubenswrapper[4897]: I1121 14:30:45.712659 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a69b6579-ce5a-44d7-ab3d-5454dd5f85df-operator-scripts\") pod \"heat-5f6b-account-create-29rlc\" (UID: \"a69b6579-ce5a-44d7-ab3d-5454dd5f85df\") " pod="openstack/heat-5f6b-account-create-29rlc" Nov 21 14:30:45 crc kubenswrapper[4897]: I1121 14:30:45.712793 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8bb39c02-70b6-470f-bf62-add1c2ebc02f-operator-scripts\") pod \"barbican-db-create-cgbsk\" (UID: \"8bb39c02-70b6-470f-bf62-add1c2ebc02f\") " pod="openstack/barbican-db-create-cgbsk" Nov 21 14:30:45 crc kubenswrapper[4897]: I1121 14:30:45.712921 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ksmjv\" (UniqueName: \"kubernetes.io/projected/a69b6579-ce5a-44d7-ab3d-5454dd5f85df-kube-api-access-ksmjv\") pod \"heat-5f6b-account-create-29rlc\" (UID: \"a69b6579-ce5a-44d7-ab3d-5454dd5f85df\") " pod="openstack/heat-5f6b-account-create-29rlc" Nov 21 14:30:45 crc kubenswrapper[4897]: I1121 14:30:45.712951 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4x7ln\" (UniqueName: \"kubernetes.io/projected/8bb39c02-70b6-470f-bf62-add1c2ebc02f-kube-api-access-4x7ln\") pod \"barbican-db-create-cgbsk\" (UID: \"8bb39c02-70b6-470f-bf62-add1c2ebc02f\") " pod="openstack/barbican-db-create-cgbsk" Nov 21 14:30:45 crc kubenswrapper[4897]: I1121 14:30:45.713722 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a69b6579-ce5a-44d7-ab3d-5454dd5f85df-operator-scripts\") pod \"heat-5f6b-account-create-29rlc\" (UID: \"a69b6579-ce5a-44d7-ab3d-5454dd5f85df\") " pod="openstack/heat-5f6b-account-create-29rlc" Nov 21 14:30:45 crc kubenswrapper[4897]: I1121 14:30:45.719970 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-c4c2-account-create-kzvn8"] Nov 21 14:30:45 crc kubenswrapper[4897]: I1121 14:30:45.721353 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-c4c2-account-create-kzvn8" Nov 21 14:30:45 crc kubenswrapper[4897]: I1121 14:30:45.726644 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Nov 21 14:30:45 crc kubenswrapper[4897]: I1121 14:30:45.731978 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-227rw"] Nov 21 14:30:45 crc kubenswrapper[4897]: I1121 14:30:45.733580 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-227rw" Nov 21 14:30:45 crc kubenswrapper[4897]: I1121 14:30:45.739039 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 21 14:30:45 crc kubenswrapper[4897]: I1121 14:30:45.739240 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 21 14:30:45 crc kubenswrapper[4897]: I1121 14:30:45.739464 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-wbfsb" Nov 21 14:30:45 crc kubenswrapper[4897]: I1121 14:30:45.739649 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 21 14:30:45 crc kubenswrapper[4897]: I1121 14:30:45.740013 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ksmjv\" (UniqueName: \"kubernetes.io/projected/a69b6579-ce5a-44d7-ab3d-5454dd5f85df-kube-api-access-ksmjv\") pod \"heat-5f6b-account-create-29rlc\" (UID: \"a69b6579-ce5a-44d7-ab3d-5454dd5f85df\") " pod="openstack/heat-5f6b-account-create-29rlc" Nov 21 14:30:45 crc kubenswrapper[4897]: I1121 14:30:45.783354 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-227rw"] Nov 21 14:30:45 crc kubenswrapper[4897]: I1121 14:30:45.790036 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-hg4dh" Nov 21 14:30:45 crc kubenswrapper[4897]: I1121 14:30:45.792867 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-c4c2-account-create-kzvn8"] Nov 21 14:30:45 crc kubenswrapper[4897]: I1121 14:30:45.814625 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/208ac008-5901-4b46-983a-a73c9ed207f2-config-data\") pod \"keystone-db-sync-227rw\" (UID: \"208ac008-5901-4b46-983a-a73c9ed207f2\") " pod="openstack/keystone-db-sync-227rw" Nov 21 14:30:45 crc kubenswrapper[4897]: I1121 14:30:45.814707 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xzdbv\" (UniqueName: \"kubernetes.io/projected/56c583f0-0dba-4e94-a8b7-26d7b5c0e87f-kube-api-access-xzdbv\") pod \"barbican-c4c2-account-create-kzvn8\" (UID: \"56c583f0-0dba-4e94-a8b7-26d7b5c0e87f\") " pod="openstack/barbican-c4c2-account-create-kzvn8" Nov 21 14:30:45 crc kubenswrapper[4897]: I1121 14:30:45.814746 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/208ac008-5901-4b46-983a-a73c9ed207f2-combined-ca-bundle\") pod \"keystone-db-sync-227rw\" (UID: \"208ac008-5901-4b46-983a-a73c9ed207f2\") " pod="openstack/keystone-db-sync-227rw" Nov 21 14:30:45 crc kubenswrapper[4897]: I1121 14:30:45.814798 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4x7ln\" (UniqueName: \"kubernetes.io/projected/8bb39c02-70b6-470f-bf62-add1c2ebc02f-kube-api-access-4x7ln\") pod \"barbican-db-create-cgbsk\" (UID: \"8bb39c02-70b6-470f-bf62-add1c2ebc02f\") " pod="openstack/barbican-db-create-cgbsk" Nov 21 14:30:45 crc kubenswrapper[4897]: I1121 14:30:45.814835 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/56c583f0-0dba-4e94-a8b7-26d7b5c0e87f-operator-scripts\") pod \"barbican-c4c2-account-create-kzvn8\" (UID: \"56c583f0-0dba-4e94-a8b7-26d7b5c0e87f\") " pod="openstack/barbican-c4c2-account-create-kzvn8" Nov 21 14:30:45 crc kubenswrapper[4897]: I1121 14:30:45.814919 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vd667\" (UniqueName: \"kubernetes.io/projected/208ac008-5901-4b46-983a-a73c9ed207f2-kube-api-access-vd667\") pod \"keystone-db-sync-227rw\" (UID: \"208ac008-5901-4b46-983a-a73c9ed207f2\") " pod="openstack/keystone-db-sync-227rw" Nov 21 14:30:45 crc kubenswrapper[4897]: I1121 14:30:45.814950 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8bb39c02-70b6-470f-bf62-add1c2ebc02f-operator-scripts\") pod \"barbican-db-create-cgbsk\" (UID: \"8bb39c02-70b6-470f-bf62-add1c2ebc02f\") " pod="openstack/barbican-db-create-cgbsk" Nov 21 14:30:45 crc kubenswrapper[4897]: I1121 14:30:45.815648 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8bb39c02-70b6-470f-bf62-add1c2ebc02f-operator-scripts\") pod \"barbican-db-create-cgbsk\" (UID: \"8bb39c02-70b6-470f-bf62-add1c2ebc02f\") " pod="openstack/barbican-db-create-cgbsk" Nov 21 14:30:45 crc kubenswrapper[4897]: I1121 14:30:45.831058 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4x7ln\" (UniqueName: \"kubernetes.io/projected/8bb39c02-70b6-470f-bf62-add1c2ebc02f-kube-api-access-4x7ln\") pod \"barbican-db-create-cgbsk\" (UID: \"8bb39c02-70b6-470f-bf62-add1c2ebc02f\") " pod="openstack/barbican-db-create-cgbsk" Nov 21 14:30:45 crc kubenswrapper[4897]: I1121 14:30:45.888460 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-f2a9-account-create-rx44g"] Nov 21 14:30:45 crc kubenswrapper[4897]: I1121 14:30:45.889911 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-f2a9-account-create-rx44g" Nov 21 14:30:45 crc kubenswrapper[4897]: I1121 14:30:45.904317 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-f2a9-account-create-rx44g"] Nov 21 14:30:45 crc kubenswrapper[4897]: I1121 14:30:45.928067 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-5f6b-account-create-29rlc" Nov 21 14:30:45 crc kubenswrapper[4897]: I1121 14:30:45.928896 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sf4jq\" (UniqueName: \"kubernetes.io/projected/1fd99443-7153-4711-9575-6ebc4fbfcf42-kube-api-access-sf4jq\") pod \"cinder-f2a9-account-create-rx44g\" (UID: \"1fd99443-7153-4711-9575-6ebc4fbfcf42\") " pod="openstack/cinder-f2a9-account-create-rx44g" Nov 21 14:30:45 crc kubenswrapper[4897]: I1121 14:30:45.928953 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vd667\" (UniqueName: \"kubernetes.io/projected/208ac008-5901-4b46-983a-a73c9ed207f2-kube-api-access-vd667\") pod \"keystone-db-sync-227rw\" (UID: \"208ac008-5901-4b46-983a-a73c9ed207f2\") " pod="openstack/keystone-db-sync-227rw" Nov 21 14:30:45 crc kubenswrapper[4897]: I1121 14:30:45.928985 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1fd99443-7153-4711-9575-6ebc4fbfcf42-operator-scripts\") pod \"cinder-f2a9-account-create-rx44g\" (UID: \"1fd99443-7153-4711-9575-6ebc4fbfcf42\") " pod="openstack/cinder-f2a9-account-create-rx44g" Nov 21 14:30:45 crc kubenswrapper[4897]: I1121 14:30:45.929031 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/208ac008-5901-4b46-983a-a73c9ed207f2-config-data\") pod \"keystone-db-sync-227rw\" (UID: \"208ac008-5901-4b46-983a-a73c9ed207f2\") " pod="openstack/keystone-db-sync-227rw" Nov 21 14:30:45 crc kubenswrapper[4897]: I1121 14:30:45.929078 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xzdbv\" (UniqueName: \"kubernetes.io/projected/56c583f0-0dba-4e94-a8b7-26d7b5c0e87f-kube-api-access-xzdbv\") pod \"barbican-c4c2-account-create-kzvn8\" (UID: \"56c583f0-0dba-4e94-a8b7-26d7b5c0e87f\") " pod="openstack/barbican-c4c2-account-create-kzvn8" Nov 21 14:30:45 crc kubenswrapper[4897]: I1121 14:30:45.929111 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/208ac008-5901-4b46-983a-a73c9ed207f2-combined-ca-bundle\") pod \"keystone-db-sync-227rw\" (UID: \"208ac008-5901-4b46-983a-a73c9ed207f2\") " pod="openstack/keystone-db-sync-227rw" Nov 21 14:30:45 crc kubenswrapper[4897]: I1121 14:30:45.929175 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/56c583f0-0dba-4e94-a8b7-26d7b5c0e87f-operator-scripts\") pod \"barbican-c4c2-account-create-kzvn8\" (UID: \"56c583f0-0dba-4e94-a8b7-26d7b5c0e87f\") " pod="openstack/barbican-c4c2-account-create-kzvn8" Nov 21 14:30:45 crc kubenswrapper[4897]: I1121 14:30:45.930275 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/56c583f0-0dba-4e94-a8b7-26d7b5c0e87f-operator-scripts\") pod \"barbican-c4c2-account-create-kzvn8\" (UID: \"56c583f0-0dba-4e94-a8b7-26d7b5c0e87f\") " pod="openstack/barbican-c4c2-account-create-kzvn8" Nov 21 14:30:45 crc kubenswrapper[4897]: I1121 14:30:45.932173 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Nov 21 14:30:45 crc kubenswrapper[4897]: I1121 14:30:45.932631 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/208ac008-5901-4b46-983a-a73c9ed207f2-config-data\") pod \"keystone-db-sync-227rw\" (UID: \"208ac008-5901-4b46-983a-a73c9ed207f2\") " pod="openstack/keystone-db-sync-227rw" Nov 21 14:30:45 crc kubenswrapper[4897]: I1121 14:30:45.943185 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/208ac008-5901-4b46-983a-a73c9ed207f2-combined-ca-bundle\") pod \"keystone-db-sync-227rw\" (UID: \"208ac008-5901-4b46-983a-a73c9ed207f2\") " pod="openstack/keystone-db-sync-227rw" Nov 21 14:30:45 crc kubenswrapper[4897]: I1121 14:30:45.954182 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xzdbv\" (UniqueName: \"kubernetes.io/projected/56c583f0-0dba-4e94-a8b7-26d7b5c0e87f-kube-api-access-xzdbv\") pod \"barbican-c4c2-account-create-kzvn8\" (UID: \"56c583f0-0dba-4e94-a8b7-26d7b5c0e87f\") " pod="openstack/barbican-c4c2-account-create-kzvn8" Nov 21 14:30:45 crc kubenswrapper[4897]: I1121 14:30:45.963087 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vd667\" (UniqueName: \"kubernetes.io/projected/208ac008-5901-4b46-983a-a73c9ed207f2-kube-api-access-vd667\") pod \"keystone-db-sync-227rw\" (UID: \"208ac008-5901-4b46-983a-a73c9ed207f2\") " pod="openstack/keystone-db-sync-227rw" Nov 21 14:30:46 crc kubenswrapper[4897]: I1121 14:30:46.003305 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-cgbsk" Nov 21 14:30:46 crc kubenswrapper[4897]: I1121 14:30:46.031146 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sf4jq\" (UniqueName: \"kubernetes.io/projected/1fd99443-7153-4711-9575-6ebc4fbfcf42-kube-api-access-sf4jq\") pod \"cinder-f2a9-account-create-rx44g\" (UID: \"1fd99443-7153-4711-9575-6ebc4fbfcf42\") " pod="openstack/cinder-f2a9-account-create-rx44g" Nov 21 14:30:46 crc kubenswrapper[4897]: I1121 14:30:46.031219 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1fd99443-7153-4711-9575-6ebc4fbfcf42-operator-scripts\") pod \"cinder-f2a9-account-create-rx44g\" (UID: \"1fd99443-7153-4711-9575-6ebc4fbfcf42\") " pod="openstack/cinder-f2a9-account-create-rx44g" Nov 21 14:30:46 crc kubenswrapper[4897]: I1121 14:30:46.032100 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1fd99443-7153-4711-9575-6ebc4fbfcf42-operator-scripts\") pod \"cinder-f2a9-account-create-rx44g\" (UID: \"1fd99443-7153-4711-9575-6ebc4fbfcf42\") " pod="openstack/cinder-f2a9-account-create-rx44g" Nov 21 14:30:46 crc kubenswrapper[4897]: I1121 14:30:46.046837 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sf4jq\" (UniqueName: \"kubernetes.io/projected/1fd99443-7153-4711-9575-6ebc4fbfcf42-kube-api-access-sf4jq\") pod \"cinder-f2a9-account-create-rx44g\" (UID: \"1fd99443-7153-4711-9575-6ebc4fbfcf42\") " pod="openstack/cinder-f2a9-account-create-rx44g" Nov 21 14:30:46 crc kubenswrapper[4897]: I1121 14:30:46.075329 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-hrqhb"] Nov 21 14:30:46 crc kubenswrapper[4897]: I1121 14:30:46.076828 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-hrqhb" Nov 21 14:30:46 crc kubenswrapper[4897]: I1121 14:30:46.089994 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-c4c2-account-create-kzvn8" Nov 21 14:30:46 crc kubenswrapper[4897]: I1121 14:30:46.098727 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-227rw" Nov 21 14:30:46 crc kubenswrapper[4897]: I1121 14:30:46.109251 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-hrqhb"] Nov 21 14:30:46 crc kubenswrapper[4897]: I1121 14:30:46.133650 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/982ba7b6-1d19-4bd4-bdea-c828fd678186-operator-scripts\") pod \"neutron-db-create-hrqhb\" (UID: \"982ba7b6-1d19-4bd4-bdea-c828fd678186\") " pod="openstack/neutron-db-create-hrqhb" Nov 21 14:30:46 crc kubenswrapper[4897]: I1121 14:30:46.133963 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v8zbf\" (UniqueName: \"kubernetes.io/projected/982ba7b6-1d19-4bd4-bdea-c828fd678186-kube-api-access-v8zbf\") pod \"neutron-db-create-hrqhb\" (UID: \"982ba7b6-1d19-4bd4-bdea-c828fd678186\") " pod="openstack/neutron-db-create-hrqhb" Nov 21 14:30:46 crc kubenswrapper[4897]: I1121 14:30:46.180769 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-a766-account-create-btlq4"] Nov 21 14:30:46 crc kubenswrapper[4897]: I1121 14:30:46.182203 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-a766-account-create-btlq4" Nov 21 14:30:46 crc kubenswrapper[4897]: I1121 14:30:46.184363 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Nov 21 14:30:46 crc kubenswrapper[4897]: I1121 14:30:46.189733 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-a766-account-create-btlq4"] Nov 21 14:30:46 crc kubenswrapper[4897]: I1121 14:30:46.235817 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v8zbf\" (UniqueName: \"kubernetes.io/projected/982ba7b6-1d19-4bd4-bdea-c828fd678186-kube-api-access-v8zbf\") pod \"neutron-db-create-hrqhb\" (UID: \"982ba7b6-1d19-4bd4-bdea-c828fd678186\") " pod="openstack/neutron-db-create-hrqhb" Nov 21 14:30:46 crc kubenswrapper[4897]: I1121 14:30:46.236088 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/982ba7b6-1d19-4bd4-bdea-c828fd678186-operator-scripts\") pod \"neutron-db-create-hrqhb\" (UID: \"982ba7b6-1d19-4bd4-bdea-c828fd678186\") " pod="openstack/neutron-db-create-hrqhb" Nov 21 14:30:46 crc kubenswrapper[4897]: I1121 14:30:46.236818 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/982ba7b6-1d19-4bd4-bdea-c828fd678186-operator-scripts\") pod \"neutron-db-create-hrqhb\" (UID: \"982ba7b6-1d19-4bd4-bdea-c828fd678186\") " pod="openstack/neutron-db-create-hrqhb" Nov 21 14:30:46 crc kubenswrapper[4897]: I1121 14:30:46.252178 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v8zbf\" (UniqueName: \"kubernetes.io/projected/982ba7b6-1d19-4bd4-bdea-c828fd678186-kube-api-access-v8zbf\") pod \"neutron-db-create-hrqhb\" (UID: \"982ba7b6-1d19-4bd4-bdea-c828fd678186\") " pod="openstack/neutron-db-create-hrqhb" Nov 21 14:30:46 crc kubenswrapper[4897]: I1121 14:30:46.289381 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-f2a9-account-create-rx44g" Nov 21 14:30:46 crc kubenswrapper[4897]: I1121 14:30:46.338126 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/25643e36-c5eb-468b-85ce-e85cff1ad090-operator-scripts\") pod \"neutron-a766-account-create-btlq4\" (UID: \"25643e36-c5eb-468b-85ce-e85cff1ad090\") " pod="openstack/neutron-a766-account-create-btlq4" Nov 21 14:30:46 crc kubenswrapper[4897]: I1121 14:30:46.338890 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-th28m\" (UniqueName: \"kubernetes.io/projected/25643e36-c5eb-468b-85ce-e85cff1ad090-kube-api-access-th28m\") pod \"neutron-a766-account-create-btlq4\" (UID: \"25643e36-c5eb-468b-85ce-e85cff1ad090\") " pod="openstack/neutron-a766-account-create-btlq4" Nov 21 14:30:46 crc kubenswrapper[4897]: I1121 14:30:46.397266 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-hrqhb" Nov 21 14:30:46 crc kubenswrapper[4897]: I1121 14:30:46.441308 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/25643e36-c5eb-468b-85ce-e85cff1ad090-operator-scripts\") pod \"neutron-a766-account-create-btlq4\" (UID: \"25643e36-c5eb-468b-85ce-e85cff1ad090\") " pod="openstack/neutron-a766-account-create-btlq4" Nov 21 14:30:46 crc kubenswrapper[4897]: I1121 14:30:46.441416 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-th28m\" (UniqueName: \"kubernetes.io/projected/25643e36-c5eb-468b-85ce-e85cff1ad090-kube-api-access-th28m\") pod \"neutron-a766-account-create-btlq4\" (UID: \"25643e36-c5eb-468b-85ce-e85cff1ad090\") " pod="openstack/neutron-a766-account-create-btlq4" Nov 21 14:30:46 crc kubenswrapper[4897]: I1121 14:30:46.442049 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/25643e36-c5eb-468b-85ce-e85cff1ad090-operator-scripts\") pod \"neutron-a766-account-create-btlq4\" (UID: \"25643e36-c5eb-468b-85ce-e85cff1ad090\") " pod="openstack/neutron-a766-account-create-btlq4" Nov 21 14:30:46 crc kubenswrapper[4897]: I1121 14:30:46.461952 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-th28m\" (UniqueName: \"kubernetes.io/projected/25643e36-c5eb-468b-85ce-e85cff1ad090-kube-api-access-th28m\") pod \"neutron-a766-account-create-btlq4\" (UID: \"25643e36-c5eb-468b-85ce-e85cff1ad090\") " pod="openstack/neutron-a766-account-create-btlq4" Nov 21 14:30:46 crc kubenswrapper[4897]: I1121 14:30:46.501542 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-a766-account-create-btlq4" Nov 21 14:30:46 crc kubenswrapper[4897]: I1121 14:30:46.502332 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/prometheus-metric-storage-0" podUID="91aa2270-92a7-4af4-9a92-141ea2b24ee1" containerName="prometheus" probeResult="failure" output="HTTP probe failed with statuscode: 503" Nov 21 14:30:47 crc kubenswrapper[4897]: I1121 14:30:47.431757 4897 generic.go:334] "Generic (PLEG): container finished" podID="91aa2270-92a7-4af4-9a92-141ea2b24ee1" containerID="d75dc39a201e4f83e8ba5e273e81d6969573634dc7a6223e72bea5261f0bd8db" exitCode=0 Nov 21 14:30:47 crc kubenswrapper[4897]: I1121 14:30:47.432076 4897 generic.go:334] "Generic (PLEG): container finished" podID="91aa2270-92a7-4af4-9a92-141ea2b24ee1" containerID="1ae73b0cd6fdca18483acc4c5180aa46a081811e918b88fc1fb45977279fc1f2" exitCode=0 Nov 21 14:30:47 crc kubenswrapper[4897]: I1121 14:30:47.432099 4897 generic.go:334] "Generic (PLEG): container finished" podID="91aa2270-92a7-4af4-9a92-141ea2b24ee1" containerID="d4cba41371733c06d0f9282cb6e29395e6f232e62a3ce987fe67b0eb3d37cb98" exitCode=0 Nov 21 14:30:47 crc kubenswrapper[4897]: I1121 14:30:47.431850 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"91aa2270-92a7-4af4-9a92-141ea2b24ee1","Type":"ContainerDied","Data":"d75dc39a201e4f83e8ba5e273e81d6969573634dc7a6223e72bea5261f0bd8db"} Nov 21 14:30:47 crc kubenswrapper[4897]: I1121 14:30:47.432155 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"91aa2270-92a7-4af4-9a92-141ea2b24ee1","Type":"ContainerDied","Data":"1ae73b0cd6fdca18483acc4c5180aa46a081811e918b88fc1fb45977279fc1f2"} Nov 21 14:30:47 crc kubenswrapper[4897]: I1121 14:30:47.432180 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"91aa2270-92a7-4af4-9a92-141ea2b24ee1","Type":"ContainerDied","Data":"d4cba41371733c06d0f9282cb6e29395e6f232e62a3ce987fe67b0eb3d37cb98"} Nov 21 14:30:47 crc kubenswrapper[4897]: E1121 14:30:47.771578 4897 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-glance-api:current-podified" Nov 21 14:30:47 crc kubenswrapper[4897]: E1121 14:30:47.772055 4897 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:glance-db-sync,Image:quay.io/podified-antelope-centos9/openstack-glance-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/glance/glance.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gj9bx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42415,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42415,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-db-sync-zz6kx_openstack(9c08d8b3-d052-47b7-ad42-7b0f9fbe1920): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 21 14:30:47 crc kubenswrapper[4897]: E1121 14:30:47.775793 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/glance-db-sync-zz6kx" podUID="9c08d8b3-d052-47b7-ad42-7b0f9fbe1920" Nov 21 14:30:48 crc kubenswrapper[4897]: I1121 14:30:48.252905 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Nov 21 14:30:48 crc kubenswrapper[4897]: I1121 14:30:48.281464 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7060bc90-dec6-47cf-93b2-5e0ead1b71e7\") pod \"91aa2270-92a7-4af4-9a92-141ea2b24ee1\" (UID: \"91aa2270-92a7-4af4-9a92-141ea2b24ee1\") " Nov 21 14:30:48 crc kubenswrapper[4897]: I1121 14:30:48.281523 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hxhfg\" (UniqueName: \"kubernetes.io/projected/91aa2270-92a7-4af4-9a92-141ea2b24ee1-kube-api-access-hxhfg\") pod \"91aa2270-92a7-4af4-9a92-141ea2b24ee1\" (UID: \"91aa2270-92a7-4af4-9a92-141ea2b24ee1\") " Nov 21 14:30:48 crc kubenswrapper[4897]: I1121 14:30:48.281559 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/91aa2270-92a7-4af4-9a92-141ea2b24ee1-web-config\") pod \"91aa2270-92a7-4af4-9a92-141ea2b24ee1\" (UID: \"91aa2270-92a7-4af4-9a92-141ea2b24ee1\") " Nov 21 14:30:48 crc kubenswrapper[4897]: I1121 14:30:48.281740 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/91aa2270-92a7-4af4-9a92-141ea2b24ee1-config\") pod \"91aa2270-92a7-4af4-9a92-141ea2b24ee1\" (UID: \"91aa2270-92a7-4af4-9a92-141ea2b24ee1\") " Nov 21 14:30:48 crc kubenswrapper[4897]: I1121 14:30:48.281757 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/91aa2270-92a7-4af4-9a92-141ea2b24ee1-tls-assets\") pod \"91aa2270-92a7-4af4-9a92-141ea2b24ee1\" (UID: \"91aa2270-92a7-4af4-9a92-141ea2b24ee1\") " Nov 21 14:30:48 crc kubenswrapper[4897]: I1121 14:30:48.281808 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/91aa2270-92a7-4af4-9a92-141ea2b24ee1-config-out\") pod \"91aa2270-92a7-4af4-9a92-141ea2b24ee1\" (UID: \"91aa2270-92a7-4af4-9a92-141ea2b24ee1\") " Nov 21 14:30:48 crc kubenswrapper[4897]: I1121 14:30:48.281880 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/91aa2270-92a7-4af4-9a92-141ea2b24ee1-thanos-prometheus-http-client-file\") pod \"91aa2270-92a7-4af4-9a92-141ea2b24ee1\" (UID: \"91aa2270-92a7-4af4-9a92-141ea2b24ee1\") " Nov 21 14:30:48 crc kubenswrapper[4897]: I1121 14:30:48.281926 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/91aa2270-92a7-4af4-9a92-141ea2b24ee1-prometheus-metric-storage-rulefiles-0\") pod \"91aa2270-92a7-4af4-9a92-141ea2b24ee1\" (UID: \"91aa2270-92a7-4af4-9a92-141ea2b24ee1\") " Nov 21 14:30:48 crc kubenswrapper[4897]: I1121 14:30:48.283217 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/91aa2270-92a7-4af4-9a92-141ea2b24ee1-prometheus-metric-storage-rulefiles-0" (OuterVolumeSpecName: "prometheus-metric-storage-rulefiles-0") pod "91aa2270-92a7-4af4-9a92-141ea2b24ee1" (UID: "91aa2270-92a7-4af4-9a92-141ea2b24ee1"). InnerVolumeSpecName "prometheus-metric-storage-rulefiles-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:30:48 crc kubenswrapper[4897]: I1121 14:30:48.342639 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91aa2270-92a7-4af4-9a92-141ea2b24ee1-tls-assets" (OuterVolumeSpecName: "tls-assets") pod "91aa2270-92a7-4af4-9a92-141ea2b24ee1" (UID: "91aa2270-92a7-4af4-9a92-141ea2b24ee1"). InnerVolumeSpecName "tls-assets". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:30:48 crc kubenswrapper[4897]: I1121 14:30:48.344810 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/91aa2270-92a7-4af4-9a92-141ea2b24ee1-config-out" (OuterVolumeSpecName: "config-out") pod "91aa2270-92a7-4af4-9a92-141ea2b24ee1" (UID: "91aa2270-92a7-4af4-9a92-141ea2b24ee1"). InnerVolumeSpecName "config-out". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:30:48 crc kubenswrapper[4897]: I1121 14:30:48.348775 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91aa2270-92a7-4af4-9a92-141ea2b24ee1-thanos-prometheus-http-client-file" (OuterVolumeSpecName: "thanos-prometheus-http-client-file") pod "91aa2270-92a7-4af4-9a92-141ea2b24ee1" (UID: "91aa2270-92a7-4af4-9a92-141ea2b24ee1"). InnerVolumeSpecName "thanos-prometheus-http-client-file". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:30:48 crc kubenswrapper[4897]: I1121 14:30:48.349729 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91aa2270-92a7-4af4-9a92-141ea2b24ee1-kube-api-access-hxhfg" (OuterVolumeSpecName: "kube-api-access-hxhfg") pod "91aa2270-92a7-4af4-9a92-141ea2b24ee1" (UID: "91aa2270-92a7-4af4-9a92-141ea2b24ee1"). InnerVolumeSpecName "kube-api-access-hxhfg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:30:48 crc kubenswrapper[4897]: I1121 14:30:48.350928 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91aa2270-92a7-4af4-9a92-141ea2b24ee1-config" (OuterVolumeSpecName: "config") pod "91aa2270-92a7-4af4-9a92-141ea2b24ee1" (UID: "91aa2270-92a7-4af4-9a92-141ea2b24ee1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:30:48 crc kubenswrapper[4897]: I1121 14:30:48.368029 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7060bc90-dec6-47cf-93b2-5e0ead1b71e7" (OuterVolumeSpecName: "prometheus-metric-storage-db") pod "91aa2270-92a7-4af4-9a92-141ea2b24ee1" (UID: "91aa2270-92a7-4af4-9a92-141ea2b24ee1"). InnerVolumeSpecName "pvc-7060bc90-dec6-47cf-93b2-5e0ead1b71e7". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 21 14:30:48 crc kubenswrapper[4897]: I1121 14:30:48.386833 4897 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/91aa2270-92a7-4af4-9a92-141ea2b24ee1-config\") on node \"crc\" DevicePath \"\"" Nov 21 14:30:48 crc kubenswrapper[4897]: I1121 14:30:48.386868 4897 reconciler_common.go:293] "Volume detached for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/91aa2270-92a7-4af4-9a92-141ea2b24ee1-tls-assets\") on node \"crc\" DevicePath \"\"" Nov 21 14:30:48 crc kubenswrapper[4897]: I1121 14:30:48.386877 4897 reconciler_common.go:293] "Volume detached for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/91aa2270-92a7-4af4-9a92-141ea2b24ee1-config-out\") on node \"crc\" DevicePath \"\"" Nov 21 14:30:48 crc kubenswrapper[4897]: I1121 14:30:48.386886 4897 reconciler_common.go:293] "Volume detached for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/91aa2270-92a7-4af4-9a92-141ea2b24ee1-thanos-prometheus-http-client-file\") on node \"crc\" DevicePath \"\"" Nov 21 14:30:48 crc kubenswrapper[4897]: I1121 14:30:48.386897 4897 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/91aa2270-92a7-4af4-9a92-141ea2b24ee1-prometheus-metric-storage-rulefiles-0\") on node \"crc\" DevicePath \"\"" Nov 21 14:30:48 crc kubenswrapper[4897]: I1121 14:30:48.386933 4897 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-7060bc90-dec6-47cf-93b2-5e0ead1b71e7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7060bc90-dec6-47cf-93b2-5e0ead1b71e7\") on node \"crc\" " Nov 21 14:30:48 crc kubenswrapper[4897]: I1121 14:30:48.386946 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hxhfg\" (UniqueName: \"kubernetes.io/projected/91aa2270-92a7-4af4-9a92-141ea2b24ee1-kube-api-access-hxhfg\") on node \"crc\" DevicePath \"\"" Nov 21 14:30:48 crc kubenswrapper[4897]: I1121 14:30:48.390732 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91aa2270-92a7-4af4-9a92-141ea2b24ee1-web-config" (OuterVolumeSpecName: "web-config") pod "91aa2270-92a7-4af4-9a92-141ea2b24ee1" (UID: "91aa2270-92a7-4af4-9a92-141ea2b24ee1"). InnerVolumeSpecName "web-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:30:48 crc kubenswrapper[4897]: I1121 14:30:48.421841 4897 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Nov 21 14:30:48 crc kubenswrapper[4897]: I1121 14:30:48.421968 4897 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-7060bc90-dec6-47cf-93b2-5e0ead1b71e7" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7060bc90-dec6-47cf-93b2-5e0ead1b71e7") on node "crc" Nov 21 14:30:48 crc kubenswrapper[4897]: I1121 14:30:48.450949 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Nov 21 14:30:48 crc kubenswrapper[4897]: I1121 14:30:48.451142 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"91aa2270-92a7-4af4-9a92-141ea2b24ee1","Type":"ContainerDied","Data":"751b0fdc381aeca542be156ad6c66f3e7d3df396c03bdd04b0819a816b3e00e9"} Nov 21 14:30:48 crc kubenswrapper[4897]: I1121 14:30:48.451195 4897 scope.go:117] "RemoveContainer" containerID="d75dc39a201e4f83e8ba5e273e81d6969573634dc7a6223e72bea5261f0bd8db" Nov 21 14:30:48 crc kubenswrapper[4897]: E1121 14:30:48.462797 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-glance-api:current-podified\\\"\"" pod="openstack/glance-db-sync-zz6kx" podUID="9c08d8b3-d052-47b7-ad42-7b0f9fbe1920" Nov 21 14:30:48 crc kubenswrapper[4897]: I1121 14:30:48.492278 4897 reconciler_common.go:293] "Volume detached for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/91aa2270-92a7-4af4-9a92-141ea2b24ee1-web-config\") on node \"crc\" DevicePath \"\"" Nov 21 14:30:48 crc kubenswrapper[4897]: I1121 14:30:48.492336 4897 reconciler_common.go:293] "Volume detached for volume \"pvc-7060bc90-dec6-47cf-93b2-5e0ead1b71e7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7060bc90-dec6-47cf-93b2-5e0ead1b71e7\") on node \"crc\" DevicePath \"\"" Nov 21 14:30:48 crc kubenswrapper[4897]: I1121 14:30:48.635970 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 21 14:30:48 crc kubenswrapper[4897]: I1121 14:30:48.667619 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 21 14:30:48 crc kubenswrapper[4897]: I1121 14:30:48.685631 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 21 14:30:48 crc kubenswrapper[4897]: E1121 14:30:48.686314 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91aa2270-92a7-4af4-9a92-141ea2b24ee1" containerName="init-config-reloader" Nov 21 14:30:48 crc kubenswrapper[4897]: I1121 14:30:48.686337 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="91aa2270-92a7-4af4-9a92-141ea2b24ee1" containerName="init-config-reloader" Nov 21 14:30:48 crc kubenswrapper[4897]: E1121 14:30:48.686363 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91aa2270-92a7-4af4-9a92-141ea2b24ee1" containerName="prometheus" Nov 21 14:30:48 crc kubenswrapper[4897]: I1121 14:30:48.686370 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="91aa2270-92a7-4af4-9a92-141ea2b24ee1" containerName="prometheus" Nov 21 14:30:48 crc kubenswrapper[4897]: E1121 14:30:48.686392 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91aa2270-92a7-4af4-9a92-141ea2b24ee1" containerName="config-reloader" Nov 21 14:30:48 crc kubenswrapper[4897]: I1121 14:30:48.686398 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="91aa2270-92a7-4af4-9a92-141ea2b24ee1" containerName="config-reloader" Nov 21 14:30:48 crc kubenswrapper[4897]: E1121 14:30:48.686408 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91aa2270-92a7-4af4-9a92-141ea2b24ee1" containerName="thanos-sidecar" Nov 21 14:30:48 crc kubenswrapper[4897]: I1121 14:30:48.686415 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="91aa2270-92a7-4af4-9a92-141ea2b24ee1" containerName="thanos-sidecar" Nov 21 14:30:48 crc kubenswrapper[4897]: I1121 14:30:48.686628 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="91aa2270-92a7-4af4-9a92-141ea2b24ee1" containerName="prometheus" Nov 21 14:30:48 crc kubenswrapper[4897]: I1121 14:30:48.686645 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="91aa2270-92a7-4af4-9a92-141ea2b24ee1" containerName="thanos-sidecar" Nov 21 14:30:48 crc kubenswrapper[4897]: I1121 14:30:48.686656 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="91aa2270-92a7-4af4-9a92-141ea2b24ee1" containerName="config-reloader" Nov 21 14:30:48 crc kubenswrapper[4897]: I1121 14:30:48.688580 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Nov 21 14:30:48 crc kubenswrapper[4897]: I1121 14:30:48.691682 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-vshbl" Nov 21 14:30:48 crc kubenswrapper[4897]: I1121 14:30:48.691834 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Nov 21 14:30:48 crc kubenswrapper[4897]: I1121 14:30:48.691982 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-metric-storage-prometheus-svc" Nov 21 14:30:48 crc kubenswrapper[4897]: I1121 14:30:48.692067 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Nov 21 14:30:48 crc kubenswrapper[4897]: I1121 14:30:48.692275 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Nov 21 14:30:48 crc kubenswrapper[4897]: I1121 14:30:48.692455 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Nov 21 14:30:48 crc kubenswrapper[4897]: I1121 14:30:48.700838 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 21 14:30:48 crc kubenswrapper[4897]: I1121 14:30:48.704947 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Nov 21 14:30:48 crc kubenswrapper[4897]: I1121 14:30:48.803423 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7znnk\" (UniqueName: \"kubernetes.io/projected/fbc040d4-3c4c-4b07-8fd4-1eb61db44e5d-kube-api-access-7znnk\") pod \"prometheus-metric-storage-0\" (UID: \"fbc040d4-3c4c-4b07-8fd4-1eb61db44e5d\") " pod="openstack/prometheus-metric-storage-0" Nov 21 14:30:48 crc kubenswrapper[4897]: I1121 14:30:48.803466 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/fbc040d4-3c4c-4b07-8fd4-1eb61db44e5d-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"fbc040d4-3c4c-4b07-8fd4-1eb61db44e5d\") " pod="openstack/prometheus-metric-storage-0" Nov 21 14:30:48 crc kubenswrapper[4897]: I1121 14:30:48.803486 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/fbc040d4-3c4c-4b07-8fd4-1eb61db44e5d-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"fbc040d4-3c4c-4b07-8fd4-1eb61db44e5d\") " pod="openstack/prometheus-metric-storage-0" Nov 21 14:30:48 crc kubenswrapper[4897]: I1121 14:30:48.803534 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/fbc040d4-3c4c-4b07-8fd4-1eb61db44e5d-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"fbc040d4-3c4c-4b07-8fd4-1eb61db44e5d\") " pod="openstack/prometheus-metric-storage-0" Nov 21 14:30:48 crc kubenswrapper[4897]: I1121 14:30:48.803560 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-7060bc90-dec6-47cf-93b2-5e0ead1b71e7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7060bc90-dec6-47cf-93b2-5e0ead1b71e7\") pod \"prometheus-metric-storage-0\" (UID: \"fbc040d4-3c4c-4b07-8fd4-1eb61db44e5d\") " pod="openstack/prometheus-metric-storage-0" Nov 21 14:30:48 crc kubenswrapper[4897]: I1121 14:30:48.803584 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/fbc040d4-3c4c-4b07-8fd4-1eb61db44e5d-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"fbc040d4-3c4c-4b07-8fd4-1eb61db44e5d\") " pod="openstack/prometheus-metric-storage-0" Nov 21 14:30:48 crc kubenswrapper[4897]: I1121 14:30:48.803632 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/fbc040d4-3c4c-4b07-8fd4-1eb61db44e5d-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"fbc040d4-3c4c-4b07-8fd4-1eb61db44e5d\") " pod="openstack/prometheus-metric-storage-0" Nov 21 14:30:48 crc kubenswrapper[4897]: I1121 14:30:48.803775 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/fbc040d4-3c4c-4b07-8fd4-1eb61db44e5d-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"fbc040d4-3c4c-4b07-8fd4-1eb61db44e5d\") " pod="openstack/prometheus-metric-storage-0" Nov 21 14:30:48 crc kubenswrapper[4897]: I1121 14:30:48.803976 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/fbc040d4-3c4c-4b07-8fd4-1eb61db44e5d-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"fbc040d4-3c4c-4b07-8fd4-1eb61db44e5d\") " pod="openstack/prometheus-metric-storage-0" Nov 21 14:30:48 crc kubenswrapper[4897]: I1121 14:30:48.804033 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbc040d4-3c4c-4b07-8fd4-1eb61db44e5d-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"fbc040d4-3c4c-4b07-8fd4-1eb61db44e5d\") " pod="openstack/prometheus-metric-storage-0" Nov 21 14:30:48 crc kubenswrapper[4897]: I1121 14:30:48.804363 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/fbc040d4-3c4c-4b07-8fd4-1eb61db44e5d-config\") pod \"prometheus-metric-storage-0\" (UID: \"fbc040d4-3c4c-4b07-8fd4-1eb61db44e5d\") " pod="openstack/prometheus-metric-storage-0" Nov 21 14:30:48 crc kubenswrapper[4897]: E1121 14:30:48.817759 4897 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod91aa2270_92a7_4af4_9a92_141ea2b24ee1.slice/crio-751b0fdc381aeca542be156ad6c66f3e7d3df396c03bdd04b0819a816b3e00e9\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod91aa2270_92a7_4af4_9a92_141ea2b24ee1.slice\": RecentStats: unable to find data in memory cache]" Nov 21 14:30:48 crc kubenswrapper[4897]: E1121 14:30:48.825095 4897 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod91aa2270_92a7_4af4_9a92_141ea2b24ee1.slice\": RecentStats: unable to find data in memory cache]" Nov 21 14:30:48 crc kubenswrapper[4897]: I1121 14:30:48.906307 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-7060bc90-dec6-47cf-93b2-5e0ead1b71e7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7060bc90-dec6-47cf-93b2-5e0ead1b71e7\") pod \"prometheus-metric-storage-0\" (UID: \"fbc040d4-3c4c-4b07-8fd4-1eb61db44e5d\") " pod="openstack/prometheus-metric-storage-0" Nov 21 14:30:48 crc kubenswrapper[4897]: I1121 14:30:48.906349 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/fbc040d4-3c4c-4b07-8fd4-1eb61db44e5d-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"fbc040d4-3c4c-4b07-8fd4-1eb61db44e5d\") " pod="openstack/prometheus-metric-storage-0" Nov 21 14:30:48 crc kubenswrapper[4897]: I1121 14:30:48.906406 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/fbc040d4-3c4c-4b07-8fd4-1eb61db44e5d-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"fbc040d4-3c4c-4b07-8fd4-1eb61db44e5d\") " pod="openstack/prometheus-metric-storage-0" Nov 21 14:30:48 crc kubenswrapper[4897]: I1121 14:30:48.906431 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/fbc040d4-3c4c-4b07-8fd4-1eb61db44e5d-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"fbc040d4-3c4c-4b07-8fd4-1eb61db44e5d\") " pod="openstack/prometheus-metric-storage-0" Nov 21 14:30:48 crc kubenswrapper[4897]: I1121 14:30:48.906485 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/fbc040d4-3c4c-4b07-8fd4-1eb61db44e5d-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"fbc040d4-3c4c-4b07-8fd4-1eb61db44e5d\") " pod="openstack/prometheus-metric-storage-0" Nov 21 14:30:48 crc kubenswrapper[4897]: I1121 14:30:48.906544 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbc040d4-3c4c-4b07-8fd4-1eb61db44e5d-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"fbc040d4-3c4c-4b07-8fd4-1eb61db44e5d\") " pod="openstack/prometheus-metric-storage-0" Nov 21 14:30:48 crc kubenswrapper[4897]: I1121 14:30:48.906611 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/fbc040d4-3c4c-4b07-8fd4-1eb61db44e5d-config\") pod \"prometheus-metric-storage-0\" (UID: \"fbc040d4-3c4c-4b07-8fd4-1eb61db44e5d\") " pod="openstack/prometheus-metric-storage-0" Nov 21 14:30:48 crc kubenswrapper[4897]: I1121 14:30:48.906642 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7znnk\" (UniqueName: \"kubernetes.io/projected/fbc040d4-3c4c-4b07-8fd4-1eb61db44e5d-kube-api-access-7znnk\") pod \"prometheus-metric-storage-0\" (UID: \"fbc040d4-3c4c-4b07-8fd4-1eb61db44e5d\") " pod="openstack/prometheus-metric-storage-0" Nov 21 14:30:48 crc kubenswrapper[4897]: I1121 14:30:48.906659 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/fbc040d4-3c4c-4b07-8fd4-1eb61db44e5d-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"fbc040d4-3c4c-4b07-8fd4-1eb61db44e5d\") " pod="openstack/prometheus-metric-storage-0" Nov 21 14:30:48 crc kubenswrapper[4897]: I1121 14:30:48.906677 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/fbc040d4-3c4c-4b07-8fd4-1eb61db44e5d-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"fbc040d4-3c4c-4b07-8fd4-1eb61db44e5d\") " pod="openstack/prometheus-metric-storage-0" Nov 21 14:30:48 crc kubenswrapper[4897]: I1121 14:30:48.906705 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/fbc040d4-3c4c-4b07-8fd4-1eb61db44e5d-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"fbc040d4-3c4c-4b07-8fd4-1eb61db44e5d\") " pod="openstack/prometheus-metric-storage-0" Nov 21 14:30:48 crc kubenswrapper[4897]: I1121 14:30:48.907290 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/fbc040d4-3c4c-4b07-8fd4-1eb61db44e5d-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"fbc040d4-3c4c-4b07-8fd4-1eb61db44e5d\") " pod="openstack/prometheus-metric-storage-0" Nov 21 14:30:48 crc kubenswrapper[4897]: I1121 14:30:48.912788 4897 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 21 14:30:48 crc kubenswrapper[4897]: I1121 14:30:48.912979 4897 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-7060bc90-dec6-47cf-93b2-5e0ead1b71e7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7060bc90-dec6-47cf-93b2-5e0ead1b71e7\") pod \"prometheus-metric-storage-0\" (UID: \"fbc040d4-3c4c-4b07-8fd4-1eb61db44e5d\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/90579abefb2729890b6253e9dfb2d9cca2541535dfd7807baf8b98857bf5cb7e/globalmount\"" pod="openstack/prometheus-metric-storage-0" Nov 21 14:30:48 crc kubenswrapper[4897]: I1121 14:30:48.918783 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/fbc040d4-3c4c-4b07-8fd4-1eb61db44e5d-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"fbc040d4-3c4c-4b07-8fd4-1eb61db44e5d\") " pod="openstack/prometheus-metric-storage-0" Nov 21 14:30:48 crc kubenswrapper[4897]: I1121 14:30:48.924705 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/fbc040d4-3c4c-4b07-8fd4-1eb61db44e5d-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"fbc040d4-3c4c-4b07-8fd4-1eb61db44e5d\") " pod="openstack/prometheus-metric-storage-0" Nov 21 14:30:48 crc kubenswrapper[4897]: I1121 14:30:48.925160 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/fbc040d4-3c4c-4b07-8fd4-1eb61db44e5d-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"fbc040d4-3c4c-4b07-8fd4-1eb61db44e5d\") " pod="openstack/prometheus-metric-storage-0" Nov 21 14:30:48 crc kubenswrapper[4897]: I1121 14:30:48.927397 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/fbc040d4-3c4c-4b07-8fd4-1eb61db44e5d-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"fbc040d4-3c4c-4b07-8fd4-1eb61db44e5d\") " pod="openstack/prometheus-metric-storage-0" Nov 21 14:30:48 crc kubenswrapper[4897]: I1121 14:30:48.927545 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbc040d4-3c4c-4b07-8fd4-1eb61db44e5d-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"fbc040d4-3c4c-4b07-8fd4-1eb61db44e5d\") " pod="openstack/prometheus-metric-storage-0" Nov 21 14:30:48 crc kubenswrapper[4897]: I1121 14:30:48.927944 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/fbc040d4-3c4c-4b07-8fd4-1eb61db44e5d-config\") pod \"prometheus-metric-storage-0\" (UID: \"fbc040d4-3c4c-4b07-8fd4-1eb61db44e5d\") " pod="openstack/prometheus-metric-storage-0" Nov 21 14:30:48 crc kubenswrapper[4897]: I1121 14:30:48.928004 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/fbc040d4-3c4c-4b07-8fd4-1eb61db44e5d-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"fbc040d4-3c4c-4b07-8fd4-1eb61db44e5d\") " pod="openstack/prometheus-metric-storage-0" Nov 21 14:30:48 crc kubenswrapper[4897]: I1121 14:30:48.928443 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7znnk\" (UniqueName: \"kubernetes.io/projected/fbc040d4-3c4c-4b07-8fd4-1eb61db44e5d-kube-api-access-7znnk\") pod \"prometheus-metric-storage-0\" (UID: \"fbc040d4-3c4c-4b07-8fd4-1eb61db44e5d\") " pod="openstack/prometheus-metric-storage-0" Nov 21 14:30:48 crc kubenswrapper[4897]: I1121 14:30:48.931138 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/fbc040d4-3c4c-4b07-8fd4-1eb61db44e5d-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"fbc040d4-3c4c-4b07-8fd4-1eb61db44e5d\") " pod="openstack/prometheus-metric-storage-0" Nov 21 14:30:49 crc kubenswrapper[4897]: I1121 14:30:49.063851 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-hrqhb"] Nov 21 14:30:49 crc kubenswrapper[4897]: I1121 14:30:49.083563 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d5b6d6b67-qc7fs"] Nov 21 14:30:49 crc kubenswrapper[4897]: I1121 14:30:49.090434 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-c4c2-account-create-kzvn8"] Nov 21 14:30:49 crc kubenswrapper[4897]: I1121 14:30:49.102532 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-hg4dh"] Nov 21 14:30:49 crc kubenswrapper[4897]: I1121 14:30:49.176688 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-7060bc90-dec6-47cf-93b2-5e0ead1b71e7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7060bc90-dec6-47cf-93b2-5e0ead1b71e7\") pod \"prometheus-metric-storage-0\" (UID: \"fbc040d4-3c4c-4b07-8fd4-1eb61db44e5d\") " pod="openstack/prometheus-metric-storage-0" Nov 21 14:30:49 crc kubenswrapper[4897]: I1121 14:30:49.313978 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-dpnlm"] Nov 21 14:30:49 crc kubenswrapper[4897]: I1121 14:30:49.325265 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-5f6b-account-create-29rlc"] Nov 21 14:30:49 crc kubenswrapper[4897]: I1121 14:30:49.333786 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-f2a9-account-create-rx44g"] Nov 21 14:30:49 crc kubenswrapper[4897]: I1121 14:30:49.349735 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Nov 21 14:30:49 crc kubenswrapper[4897]: I1121 14:30:49.355987 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-227rw"] Nov 21 14:30:49 crc kubenswrapper[4897]: I1121 14:30:49.503806 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-a766-account-create-btlq4"] Nov 21 14:30:49 crc kubenswrapper[4897]: I1121 14:30:49.515879 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-cgbsk"] Nov 21 14:30:50 crc kubenswrapper[4897]: I1121 14:30:50.104494 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="91aa2270-92a7-4af4-9a92-141ea2b24ee1" path="/var/lib/kubelet/pods/91aa2270-92a7-4af4-9a92-141ea2b24ee1/volumes" Nov 21 14:30:50 crc kubenswrapper[4897]: W1121 14:30:50.765022 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod982ba7b6_1d19_4bd4_bdea_c828fd678186.slice/crio-1614aeb618cc32854fe6fd6a3661255a445d2ecf1db6ebac4a176700ce427a76 WatchSource:0}: Error finding container 1614aeb618cc32854fe6fd6a3661255a445d2ecf1db6ebac4a176700ce427a76: Status 404 returned error can't find the container with id 1614aeb618cc32854fe6fd6a3661255a445d2ecf1db6ebac4a176700ce427a76 Nov 21 14:30:50 crc kubenswrapper[4897]: W1121 14:30:50.777572 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod17a6c4a6_4493_4531_8f5c_814c7333a6ae.slice/crio-7c030fe110868de3a3e43d878600b726357395b387e561ca836f1ad54225b6f0 WatchSource:0}: Error finding container 7c030fe110868de3a3e43d878600b726357395b387e561ca836f1ad54225b6f0: Status 404 returned error can't find the container with id 7c030fe110868de3a3e43d878600b726357395b387e561ca836f1ad54225b6f0 Nov 21 14:30:50 crc kubenswrapper[4897]: W1121 14:30:50.781756 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda69b6579_ce5a_44d7_ab3d_5454dd5f85df.slice/crio-028b8a3f8f8da9052c2a1e2d67fb9be783942b649225fd9497474f2985080dd4 WatchSource:0}: Error finding container 028b8a3f8f8da9052c2a1e2d67fb9be783942b649225fd9497474f2985080dd4: Status 404 returned error can't find the container with id 028b8a3f8f8da9052c2a1e2d67fb9be783942b649225fd9497474f2985080dd4 Nov 21 14:30:50 crc kubenswrapper[4897]: W1121 14:30:50.787027 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8bb39c02_70b6_470f_bf62_add1c2ebc02f.slice/crio-00c88f9203968f749543de4dbe8afae7e77dc45520c2698a5026f4e9226730fb WatchSource:0}: Error finding container 00c88f9203968f749543de4dbe8afae7e77dc45520c2698a5026f4e9226730fb: Status 404 returned error can't find the container with id 00c88f9203968f749543de4dbe8afae7e77dc45520c2698a5026f4e9226730fb Nov 21 14:30:50 crc kubenswrapper[4897]: W1121 14:30:50.806400 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod56c583f0_0dba_4e94_a8b7_26d7b5c0e87f.slice/crio-1090d2af284c51892bcb7fa7a6a0de257271ebb6659e0178eaca286f47a34b7f WatchSource:0}: Error finding container 1090d2af284c51892bcb7fa7a6a0de257271ebb6659e0178eaca286f47a34b7f: Status 404 returned error can't find the container with id 1090d2af284c51892bcb7fa7a6a0de257271ebb6659e0178eaca286f47a34b7f Nov 21 14:30:50 crc kubenswrapper[4897]: W1121 14:30:50.808682 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod208ac008_5901_4b46_983a_a73c9ed207f2.slice/crio-114beb74200ad93f12e7ed339efdc21758f06e09563638274a1a6ac884304735 WatchSource:0}: Error finding container 114beb74200ad93f12e7ed339efdc21758f06e09563638274a1a6ac884304735: Status 404 returned error can't find the container with id 114beb74200ad93f12e7ed339efdc21758f06e09563638274a1a6ac884304735 Nov 21 14:30:50 crc kubenswrapper[4897]: W1121 14:30:50.812852 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1fd99443_7153_4711_9575_6ebc4fbfcf42.slice/crio-1cb5d006c2eb51c868e0945688cad3f46e74fef8b259852eb7f281589a14a76b WatchSource:0}: Error finding container 1cb5d006c2eb51c868e0945688cad3f46e74fef8b259852eb7f281589a14a76b: Status 404 returned error can't find the container with id 1cb5d006c2eb51c868e0945688cad3f46e74fef8b259852eb7f281589a14a76b Nov 21 14:30:51 crc kubenswrapper[4897]: I1121 14:30:51.070028 4897 scope.go:117] "RemoveContainer" containerID="1ae73b0cd6fdca18483acc4c5180aa46a081811e918b88fc1fb45977279fc1f2" Nov 21 14:30:51 crc kubenswrapper[4897]: I1121 14:30:51.282037 4897 scope.go:117] "RemoveContainer" containerID="d4cba41371733c06d0f9282cb6e29395e6f232e62a3ce987fe67b0eb3d37cb98" Nov 21 14:30:51 crc kubenswrapper[4897]: I1121 14:30:51.501567 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-227rw" event={"ID":"208ac008-5901-4b46-983a-a73c9ed207f2","Type":"ContainerStarted","Data":"114beb74200ad93f12e7ed339efdc21758f06e09563638274a1a6ac884304735"} Nov 21 14:30:51 crc kubenswrapper[4897]: I1121 14:30:51.506230 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-dpnlm" event={"ID":"67dd6a02-c8f6-4035-ad18-c6bbea13c688","Type":"ContainerStarted","Data":"c7befc62974d1f8c10dcf4aa82eebb822773d3e6f69cce8bd59098b4a662a6d6"} Nov 21 14:30:51 crc kubenswrapper[4897]: I1121 14:30:51.507480 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-hrqhb" event={"ID":"982ba7b6-1d19-4bd4-bdea-c828fd678186","Type":"ContainerStarted","Data":"1614aeb618cc32854fe6fd6a3661255a445d2ecf1db6ebac4a176700ce427a76"} Nov 21 14:30:51 crc kubenswrapper[4897]: I1121 14:30:51.508815 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-hg4dh" event={"ID":"ee8a4380-421f-45b1-ba19-9832b69935fe","Type":"ContainerStarted","Data":"35c08b523620f504553fdab9f8f75b551b43014d9ba7b0404b6e7f81ed83b244"} Nov 21 14:30:51 crc kubenswrapper[4897]: I1121 14:30:51.511948 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-f2a9-account-create-rx44g" event={"ID":"1fd99443-7153-4711-9575-6ebc4fbfcf42","Type":"ContainerStarted","Data":"1cb5d006c2eb51c868e0945688cad3f46e74fef8b259852eb7f281589a14a76b"} Nov 21 14:30:51 crc kubenswrapper[4897]: I1121 14:30:51.513703 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d5b6d6b67-qc7fs" event={"ID":"17a6c4a6-4493-4531-8f5c-814c7333a6ae","Type":"ContainerStarted","Data":"7c030fe110868de3a3e43d878600b726357395b387e561ca836f1ad54225b6f0"} Nov 21 14:30:51 crc kubenswrapper[4897]: I1121 14:30:51.516237 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-a766-account-create-btlq4" event={"ID":"25643e36-c5eb-468b-85ce-e85cff1ad090","Type":"ContainerStarted","Data":"6e28d7e2796437459c660202dd15e5b41b2473a09bb2412c98c55cc5d74c5d1b"} Nov 21 14:30:51 crc kubenswrapper[4897]: I1121 14:30:51.518163 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-c4c2-account-create-kzvn8" event={"ID":"56c583f0-0dba-4e94-a8b7-26d7b5c0e87f","Type":"ContainerStarted","Data":"1090d2af284c51892bcb7fa7a6a0de257271ebb6659e0178eaca286f47a34b7f"} Nov 21 14:30:51 crc kubenswrapper[4897]: I1121 14:30:51.520294 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-5f6b-account-create-29rlc" event={"ID":"a69b6579-ce5a-44d7-ab3d-5454dd5f85df","Type":"ContainerStarted","Data":"028b8a3f8f8da9052c2a1e2d67fb9be783942b649225fd9497474f2985080dd4"} Nov 21 14:30:51 crc kubenswrapper[4897]: I1121 14:30:51.523777 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-cgbsk" event={"ID":"8bb39c02-70b6-470f-bf62-add1c2ebc02f","Type":"ContainerStarted","Data":"00c88f9203968f749543de4dbe8afae7e77dc45520c2698a5026f4e9226730fb"} Nov 21 14:30:51 crc kubenswrapper[4897]: I1121 14:30:51.527300 4897 scope.go:117] "RemoveContainer" containerID="ae9d2f28e206da418bad9d39e2d6c5076e3d741d413ec6b6cb6c3b13dfa61274" Nov 21 14:30:51 crc kubenswrapper[4897]: I1121 14:30:51.750845 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 21 14:30:51 crc kubenswrapper[4897]: W1121 14:30:51.762748 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfbc040d4_3c4c_4b07_8fd4_1eb61db44e5d.slice/crio-034ae314edde747738d9db8ffe7a3bca8b0a7b136f4745c3ff59c50c3d376d5c WatchSource:0}: Error finding container 034ae314edde747738d9db8ffe7a3bca8b0a7b136f4745c3ff59c50c3d376d5c: Status 404 returned error can't find the container with id 034ae314edde747738d9db8ffe7a3bca8b0a7b136f4745c3ff59c50c3d376d5c Nov 21 14:30:52 crc kubenswrapper[4897]: I1121 14:30:52.537382 4897 generic.go:334] "Generic (PLEG): container finished" podID="8bb39c02-70b6-470f-bf62-add1c2ebc02f" containerID="8781c24d75d3e6b5d39b1e57310980d15f0af48994ee70807841ba07c7ff9b3c" exitCode=0 Nov 21 14:30:52 crc kubenswrapper[4897]: I1121 14:30:52.537425 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-cgbsk" event={"ID":"8bb39c02-70b6-470f-bf62-add1c2ebc02f","Type":"ContainerDied","Data":"8781c24d75d3e6b5d39b1e57310980d15f0af48994ee70807841ba07c7ff9b3c"} Nov 21 14:30:52 crc kubenswrapper[4897]: I1121 14:30:52.539898 4897 generic.go:334] "Generic (PLEG): container finished" podID="982ba7b6-1d19-4bd4-bdea-c828fd678186" containerID="f611821b857e8dab1b5c82f556bf92d1d63333064f5fd2e8326c8cc267cfa952" exitCode=0 Nov 21 14:30:52 crc kubenswrapper[4897]: I1121 14:30:52.539973 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-hrqhb" event={"ID":"982ba7b6-1d19-4bd4-bdea-c828fd678186","Type":"ContainerDied","Data":"f611821b857e8dab1b5c82f556bf92d1d63333064f5fd2e8326c8cc267cfa952"} Nov 21 14:30:52 crc kubenswrapper[4897]: I1121 14:30:52.542535 4897 generic.go:334] "Generic (PLEG): container finished" podID="ee8a4380-421f-45b1-ba19-9832b69935fe" containerID="0cb696c4ee7260fa5bb8919890a7625680b2a0d45b07bff3285343319b1ea02f" exitCode=0 Nov 21 14:30:52 crc kubenswrapper[4897]: I1121 14:30:52.542627 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-hg4dh" event={"ID":"ee8a4380-421f-45b1-ba19-9832b69935fe","Type":"ContainerDied","Data":"0cb696c4ee7260fa5bb8919890a7625680b2a0d45b07bff3285343319b1ea02f"} Nov 21 14:30:52 crc kubenswrapper[4897]: I1121 14:30:52.545338 4897 generic.go:334] "Generic (PLEG): container finished" podID="1fd99443-7153-4711-9575-6ebc4fbfcf42" containerID="940d3d57b3424e16b5ced6b938dfcf9dc5e5dff3244c38fa7a71385046318ac2" exitCode=0 Nov 21 14:30:52 crc kubenswrapper[4897]: I1121 14:30:52.545367 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-f2a9-account-create-rx44g" event={"ID":"1fd99443-7153-4711-9575-6ebc4fbfcf42","Type":"ContainerDied","Data":"940d3d57b3424e16b5ced6b938dfcf9dc5e5dff3244c38fa7a71385046318ac2"} Nov 21 14:30:52 crc kubenswrapper[4897]: I1121 14:30:52.546491 4897 generic.go:334] "Generic (PLEG): container finished" podID="17a6c4a6-4493-4531-8f5c-814c7333a6ae" containerID="164ea673a2b950fa6284d85e4eb1fe8e329dd37706927daf4eaa21b953f48acb" exitCode=0 Nov 21 14:30:52 crc kubenswrapper[4897]: I1121 14:30:52.546578 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d5b6d6b67-qc7fs" event={"ID":"17a6c4a6-4493-4531-8f5c-814c7333a6ae","Type":"ContainerDied","Data":"164ea673a2b950fa6284d85e4eb1fe8e329dd37706927daf4eaa21b953f48acb"} Nov 21 14:30:52 crc kubenswrapper[4897]: I1121 14:30:52.549146 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"fbc040d4-3c4c-4b07-8fd4-1eb61db44e5d","Type":"ContainerStarted","Data":"034ae314edde747738d9db8ffe7a3bca8b0a7b136f4745c3ff59c50c3d376d5c"} Nov 21 14:30:52 crc kubenswrapper[4897]: I1121 14:30:52.552299 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0" event={"ID":"813eae15-fec1-4d3a-bae2-1dc597c528f7","Type":"ContainerStarted","Data":"73c0e396184e52ce4ab802d647e0f9d82f1f11f7f86b7c1d9df80065526d91e9"} Nov 21 14:30:52 crc kubenswrapper[4897]: I1121 14:30:52.553934 4897 generic.go:334] "Generic (PLEG): container finished" podID="25643e36-c5eb-468b-85ce-e85cff1ad090" containerID="ede6a9437617ca4a9d2c0f1600f1fe7faedf4bcf189909eb8f3195b61567ed58" exitCode=0 Nov 21 14:30:52 crc kubenswrapper[4897]: I1121 14:30:52.553998 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-a766-account-create-btlq4" event={"ID":"25643e36-c5eb-468b-85ce-e85cff1ad090","Type":"ContainerDied","Data":"ede6a9437617ca4a9d2c0f1600f1fe7faedf4bcf189909eb8f3195b61567ed58"} Nov 21 14:30:52 crc kubenswrapper[4897]: I1121 14:30:52.555763 4897 generic.go:334] "Generic (PLEG): container finished" podID="56c583f0-0dba-4e94-a8b7-26d7b5c0e87f" containerID="8de5ff5ee2cf9cf2c61e0a6143f2ebc4b16a007274b3b322e3a275de8e4ce95a" exitCode=0 Nov 21 14:30:52 crc kubenswrapper[4897]: I1121 14:30:52.555811 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-c4c2-account-create-kzvn8" event={"ID":"56c583f0-0dba-4e94-a8b7-26d7b5c0e87f","Type":"ContainerDied","Data":"8de5ff5ee2cf9cf2c61e0a6143f2ebc4b16a007274b3b322e3a275de8e4ce95a"} Nov 21 14:30:52 crc kubenswrapper[4897]: I1121 14:30:52.560760 4897 generic.go:334] "Generic (PLEG): container finished" podID="67dd6a02-c8f6-4035-ad18-c6bbea13c688" containerID="2a2f30c93a24730d081746bf1340d27054ff14feeb84ac94d2f98b5707eb77f6" exitCode=0 Nov 21 14:30:52 crc kubenswrapper[4897]: I1121 14:30:52.560848 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-dpnlm" event={"ID":"67dd6a02-c8f6-4035-ad18-c6bbea13c688","Type":"ContainerDied","Data":"2a2f30c93a24730d081746bf1340d27054ff14feeb84ac94d2f98b5707eb77f6"} Nov 21 14:30:52 crc kubenswrapper[4897]: I1121 14:30:52.565466 4897 generic.go:334] "Generic (PLEG): container finished" podID="a69b6579-ce5a-44d7-ab3d-5454dd5f85df" containerID="fce9d4dbbab86c5b7bd18a3bc4637c387571935a1eb7826370d6345f299fe008" exitCode=0 Nov 21 14:30:52 crc kubenswrapper[4897]: I1121 14:30:52.565525 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-5f6b-account-create-29rlc" event={"ID":"a69b6579-ce5a-44d7-ab3d-5454dd5f85df","Type":"ContainerDied","Data":"fce9d4dbbab86c5b7bd18a3bc4637c387571935a1eb7826370d6345f299fe008"} Nov 21 14:30:52 crc kubenswrapper[4897]: I1121 14:30:52.683524 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mysqld-exporter-0" podStartSLOduration=2.968624923 podStartE2EDuration="17.683488903s" podCreationTimestamp="2025-11-21 14:30:35 +0000 UTC" firstStartedPulling="2025-11-21 14:30:36.606757443 +0000 UTC m=+1313.891350918" lastFinishedPulling="2025-11-21 14:30:51.321621423 +0000 UTC m=+1328.606214898" observedRunningTime="2025-11-21 14:30:52.67330399 +0000 UTC m=+1329.957897465" watchObservedRunningTime="2025-11-21 14:30:52.683488903 +0000 UTC m=+1329.968082378" Nov 21 14:30:53 crc kubenswrapper[4897]: I1121 14:30:53.577435 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d5b6d6b67-qc7fs" event={"ID":"17a6c4a6-4493-4531-8f5c-814c7333a6ae","Type":"ContainerStarted","Data":"afe2abe49994bcaffbe92412ab26a0240d4102be177179d0e806cea09b185b65"} Nov 21 14:30:53 crc kubenswrapper[4897]: I1121 14:30:53.579107 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6d5b6d6b67-qc7fs" Nov 21 14:30:53 crc kubenswrapper[4897]: I1121 14:30:53.604921 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6d5b6d6b67-qc7fs" podStartSLOduration=16.604899658 podStartE2EDuration="16.604899658s" podCreationTimestamp="2025-11-21 14:30:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:30:53.597056738 +0000 UTC m=+1330.881650223" watchObservedRunningTime="2025-11-21 14:30:53.604899658 +0000 UTC m=+1330.889493133" Nov 21 14:30:55 crc kubenswrapper[4897]: I1121 14:30:55.596689 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"fbc040d4-3c4c-4b07-8fd4-1eb61db44e5d","Type":"ContainerStarted","Data":"81c07f5180c87605c31b38047cf5d4bc886014b4de5781764c005b00293f6cde"} Nov 21 14:30:55 crc kubenswrapper[4897]: I1121 14:30:55.855905 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-cgbsk" Nov 21 14:30:55 crc kubenswrapper[4897]: I1121 14:30:55.861362 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-hrqhb" Nov 21 14:30:55 crc kubenswrapper[4897]: I1121 14:30:55.892539 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-dpnlm" Nov 21 14:30:55 crc kubenswrapper[4897]: I1121 14:30:55.904344 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-f2a9-account-create-rx44g" Nov 21 14:30:55 crc kubenswrapper[4897]: I1121 14:30:55.922976 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-5f6b-account-create-29rlc" Nov 21 14:30:55 crc kubenswrapper[4897]: I1121 14:30:55.939106 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-c4c2-account-create-kzvn8" Nov 21 14:30:55 crc kubenswrapper[4897]: I1121 14:30:55.945128 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-a766-account-create-btlq4" Nov 21 14:30:55 crc kubenswrapper[4897]: I1121 14:30:55.968865 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-hg4dh" Nov 21 14:30:55 crc kubenswrapper[4897]: I1121 14:30:55.973206 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4x7ln\" (UniqueName: \"kubernetes.io/projected/8bb39c02-70b6-470f-bf62-add1c2ebc02f-kube-api-access-4x7ln\") pod \"8bb39c02-70b6-470f-bf62-add1c2ebc02f\" (UID: \"8bb39c02-70b6-470f-bf62-add1c2ebc02f\") " Nov 21 14:30:55 crc kubenswrapper[4897]: I1121 14:30:55.973261 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/67dd6a02-c8f6-4035-ad18-c6bbea13c688-operator-scripts\") pod \"67dd6a02-c8f6-4035-ad18-c6bbea13c688\" (UID: \"67dd6a02-c8f6-4035-ad18-c6bbea13c688\") " Nov 21 14:30:55 crc kubenswrapper[4897]: I1121 14:30:55.973410 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8bb39c02-70b6-470f-bf62-add1c2ebc02f-operator-scripts\") pod \"8bb39c02-70b6-470f-bf62-add1c2ebc02f\" (UID: \"8bb39c02-70b6-470f-bf62-add1c2ebc02f\") " Nov 21 14:30:55 crc kubenswrapper[4897]: I1121 14:30:55.973461 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v8zbf\" (UniqueName: \"kubernetes.io/projected/982ba7b6-1d19-4bd4-bdea-c828fd678186-kube-api-access-v8zbf\") pod \"982ba7b6-1d19-4bd4-bdea-c828fd678186\" (UID: \"982ba7b6-1d19-4bd4-bdea-c828fd678186\") " Nov 21 14:30:55 crc kubenswrapper[4897]: I1121 14:30:55.973572 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-49wwn\" (UniqueName: \"kubernetes.io/projected/67dd6a02-c8f6-4035-ad18-c6bbea13c688-kube-api-access-49wwn\") pod \"67dd6a02-c8f6-4035-ad18-c6bbea13c688\" (UID: \"67dd6a02-c8f6-4035-ad18-c6bbea13c688\") " Nov 21 14:30:55 crc kubenswrapper[4897]: I1121 14:30:55.973649 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/982ba7b6-1d19-4bd4-bdea-c828fd678186-operator-scripts\") pod \"982ba7b6-1d19-4bd4-bdea-c828fd678186\" (UID: \"982ba7b6-1d19-4bd4-bdea-c828fd678186\") " Nov 21 14:30:55 crc kubenswrapper[4897]: I1121 14:30:55.975152 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/982ba7b6-1d19-4bd4-bdea-c828fd678186-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "982ba7b6-1d19-4bd4-bdea-c828fd678186" (UID: "982ba7b6-1d19-4bd4-bdea-c828fd678186"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:30:55 crc kubenswrapper[4897]: I1121 14:30:55.977925 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8bb39c02-70b6-470f-bf62-add1c2ebc02f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8bb39c02-70b6-470f-bf62-add1c2ebc02f" (UID: "8bb39c02-70b6-470f-bf62-add1c2ebc02f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:30:55 crc kubenswrapper[4897]: I1121 14:30:55.977964 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/67dd6a02-c8f6-4035-ad18-c6bbea13c688-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "67dd6a02-c8f6-4035-ad18-c6bbea13c688" (UID: "67dd6a02-c8f6-4035-ad18-c6bbea13c688"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:30:55 crc kubenswrapper[4897]: I1121 14:30:55.995038 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/982ba7b6-1d19-4bd4-bdea-c828fd678186-kube-api-access-v8zbf" (OuterVolumeSpecName: "kube-api-access-v8zbf") pod "982ba7b6-1d19-4bd4-bdea-c828fd678186" (UID: "982ba7b6-1d19-4bd4-bdea-c828fd678186"). InnerVolumeSpecName "kube-api-access-v8zbf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:30:56 crc kubenswrapper[4897]: I1121 14:30:55.999965 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8bb39c02-70b6-470f-bf62-add1c2ebc02f-kube-api-access-4x7ln" (OuterVolumeSpecName: "kube-api-access-4x7ln") pod "8bb39c02-70b6-470f-bf62-add1c2ebc02f" (UID: "8bb39c02-70b6-470f-bf62-add1c2ebc02f"). InnerVolumeSpecName "kube-api-access-4x7ln". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:30:56 crc kubenswrapper[4897]: I1121 14:30:56.003391 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67dd6a02-c8f6-4035-ad18-c6bbea13c688-kube-api-access-49wwn" (OuterVolumeSpecName: "kube-api-access-49wwn") pod "67dd6a02-c8f6-4035-ad18-c6bbea13c688" (UID: "67dd6a02-c8f6-4035-ad18-c6bbea13c688"). InnerVolumeSpecName "kube-api-access-49wwn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:30:56 crc kubenswrapper[4897]: I1121 14:30:56.075587 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ksmjv\" (UniqueName: \"kubernetes.io/projected/a69b6579-ce5a-44d7-ab3d-5454dd5f85df-kube-api-access-ksmjv\") pod \"a69b6579-ce5a-44d7-ab3d-5454dd5f85df\" (UID: \"a69b6579-ce5a-44d7-ab3d-5454dd5f85df\") " Nov 21 14:30:56 crc kubenswrapper[4897]: I1121 14:30:56.075680 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pzsg2\" (UniqueName: \"kubernetes.io/projected/ee8a4380-421f-45b1-ba19-9832b69935fe-kube-api-access-pzsg2\") pod \"ee8a4380-421f-45b1-ba19-9832b69935fe\" (UID: \"ee8a4380-421f-45b1-ba19-9832b69935fe\") " Nov 21 14:30:56 crc kubenswrapper[4897]: I1121 14:30:56.075701 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/25643e36-c5eb-468b-85ce-e85cff1ad090-operator-scripts\") pod \"25643e36-c5eb-468b-85ce-e85cff1ad090\" (UID: \"25643e36-c5eb-468b-85ce-e85cff1ad090\") " Nov 21 14:30:56 crc kubenswrapper[4897]: I1121 14:30:56.075742 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sf4jq\" (UniqueName: \"kubernetes.io/projected/1fd99443-7153-4711-9575-6ebc4fbfcf42-kube-api-access-sf4jq\") pod \"1fd99443-7153-4711-9575-6ebc4fbfcf42\" (UID: \"1fd99443-7153-4711-9575-6ebc4fbfcf42\") " Nov 21 14:30:56 crc kubenswrapper[4897]: I1121 14:30:56.075790 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1fd99443-7153-4711-9575-6ebc4fbfcf42-operator-scripts\") pod \"1fd99443-7153-4711-9575-6ebc4fbfcf42\" (UID: \"1fd99443-7153-4711-9575-6ebc4fbfcf42\") " Nov 21 14:30:56 crc kubenswrapper[4897]: I1121 14:30:56.075831 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ee8a4380-421f-45b1-ba19-9832b69935fe-operator-scripts\") pod \"ee8a4380-421f-45b1-ba19-9832b69935fe\" (UID: \"ee8a4380-421f-45b1-ba19-9832b69935fe\") " Nov 21 14:30:56 crc kubenswrapper[4897]: I1121 14:30:56.075867 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a69b6579-ce5a-44d7-ab3d-5454dd5f85df-operator-scripts\") pod \"a69b6579-ce5a-44d7-ab3d-5454dd5f85df\" (UID: \"a69b6579-ce5a-44d7-ab3d-5454dd5f85df\") " Nov 21 14:30:56 crc kubenswrapper[4897]: I1121 14:30:56.075895 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/56c583f0-0dba-4e94-a8b7-26d7b5c0e87f-operator-scripts\") pod \"56c583f0-0dba-4e94-a8b7-26d7b5c0e87f\" (UID: \"56c583f0-0dba-4e94-a8b7-26d7b5c0e87f\") " Nov 21 14:30:56 crc kubenswrapper[4897]: I1121 14:30:56.075988 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-th28m\" (UniqueName: \"kubernetes.io/projected/25643e36-c5eb-468b-85ce-e85cff1ad090-kube-api-access-th28m\") pod \"25643e36-c5eb-468b-85ce-e85cff1ad090\" (UID: \"25643e36-c5eb-468b-85ce-e85cff1ad090\") " Nov 21 14:30:56 crc kubenswrapper[4897]: I1121 14:30:56.076040 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xzdbv\" (UniqueName: \"kubernetes.io/projected/56c583f0-0dba-4e94-a8b7-26d7b5c0e87f-kube-api-access-xzdbv\") pod \"56c583f0-0dba-4e94-a8b7-26d7b5c0e87f\" (UID: \"56c583f0-0dba-4e94-a8b7-26d7b5c0e87f\") " Nov 21 14:30:56 crc kubenswrapper[4897]: I1121 14:30:56.076301 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25643e36-c5eb-468b-85ce-e85cff1ad090-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "25643e36-c5eb-468b-85ce-e85cff1ad090" (UID: "25643e36-c5eb-468b-85ce-e85cff1ad090"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:30:56 crc kubenswrapper[4897]: I1121 14:30:56.076483 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1fd99443-7153-4711-9575-6ebc4fbfcf42-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1fd99443-7153-4711-9575-6ebc4fbfcf42" (UID: "1fd99443-7153-4711-9575-6ebc4fbfcf42"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:30:56 crc kubenswrapper[4897]: I1121 14:30:56.076561 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-49wwn\" (UniqueName: \"kubernetes.io/projected/67dd6a02-c8f6-4035-ad18-c6bbea13c688-kube-api-access-49wwn\") on node \"crc\" DevicePath \"\"" Nov 21 14:30:56 crc kubenswrapper[4897]: I1121 14:30:56.076580 4897 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/982ba7b6-1d19-4bd4-bdea-c828fd678186-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 21 14:30:56 crc kubenswrapper[4897]: I1121 14:30:56.076589 4897 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/25643e36-c5eb-468b-85ce-e85cff1ad090-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 21 14:30:56 crc kubenswrapper[4897]: I1121 14:30:56.076600 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4x7ln\" (UniqueName: \"kubernetes.io/projected/8bb39c02-70b6-470f-bf62-add1c2ebc02f-kube-api-access-4x7ln\") on node \"crc\" DevicePath \"\"" Nov 21 14:30:56 crc kubenswrapper[4897]: I1121 14:30:56.076608 4897 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/67dd6a02-c8f6-4035-ad18-c6bbea13c688-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 21 14:30:56 crc kubenswrapper[4897]: I1121 14:30:56.076604 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee8a4380-421f-45b1-ba19-9832b69935fe-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ee8a4380-421f-45b1-ba19-9832b69935fe" (UID: "ee8a4380-421f-45b1-ba19-9832b69935fe"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:30:56 crc kubenswrapper[4897]: I1121 14:30:56.076617 4897 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8bb39c02-70b6-470f-bf62-add1c2ebc02f-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 21 14:30:56 crc kubenswrapper[4897]: I1121 14:30:56.076643 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v8zbf\" (UniqueName: \"kubernetes.io/projected/982ba7b6-1d19-4bd4-bdea-c828fd678186-kube-api-access-v8zbf\") on node \"crc\" DevicePath \"\"" Nov 21 14:30:56 crc kubenswrapper[4897]: I1121 14:30:56.076644 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a69b6579-ce5a-44d7-ab3d-5454dd5f85df-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a69b6579-ce5a-44d7-ab3d-5454dd5f85df" (UID: "a69b6579-ce5a-44d7-ab3d-5454dd5f85df"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:30:56 crc kubenswrapper[4897]: I1121 14:30:56.076687 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/56c583f0-0dba-4e94-a8b7-26d7b5c0e87f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "56c583f0-0dba-4e94-a8b7-26d7b5c0e87f" (UID: "56c583f0-0dba-4e94-a8b7-26d7b5c0e87f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:30:56 crc kubenswrapper[4897]: I1121 14:30:56.079226 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a69b6579-ce5a-44d7-ab3d-5454dd5f85df-kube-api-access-ksmjv" (OuterVolumeSpecName: "kube-api-access-ksmjv") pod "a69b6579-ce5a-44d7-ab3d-5454dd5f85df" (UID: "a69b6579-ce5a-44d7-ab3d-5454dd5f85df"). InnerVolumeSpecName "kube-api-access-ksmjv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:30:56 crc kubenswrapper[4897]: I1121 14:30:56.079740 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56c583f0-0dba-4e94-a8b7-26d7b5c0e87f-kube-api-access-xzdbv" (OuterVolumeSpecName: "kube-api-access-xzdbv") pod "56c583f0-0dba-4e94-a8b7-26d7b5c0e87f" (UID: "56c583f0-0dba-4e94-a8b7-26d7b5c0e87f"). InnerVolumeSpecName "kube-api-access-xzdbv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:30:56 crc kubenswrapper[4897]: I1121 14:30:56.080087 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee8a4380-421f-45b1-ba19-9832b69935fe-kube-api-access-pzsg2" (OuterVolumeSpecName: "kube-api-access-pzsg2") pod "ee8a4380-421f-45b1-ba19-9832b69935fe" (UID: "ee8a4380-421f-45b1-ba19-9832b69935fe"). InnerVolumeSpecName "kube-api-access-pzsg2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:30:56 crc kubenswrapper[4897]: I1121 14:30:56.080606 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25643e36-c5eb-468b-85ce-e85cff1ad090-kube-api-access-th28m" (OuterVolumeSpecName: "kube-api-access-th28m") pod "25643e36-c5eb-468b-85ce-e85cff1ad090" (UID: "25643e36-c5eb-468b-85ce-e85cff1ad090"). InnerVolumeSpecName "kube-api-access-th28m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:30:56 crc kubenswrapper[4897]: I1121 14:30:56.081011 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1fd99443-7153-4711-9575-6ebc4fbfcf42-kube-api-access-sf4jq" (OuterVolumeSpecName: "kube-api-access-sf4jq") pod "1fd99443-7153-4711-9575-6ebc4fbfcf42" (UID: "1fd99443-7153-4711-9575-6ebc4fbfcf42"). InnerVolumeSpecName "kube-api-access-sf4jq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:30:56 crc kubenswrapper[4897]: I1121 14:30:56.178534 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xzdbv\" (UniqueName: \"kubernetes.io/projected/56c583f0-0dba-4e94-a8b7-26d7b5c0e87f-kube-api-access-xzdbv\") on node \"crc\" DevicePath \"\"" Nov 21 14:30:56 crc kubenswrapper[4897]: I1121 14:30:56.178566 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ksmjv\" (UniqueName: \"kubernetes.io/projected/a69b6579-ce5a-44d7-ab3d-5454dd5f85df-kube-api-access-ksmjv\") on node \"crc\" DevicePath \"\"" Nov 21 14:30:56 crc kubenswrapper[4897]: I1121 14:30:56.178575 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pzsg2\" (UniqueName: \"kubernetes.io/projected/ee8a4380-421f-45b1-ba19-9832b69935fe-kube-api-access-pzsg2\") on node \"crc\" DevicePath \"\"" Nov 21 14:30:56 crc kubenswrapper[4897]: I1121 14:30:56.178585 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sf4jq\" (UniqueName: \"kubernetes.io/projected/1fd99443-7153-4711-9575-6ebc4fbfcf42-kube-api-access-sf4jq\") on node \"crc\" DevicePath \"\"" Nov 21 14:30:56 crc kubenswrapper[4897]: I1121 14:30:56.178596 4897 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1fd99443-7153-4711-9575-6ebc4fbfcf42-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 21 14:30:56 crc kubenswrapper[4897]: I1121 14:30:56.178605 4897 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ee8a4380-421f-45b1-ba19-9832b69935fe-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 21 14:30:56 crc kubenswrapper[4897]: I1121 14:30:56.178614 4897 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a69b6579-ce5a-44d7-ab3d-5454dd5f85df-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 21 14:30:56 crc kubenswrapper[4897]: I1121 14:30:56.178621 4897 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/56c583f0-0dba-4e94-a8b7-26d7b5c0e87f-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 21 14:30:56 crc kubenswrapper[4897]: I1121 14:30:56.178629 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-th28m\" (UniqueName: \"kubernetes.io/projected/25643e36-c5eb-468b-85ce-e85cff1ad090-kube-api-access-th28m\") on node \"crc\" DevicePath \"\"" Nov 21 14:30:56 crc kubenswrapper[4897]: I1121 14:30:56.611731 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-a766-account-create-btlq4" event={"ID":"25643e36-c5eb-468b-85ce-e85cff1ad090","Type":"ContainerDied","Data":"6e28d7e2796437459c660202dd15e5b41b2473a09bb2412c98c55cc5d74c5d1b"} Nov 21 14:30:56 crc kubenswrapper[4897]: I1121 14:30:56.611781 4897 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6e28d7e2796437459c660202dd15e5b41b2473a09bb2412c98c55cc5d74c5d1b" Nov 21 14:30:56 crc kubenswrapper[4897]: I1121 14:30:56.611870 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-a766-account-create-btlq4" Nov 21 14:30:56 crc kubenswrapper[4897]: I1121 14:30:56.617836 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-c4c2-account-create-kzvn8" event={"ID":"56c583f0-0dba-4e94-a8b7-26d7b5c0e87f","Type":"ContainerDied","Data":"1090d2af284c51892bcb7fa7a6a0de257271ebb6659e0178eaca286f47a34b7f"} Nov 21 14:30:56 crc kubenswrapper[4897]: I1121 14:30:56.617905 4897 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1090d2af284c51892bcb7fa7a6a0de257271ebb6659e0178eaca286f47a34b7f" Nov 21 14:30:56 crc kubenswrapper[4897]: I1121 14:30:56.617904 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-c4c2-account-create-kzvn8" Nov 21 14:30:56 crc kubenswrapper[4897]: I1121 14:30:56.619805 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-f2a9-account-create-rx44g" Nov 21 14:30:56 crc kubenswrapper[4897]: I1121 14:30:56.619797 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-f2a9-account-create-rx44g" event={"ID":"1fd99443-7153-4711-9575-6ebc4fbfcf42","Type":"ContainerDied","Data":"1cb5d006c2eb51c868e0945688cad3f46e74fef8b259852eb7f281589a14a76b"} Nov 21 14:30:56 crc kubenswrapper[4897]: I1121 14:30:56.619960 4897 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1cb5d006c2eb51c868e0945688cad3f46e74fef8b259852eb7f281589a14a76b" Nov 21 14:30:56 crc kubenswrapper[4897]: I1121 14:30:56.622128 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-5f6b-account-create-29rlc" event={"ID":"a69b6579-ce5a-44d7-ab3d-5454dd5f85df","Type":"ContainerDied","Data":"028b8a3f8f8da9052c2a1e2d67fb9be783942b649225fd9497474f2985080dd4"} Nov 21 14:30:56 crc kubenswrapper[4897]: I1121 14:30:56.622170 4897 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="028b8a3f8f8da9052c2a1e2d67fb9be783942b649225fd9497474f2985080dd4" Nov 21 14:30:56 crc kubenswrapper[4897]: I1121 14:30:56.622241 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-5f6b-account-create-29rlc" Nov 21 14:30:56 crc kubenswrapper[4897]: I1121 14:30:56.625257 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-cgbsk" event={"ID":"8bb39c02-70b6-470f-bf62-add1c2ebc02f","Type":"ContainerDied","Data":"00c88f9203968f749543de4dbe8afae7e77dc45520c2698a5026f4e9226730fb"} Nov 21 14:30:56 crc kubenswrapper[4897]: I1121 14:30:56.625298 4897 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="00c88f9203968f749543de4dbe8afae7e77dc45520c2698a5026f4e9226730fb" Nov 21 14:30:56 crc kubenswrapper[4897]: I1121 14:30:56.625302 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-cgbsk" Nov 21 14:30:56 crc kubenswrapper[4897]: I1121 14:30:56.627434 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-hrqhb" event={"ID":"982ba7b6-1d19-4bd4-bdea-c828fd678186","Type":"ContainerDied","Data":"1614aeb618cc32854fe6fd6a3661255a445d2ecf1db6ebac4a176700ce427a76"} Nov 21 14:30:56 crc kubenswrapper[4897]: I1121 14:30:56.627461 4897 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1614aeb618cc32854fe6fd6a3661255a445d2ecf1db6ebac4a176700ce427a76" Nov 21 14:30:56 crc kubenswrapper[4897]: I1121 14:30:56.627536 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-hrqhb" Nov 21 14:30:56 crc kubenswrapper[4897]: I1121 14:30:56.630590 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-hg4dh" event={"ID":"ee8a4380-421f-45b1-ba19-9832b69935fe","Type":"ContainerDied","Data":"35c08b523620f504553fdab9f8f75b551b43014d9ba7b0404b6e7f81ed83b244"} Nov 21 14:30:56 crc kubenswrapper[4897]: I1121 14:30:56.630624 4897 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="35c08b523620f504553fdab9f8f75b551b43014d9ba7b0404b6e7f81ed83b244" Nov 21 14:30:56 crc kubenswrapper[4897]: I1121 14:30:56.630689 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-hg4dh" Nov 21 14:30:56 crc kubenswrapper[4897]: I1121 14:30:56.634184 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-227rw" event={"ID":"208ac008-5901-4b46-983a-a73c9ed207f2","Type":"ContainerStarted","Data":"22185b922cd0f1f68e876cd66d956a126099e34ca5c3647d107d2adc78cb4383"} Nov 21 14:30:56 crc kubenswrapper[4897]: I1121 14:30:56.638191 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-dpnlm" Nov 21 14:30:56 crc kubenswrapper[4897]: I1121 14:30:56.638633 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-dpnlm" event={"ID":"67dd6a02-c8f6-4035-ad18-c6bbea13c688","Type":"ContainerDied","Data":"c7befc62974d1f8c10dcf4aa82eebb822773d3e6f69cce8bd59098b4a662a6d6"} Nov 21 14:30:56 crc kubenswrapper[4897]: I1121 14:30:56.638666 4897 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c7befc62974d1f8c10dcf4aa82eebb822773d3e6f69cce8bd59098b4a662a6d6" Nov 21 14:30:56 crc kubenswrapper[4897]: I1121 14:30:56.652935 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-227rw" podStartSLOduration=6.76253456 podStartE2EDuration="11.652918426s" podCreationTimestamp="2025-11-21 14:30:45 +0000 UTC" firstStartedPulling="2025-11-21 14:30:50.811728256 +0000 UTC m=+1328.096321761" lastFinishedPulling="2025-11-21 14:30:55.702112162 +0000 UTC m=+1332.986705627" observedRunningTime="2025-11-21 14:30:56.652059703 +0000 UTC m=+1333.936653188" watchObservedRunningTime="2025-11-21 14:30:56.652918426 +0000 UTC m=+1333.937511901" Nov 21 14:30:58 crc kubenswrapper[4897]: I1121 14:30:58.059242 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6d5b6d6b67-qc7fs" Nov 21 14:30:58 crc kubenswrapper[4897]: I1121 14:30:58.117568 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-7kmrw"] Nov 21 14:30:58 crc kubenswrapper[4897]: I1121 14:30:58.117837 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-b8fbc5445-7kmrw" podUID="7a31e75c-6e03-40e3-927f-6b4e04c1efee" containerName="dnsmasq-dns" containerID="cri-o://3e1bff1bbd0769c15d6aa8951ca8e1ab60f58ea1d63e68a36263aec05a1a078e" gracePeriod=10 Nov 21 14:30:58 crc kubenswrapper[4897]: I1121 14:30:58.672775 4897 generic.go:334] "Generic (PLEG): container finished" podID="7a31e75c-6e03-40e3-927f-6b4e04c1efee" containerID="3e1bff1bbd0769c15d6aa8951ca8e1ab60f58ea1d63e68a36263aec05a1a078e" exitCode=0 Nov 21 14:30:58 crc kubenswrapper[4897]: I1121 14:30:58.672841 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-7kmrw" event={"ID":"7a31e75c-6e03-40e3-927f-6b4e04c1efee","Type":"ContainerDied","Data":"3e1bff1bbd0769c15d6aa8951ca8e1ab60f58ea1d63e68a36263aec05a1a078e"} Nov 21 14:30:58 crc kubenswrapper[4897]: I1121 14:30:58.672879 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-7kmrw" event={"ID":"7a31e75c-6e03-40e3-927f-6b4e04c1efee","Type":"ContainerDied","Data":"fe1b324f77ac15b5c51888eefead22f6b6c637ac576efe5d64e306f14244f92e"} Nov 21 14:30:58 crc kubenswrapper[4897]: I1121 14:30:58.672902 4897 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fe1b324f77ac15b5c51888eefead22f6b6c637ac576efe5d64e306f14244f92e" Nov 21 14:30:58 crc kubenswrapper[4897]: I1121 14:30:58.684220 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-7kmrw" Nov 21 14:30:58 crc kubenswrapper[4897]: I1121 14:30:58.842536 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7a31e75c-6e03-40e3-927f-6b4e04c1efee-ovsdbserver-nb\") pod \"7a31e75c-6e03-40e3-927f-6b4e04c1efee\" (UID: \"7a31e75c-6e03-40e3-927f-6b4e04c1efee\") " Nov 21 14:30:58 crc kubenswrapper[4897]: I1121 14:30:58.842614 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5vxc9\" (UniqueName: \"kubernetes.io/projected/7a31e75c-6e03-40e3-927f-6b4e04c1efee-kube-api-access-5vxc9\") pod \"7a31e75c-6e03-40e3-927f-6b4e04c1efee\" (UID: \"7a31e75c-6e03-40e3-927f-6b4e04c1efee\") " Nov 21 14:30:58 crc kubenswrapper[4897]: I1121 14:30:58.842688 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a31e75c-6e03-40e3-927f-6b4e04c1efee-config\") pod \"7a31e75c-6e03-40e3-927f-6b4e04c1efee\" (UID: \"7a31e75c-6e03-40e3-927f-6b4e04c1efee\") " Nov 21 14:30:58 crc kubenswrapper[4897]: I1121 14:30:58.842746 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7a31e75c-6e03-40e3-927f-6b4e04c1efee-dns-svc\") pod \"7a31e75c-6e03-40e3-927f-6b4e04c1efee\" (UID: \"7a31e75c-6e03-40e3-927f-6b4e04c1efee\") " Nov 21 14:30:58 crc kubenswrapper[4897]: I1121 14:30:58.842781 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7a31e75c-6e03-40e3-927f-6b4e04c1efee-ovsdbserver-sb\") pod \"7a31e75c-6e03-40e3-927f-6b4e04c1efee\" (UID: \"7a31e75c-6e03-40e3-927f-6b4e04c1efee\") " Nov 21 14:30:58 crc kubenswrapper[4897]: I1121 14:30:58.850490 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a31e75c-6e03-40e3-927f-6b4e04c1efee-kube-api-access-5vxc9" (OuterVolumeSpecName: "kube-api-access-5vxc9") pod "7a31e75c-6e03-40e3-927f-6b4e04c1efee" (UID: "7a31e75c-6e03-40e3-927f-6b4e04c1efee"). InnerVolumeSpecName "kube-api-access-5vxc9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:30:58 crc kubenswrapper[4897]: I1121 14:30:58.905566 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7a31e75c-6e03-40e3-927f-6b4e04c1efee-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7a31e75c-6e03-40e3-927f-6b4e04c1efee" (UID: "7a31e75c-6e03-40e3-927f-6b4e04c1efee"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:30:58 crc kubenswrapper[4897]: I1121 14:30:58.912285 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7a31e75c-6e03-40e3-927f-6b4e04c1efee-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7a31e75c-6e03-40e3-927f-6b4e04c1efee" (UID: "7a31e75c-6e03-40e3-927f-6b4e04c1efee"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:30:58 crc kubenswrapper[4897]: I1121 14:30:58.918615 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7a31e75c-6e03-40e3-927f-6b4e04c1efee-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7a31e75c-6e03-40e3-927f-6b4e04c1efee" (UID: "7a31e75c-6e03-40e3-927f-6b4e04c1efee"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:30:58 crc kubenswrapper[4897]: I1121 14:30:58.930935 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7a31e75c-6e03-40e3-927f-6b4e04c1efee-config" (OuterVolumeSpecName: "config") pod "7a31e75c-6e03-40e3-927f-6b4e04c1efee" (UID: "7a31e75c-6e03-40e3-927f-6b4e04c1efee"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:30:58 crc kubenswrapper[4897]: I1121 14:30:58.944676 4897 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7a31e75c-6e03-40e3-927f-6b4e04c1efee-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 21 14:30:58 crc kubenswrapper[4897]: I1121 14:30:58.944704 4897 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7a31e75c-6e03-40e3-927f-6b4e04c1efee-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 21 14:30:58 crc kubenswrapper[4897]: I1121 14:30:58.944717 4897 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7a31e75c-6e03-40e3-927f-6b4e04c1efee-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 21 14:30:58 crc kubenswrapper[4897]: I1121 14:30:58.944727 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5vxc9\" (UniqueName: \"kubernetes.io/projected/7a31e75c-6e03-40e3-927f-6b4e04c1efee-kube-api-access-5vxc9\") on node \"crc\" DevicePath \"\"" Nov 21 14:30:58 crc kubenswrapper[4897]: I1121 14:30:58.944738 4897 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a31e75c-6e03-40e3-927f-6b4e04c1efee-config\") on node \"crc\" DevicePath \"\"" Nov 21 14:30:59 crc kubenswrapper[4897]: I1121 14:30:59.692496 4897 generic.go:334] "Generic (PLEG): container finished" podID="208ac008-5901-4b46-983a-a73c9ed207f2" containerID="22185b922cd0f1f68e876cd66d956a126099e34ca5c3647d107d2adc78cb4383" exitCode=0 Nov 21 14:30:59 crc kubenswrapper[4897]: I1121 14:30:59.692574 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-227rw" event={"ID":"208ac008-5901-4b46-983a-a73c9ed207f2","Type":"ContainerDied","Data":"22185b922cd0f1f68e876cd66d956a126099e34ca5c3647d107d2adc78cb4383"} Nov 21 14:30:59 crc kubenswrapper[4897]: I1121 14:30:59.693274 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-7kmrw" Nov 21 14:30:59 crc kubenswrapper[4897]: I1121 14:30:59.762376 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-7kmrw"] Nov 21 14:30:59 crc kubenswrapper[4897]: I1121 14:30:59.773933 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-7kmrw"] Nov 21 14:31:00 crc kubenswrapper[4897]: I1121 14:31:00.119052 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a31e75c-6e03-40e3-927f-6b4e04c1efee" path="/var/lib/kubelet/pods/7a31e75c-6e03-40e3-927f-6b4e04c1efee/volumes" Nov 21 14:31:01 crc kubenswrapper[4897]: I1121 14:31:01.092133 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-227rw" Nov 21 14:31:01 crc kubenswrapper[4897]: I1121 14:31:01.199155 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/208ac008-5901-4b46-983a-a73c9ed207f2-combined-ca-bundle\") pod \"208ac008-5901-4b46-983a-a73c9ed207f2\" (UID: \"208ac008-5901-4b46-983a-a73c9ed207f2\") " Nov 21 14:31:01 crc kubenswrapper[4897]: I1121 14:31:01.199749 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vd667\" (UniqueName: \"kubernetes.io/projected/208ac008-5901-4b46-983a-a73c9ed207f2-kube-api-access-vd667\") pod \"208ac008-5901-4b46-983a-a73c9ed207f2\" (UID: \"208ac008-5901-4b46-983a-a73c9ed207f2\") " Nov 21 14:31:01 crc kubenswrapper[4897]: I1121 14:31:01.200014 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/208ac008-5901-4b46-983a-a73c9ed207f2-config-data\") pod \"208ac008-5901-4b46-983a-a73c9ed207f2\" (UID: \"208ac008-5901-4b46-983a-a73c9ed207f2\") " Nov 21 14:31:01 crc kubenswrapper[4897]: I1121 14:31:01.207724 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/208ac008-5901-4b46-983a-a73c9ed207f2-kube-api-access-vd667" (OuterVolumeSpecName: "kube-api-access-vd667") pod "208ac008-5901-4b46-983a-a73c9ed207f2" (UID: "208ac008-5901-4b46-983a-a73c9ed207f2"). InnerVolumeSpecName "kube-api-access-vd667". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:31:01 crc kubenswrapper[4897]: I1121 14:31:01.238411 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/208ac008-5901-4b46-983a-a73c9ed207f2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "208ac008-5901-4b46-983a-a73c9ed207f2" (UID: "208ac008-5901-4b46-983a-a73c9ed207f2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:31:01 crc kubenswrapper[4897]: I1121 14:31:01.269952 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/208ac008-5901-4b46-983a-a73c9ed207f2-config-data" (OuterVolumeSpecName: "config-data") pod "208ac008-5901-4b46-983a-a73c9ed207f2" (UID: "208ac008-5901-4b46-983a-a73c9ed207f2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:31:01 crc kubenswrapper[4897]: I1121 14:31:01.302396 4897 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/208ac008-5901-4b46-983a-a73c9ed207f2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 21 14:31:01 crc kubenswrapper[4897]: I1121 14:31:01.302433 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vd667\" (UniqueName: \"kubernetes.io/projected/208ac008-5901-4b46-983a-a73c9ed207f2-kube-api-access-vd667\") on node \"crc\" DevicePath \"\"" Nov 21 14:31:01 crc kubenswrapper[4897]: I1121 14:31:01.302447 4897 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/208ac008-5901-4b46-983a-a73c9ed207f2-config-data\") on node \"crc\" DevicePath \"\"" Nov 21 14:31:01 crc kubenswrapper[4897]: I1121 14:31:01.714028 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-zz6kx" event={"ID":"9c08d8b3-d052-47b7-ad42-7b0f9fbe1920","Type":"ContainerStarted","Data":"f0b5d22717e8bf9ca37f66f5b7cbefcb84e12420967595ff9e141eebb59aaa33"} Nov 21 14:31:01 crc kubenswrapper[4897]: I1121 14:31:01.718387 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-227rw" event={"ID":"208ac008-5901-4b46-983a-a73c9ed207f2","Type":"ContainerDied","Data":"114beb74200ad93f12e7ed339efdc21758f06e09563638274a1a6ac884304735"} Nov 21 14:31:01 crc kubenswrapper[4897]: I1121 14:31:01.718431 4897 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="114beb74200ad93f12e7ed339efdc21758f06e09563638274a1a6ac884304735" Nov 21 14:31:01 crc kubenswrapper[4897]: I1121 14:31:01.718468 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-227rw" Nov 21 14:31:01 crc kubenswrapper[4897]: I1121 14:31:01.744136 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-zz6kx" podStartSLOduration=2.947491673 podStartE2EDuration="33.744110478s" podCreationTimestamp="2025-11-21 14:30:28 +0000 UTC" firstStartedPulling="2025-11-21 14:30:29.760672918 +0000 UTC m=+1307.045266393" lastFinishedPulling="2025-11-21 14:31:00.557291713 +0000 UTC m=+1337.841885198" observedRunningTime="2025-11-21 14:31:01.744062287 +0000 UTC m=+1339.028655772" watchObservedRunningTime="2025-11-21 14:31:01.744110478 +0000 UTC m=+1339.028703953" Nov 21 14:31:01 crc kubenswrapper[4897]: I1121 14:31:01.973291 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6f8c45789f-lgbrf"] Nov 21 14:31:01 crc kubenswrapper[4897]: E1121 14:31:01.973963 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25643e36-c5eb-468b-85ce-e85cff1ad090" containerName="mariadb-account-create" Nov 21 14:31:01 crc kubenswrapper[4897]: I1121 14:31:01.973981 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="25643e36-c5eb-468b-85ce-e85cff1ad090" containerName="mariadb-account-create" Nov 21 14:31:01 crc kubenswrapper[4897]: E1121 14:31:01.973997 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="982ba7b6-1d19-4bd4-bdea-c828fd678186" containerName="mariadb-database-create" Nov 21 14:31:01 crc kubenswrapper[4897]: I1121 14:31:01.974004 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="982ba7b6-1d19-4bd4-bdea-c828fd678186" containerName="mariadb-database-create" Nov 21 14:31:01 crc kubenswrapper[4897]: E1121 14:31:01.974031 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56c583f0-0dba-4e94-a8b7-26d7b5c0e87f" containerName="mariadb-account-create" Nov 21 14:31:01 crc kubenswrapper[4897]: I1121 14:31:01.974037 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="56c583f0-0dba-4e94-a8b7-26d7b5c0e87f" containerName="mariadb-account-create" Nov 21 14:31:01 crc kubenswrapper[4897]: E1121 14:31:01.974047 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a31e75c-6e03-40e3-927f-6b4e04c1efee" containerName="init" Nov 21 14:31:01 crc kubenswrapper[4897]: I1121 14:31:01.974052 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a31e75c-6e03-40e3-927f-6b4e04c1efee" containerName="init" Nov 21 14:31:01 crc kubenswrapper[4897]: E1121 14:31:01.974064 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a31e75c-6e03-40e3-927f-6b4e04c1efee" containerName="dnsmasq-dns" Nov 21 14:31:01 crc kubenswrapper[4897]: I1121 14:31:01.974070 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a31e75c-6e03-40e3-927f-6b4e04c1efee" containerName="dnsmasq-dns" Nov 21 14:31:01 crc kubenswrapper[4897]: E1121 14:31:01.974082 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee8a4380-421f-45b1-ba19-9832b69935fe" containerName="mariadb-database-create" Nov 21 14:31:01 crc kubenswrapper[4897]: I1121 14:31:01.974088 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee8a4380-421f-45b1-ba19-9832b69935fe" containerName="mariadb-database-create" Nov 21 14:31:01 crc kubenswrapper[4897]: E1121 14:31:01.974103 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bb39c02-70b6-470f-bf62-add1c2ebc02f" containerName="mariadb-database-create" Nov 21 14:31:01 crc kubenswrapper[4897]: I1121 14:31:01.974109 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bb39c02-70b6-470f-bf62-add1c2ebc02f" containerName="mariadb-database-create" Nov 21 14:31:01 crc kubenswrapper[4897]: E1121 14:31:01.974121 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a69b6579-ce5a-44d7-ab3d-5454dd5f85df" containerName="mariadb-account-create" Nov 21 14:31:01 crc kubenswrapper[4897]: I1121 14:31:01.974126 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="a69b6579-ce5a-44d7-ab3d-5454dd5f85df" containerName="mariadb-account-create" Nov 21 14:31:01 crc kubenswrapper[4897]: E1121 14:31:01.974159 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67dd6a02-c8f6-4035-ad18-c6bbea13c688" containerName="mariadb-database-create" Nov 21 14:31:01 crc kubenswrapper[4897]: I1121 14:31:01.974166 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="67dd6a02-c8f6-4035-ad18-c6bbea13c688" containerName="mariadb-database-create" Nov 21 14:31:01 crc kubenswrapper[4897]: E1121 14:31:01.974176 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="208ac008-5901-4b46-983a-a73c9ed207f2" containerName="keystone-db-sync" Nov 21 14:31:01 crc kubenswrapper[4897]: I1121 14:31:01.974183 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="208ac008-5901-4b46-983a-a73c9ed207f2" containerName="keystone-db-sync" Nov 21 14:31:01 crc kubenswrapper[4897]: E1121 14:31:01.974192 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1fd99443-7153-4711-9575-6ebc4fbfcf42" containerName="mariadb-account-create" Nov 21 14:31:01 crc kubenswrapper[4897]: I1121 14:31:01.974199 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="1fd99443-7153-4711-9575-6ebc4fbfcf42" containerName="mariadb-account-create" Nov 21 14:31:01 crc kubenswrapper[4897]: I1121 14:31:01.974383 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a31e75c-6e03-40e3-927f-6b4e04c1efee" containerName="dnsmasq-dns" Nov 21 14:31:01 crc kubenswrapper[4897]: I1121 14:31:01.974403 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="982ba7b6-1d19-4bd4-bdea-c828fd678186" containerName="mariadb-database-create" Nov 21 14:31:01 crc kubenswrapper[4897]: I1121 14:31:01.974416 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="208ac008-5901-4b46-983a-a73c9ed207f2" containerName="keystone-db-sync" Nov 21 14:31:01 crc kubenswrapper[4897]: I1121 14:31:01.974424 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee8a4380-421f-45b1-ba19-9832b69935fe" containerName="mariadb-database-create" Nov 21 14:31:01 crc kubenswrapper[4897]: I1121 14:31:01.974433 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="56c583f0-0dba-4e94-a8b7-26d7b5c0e87f" containerName="mariadb-account-create" Nov 21 14:31:01 crc kubenswrapper[4897]: I1121 14:31:01.974442 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="67dd6a02-c8f6-4035-ad18-c6bbea13c688" containerName="mariadb-database-create" Nov 21 14:31:01 crc kubenswrapper[4897]: I1121 14:31:01.974449 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="1fd99443-7153-4711-9575-6ebc4fbfcf42" containerName="mariadb-account-create" Nov 21 14:31:01 crc kubenswrapper[4897]: I1121 14:31:01.974460 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="8bb39c02-70b6-470f-bf62-add1c2ebc02f" containerName="mariadb-database-create" Nov 21 14:31:01 crc kubenswrapper[4897]: I1121 14:31:01.974471 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="a69b6579-ce5a-44d7-ab3d-5454dd5f85df" containerName="mariadb-account-create" Nov 21 14:31:01 crc kubenswrapper[4897]: I1121 14:31:01.974481 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="25643e36-c5eb-468b-85ce-e85cff1ad090" containerName="mariadb-account-create" Nov 21 14:31:01 crc kubenswrapper[4897]: I1121 14:31:01.980200 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f8c45789f-lgbrf" Nov 21 14:31:01 crc kubenswrapper[4897]: I1121 14:31:01.987119 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6f8c45789f-lgbrf"] Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.057278 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-mrvrw"] Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.059057 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-mrvrw" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.062946 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.063113 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.063218 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.063332 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.063454 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-wbfsb" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.071381 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-mrvrw"] Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.145016 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6b7418e5-f3f0-46a8-9c0d-b84b922763be-dns-swift-storage-0\") pod \"dnsmasq-dns-6f8c45789f-lgbrf\" (UID: \"6b7418e5-f3f0-46a8-9c0d-b84b922763be\") " pod="openstack/dnsmasq-dns-6f8c45789f-lgbrf" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.145073 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6b7418e5-f3f0-46a8-9c0d-b84b922763be-ovsdbserver-nb\") pod \"dnsmasq-dns-6f8c45789f-lgbrf\" (UID: \"6b7418e5-f3f0-46a8-9c0d-b84b922763be\") " pod="openstack/dnsmasq-dns-6f8c45789f-lgbrf" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.145099 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6b7418e5-f3f0-46a8-9c0d-b84b922763be-dns-svc\") pod \"dnsmasq-dns-6f8c45789f-lgbrf\" (UID: \"6b7418e5-f3f0-46a8-9c0d-b84b922763be\") " pod="openstack/dnsmasq-dns-6f8c45789f-lgbrf" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.145187 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b7418e5-f3f0-46a8-9c0d-b84b922763be-config\") pod \"dnsmasq-dns-6f8c45789f-lgbrf\" (UID: \"6b7418e5-f3f0-46a8-9c0d-b84b922763be\") " pod="openstack/dnsmasq-dns-6f8c45789f-lgbrf" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.145416 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cpfg6\" (UniqueName: \"kubernetes.io/projected/6b7418e5-f3f0-46a8-9c0d-b84b922763be-kube-api-access-cpfg6\") pod \"dnsmasq-dns-6f8c45789f-lgbrf\" (UID: \"6b7418e5-f3f0-46a8-9c0d-b84b922763be\") " pod="openstack/dnsmasq-dns-6f8c45789f-lgbrf" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.145473 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6b7418e5-f3f0-46a8-9c0d-b84b922763be-ovsdbserver-sb\") pod \"dnsmasq-dns-6f8c45789f-lgbrf\" (UID: \"6b7418e5-f3f0-46a8-9c0d-b84b922763be\") " pod="openstack/dnsmasq-dns-6f8c45789f-lgbrf" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.194003 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-sync-tbtcb"] Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.195445 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-tbtcb" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.199395 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.199414 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-tkmk9" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.204289 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-tbtcb"] Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.247153 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6b7418e5-f3f0-46a8-9c0d-b84b922763be-dns-swift-storage-0\") pod \"dnsmasq-dns-6f8c45789f-lgbrf\" (UID: \"6b7418e5-f3f0-46a8-9c0d-b84b922763be\") " pod="openstack/dnsmasq-dns-6f8c45789f-lgbrf" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.247208 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6b7418e5-f3f0-46a8-9c0d-b84b922763be-ovsdbserver-nb\") pod \"dnsmasq-dns-6f8c45789f-lgbrf\" (UID: \"6b7418e5-f3f0-46a8-9c0d-b84b922763be\") " pod="openstack/dnsmasq-dns-6f8c45789f-lgbrf" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.247232 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72a6a702-fda5-4a79-8a0b-10b9cae6d9c7-combined-ca-bundle\") pod \"keystone-bootstrap-mrvrw\" (UID: \"72a6a702-fda5-4a79-8a0b-10b9cae6d9c7\") " pod="openstack/keystone-bootstrap-mrvrw" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.247250 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6b7418e5-f3f0-46a8-9c0d-b84b922763be-dns-svc\") pod \"dnsmasq-dns-6f8c45789f-lgbrf\" (UID: \"6b7418e5-f3f0-46a8-9c0d-b84b922763be\") " pod="openstack/dnsmasq-dns-6f8c45789f-lgbrf" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.247348 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b7418e5-f3f0-46a8-9c0d-b84b922763be-config\") pod \"dnsmasq-dns-6f8c45789f-lgbrf\" (UID: \"6b7418e5-f3f0-46a8-9c0d-b84b922763be\") " pod="openstack/dnsmasq-dns-6f8c45789f-lgbrf" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.247398 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/72a6a702-fda5-4a79-8a0b-10b9cae6d9c7-scripts\") pod \"keystone-bootstrap-mrvrw\" (UID: \"72a6a702-fda5-4a79-8a0b-10b9cae6d9c7\") " pod="openstack/keystone-bootstrap-mrvrw" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.247440 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cpfg6\" (UniqueName: \"kubernetes.io/projected/6b7418e5-f3f0-46a8-9c0d-b84b922763be-kube-api-access-cpfg6\") pod \"dnsmasq-dns-6f8c45789f-lgbrf\" (UID: \"6b7418e5-f3f0-46a8-9c0d-b84b922763be\") " pod="openstack/dnsmasq-dns-6f8c45789f-lgbrf" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.247459 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72a6a702-fda5-4a79-8a0b-10b9cae6d9c7-config-data\") pod \"keystone-bootstrap-mrvrw\" (UID: \"72a6a702-fda5-4a79-8a0b-10b9cae6d9c7\") " pod="openstack/keystone-bootstrap-mrvrw" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.247476 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/72a6a702-fda5-4a79-8a0b-10b9cae6d9c7-credential-keys\") pod \"keystone-bootstrap-mrvrw\" (UID: \"72a6a702-fda5-4a79-8a0b-10b9cae6d9c7\") " pod="openstack/keystone-bootstrap-mrvrw" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.247495 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-chhx7\" (UniqueName: \"kubernetes.io/projected/72a6a702-fda5-4a79-8a0b-10b9cae6d9c7-kube-api-access-chhx7\") pod \"keystone-bootstrap-mrvrw\" (UID: \"72a6a702-fda5-4a79-8a0b-10b9cae6d9c7\") " pod="openstack/keystone-bootstrap-mrvrw" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.247542 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6b7418e5-f3f0-46a8-9c0d-b84b922763be-ovsdbserver-sb\") pod \"dnsmasq-dns-6f8c45789f-lgbrf\" (UID: \"6b7418e5-f3f0-46a8-9c0d-b84b922763be\") " pod="openstack/dnsmasq-dns-6f8c45789f-lgbrf" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.247653 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/72a6a702-fda5-4a79-8a0b-10b9cae6d9c7-fernet-keys\") pod \"keystone-bootstrap-mrvrw\" (UID: \"72a6a702-fda5-4a79-8a0b-10b9cae6d9c7\") " pod="openstack/keystone-bootstrap-mrvrw" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.248451 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b7418e5-f3f0-46a8-9c0d-b84b922763be-config\") pod \"dnsmasq-dns-6f8c45789f-lgbrf\" (UID: \"6b7418e5-f3f0-46a8-9c0d-b84b922763be\") " pod="openstack/dnsmasq-dns-6f8c45789f-lgbrf" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.249628 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6b7418e5-f3f0-46a8-9c0d-b84b922763be-ovsdbserver-sb\") pod \"dnsmasq-dns-6f8c45789f-lgbrf\" (UID: \"6b7418e5-f3f0-46a8-9c0d-b84b922763be\") " pod="openstack/dnsmasq-dns-6f8c45789f-lgbrf" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.250223 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6b7418e5-f3f0-46a8-9c0d-b84b922763be-dns-swift-storage-0\") pod \"dnsmasq-dns-6f8c45789f-lgbrf\" (UID: \"6b7418e5-f3f0-46a8-9c0d-b84b922763be\") " pod="openstack/dnsmasq-dns-6f8c45789f-lgbrf" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.250799 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6b7418e5-f3f0-46a8-9c0d-b84b922763be-ovsdbserver-nb\") pod \"dnsmasq-dns-6f8c45789f-lgbrf\" (UID: \"6b7418e5-f3f0-46a8-9c0d-b84b922763be\") " pod="openstack/dnsmasq-dns-6f8c45789f-lgbrf" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.251431 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6b7418e5-f3f0-46a8-9c0d-b84b922763be-dns-svc\") pod \"dnsmasq-dns-6f8c45789f-lgbrf\" (UID: \"6b7418e5-f3f0-46a8-9c0d-b84b922763be\") " pod="openstack/dnsmasq-dns-6f8c45789f-lgbrf" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.264936 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-mrctr"] Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.275837 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-mrctr" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.280910 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.281139 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.281686 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-lddw4" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.283001 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-mrctr"] Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.291071 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cpfg6\" (UniqueName: \"kubernetes.io/projected/6b7418e5-f3f0-46a8-9c0d-b84b922763be-kube-api-access-cpfg6\") pod \"dnsmasq-dns-6f8c45789f-lgbrf\" (UID: \"6b7418e5-f3f0-46a8-9c0d-b84b922763be\") " pod="openstack/dnsmasq-dns-6f8c45789f-lgbrf" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.324173 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f8c45789f-lgbrf" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.351683 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72a6a702-fda5-4a79-8a0b-10b9cae6d9c7-config-data\") pod \"keystone-bootstrap-mrvrw\" (UID: \"72a6a702-fda5-4a79-8a0b-10b9cae6d9c7\") " pod="openstack/keystone-bootstrap-mrvrw" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.351722 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/72a6a702-fda5-4a79-8a0b-10b9cae6d9c7-credential-keys\") pod \"keystone-bootstrap-mrvrw\" (UID: \"72a6a702-fda5-4a79-8a0b-10b9cae6d9c7\") " pod="openstack/keystone-bootstrap-mrvrw" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.351739 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-chhx7\" (UniqueName: \"kubernetes.io/projected/72a6a702-fda5-4a79-8a0b-10b9cae6d9c7-kube-api-access-chhx7\") pod \"keystone-bootstrap-mrvrw\" (UID: \"72a6a702-fda5-4a79-8a0b-10b9cae6d9c7\") " pod="openstack/keystone-bootstrap-mrvrw" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.351782 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59fd52b7-f25b-4a81-a961-41e6cb526c89-combined-ca-bundle\") pod \"heat-db-sync-tbtcb\" (UID: \"59fd52b7-f25b-4a81-a961-41e6cb526c89\") " pod="openstack/heat-db-sync-tbtcb" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.351833 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p4csc\" (UniqueName: \"kubernetes.io/projected/59fd52b7-f25b-4a81-a961-41e6cb526c89-kube-api-access-p4csc\") pod \"heat-db-sync-tbtcb\" (UID: \"59fd52b7-f25b-4a81-a961-41e6cb526c89\") " pod="openstack/heat-db-sync-tbtcb" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.351865 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/72a6a702-fda5-4a79-8a0b-10b9cae6d9c7-fernet-keys\") pod \"keystone-bootstrap-mrvrw\" (UID: \"72a6a702-fda5-4a79-8a0b-10b9cae6d9c7\") " pod="openstack/keystone-bootstrap-mrvrw" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.351898 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72a6a702-fda5-4a79-8a0b-10b9cae6d9c7-combined-ca-bundle\") pod \"keystone-bootstrap-mrvrw\" (UID: \"72a6a702-fda5-4a79-8a0b-10b9cae6d9c7\") " pod="openstack/keystone-bootstrap-mrvrw" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.351927 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59fd52b7-f25b-4a81-a961-41e6cb526c89-config-data\") pod \"heat-db-sync-tbtcb\" (UID: \"59fd52b7-f25b-4a81-a961-41e6cb526c89\") " pod="openstack/heat-db-sync-tbtcb" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.351983 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/72a6a702-fda5-4a79-8a0b-10b9cae6d9c7-scripts\") pod \"keystone-bootstrap-mrvrw\" (UID: \"72a6a702-fda5-4a79-8a0b-10b9cae6d9c7\") " pod="openstack/keystone-bootstrap-mrvrw" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.368875 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72a6a702-fda5-4a79-8a0b-10b9cae6d9c7-combined-ca-bundle\") pod \"keystone-bootstrap-mrvrw\" (UID: \"72a6a702-fda5-4a79-8a0b-10b9cae6d9c7\") " pod="openstack/keystone-bootstrap-mrvrw" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.369524 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/72a6a702-fda5-4a79-8a0b-10b9cae6d9c7-scripts\") pod \"keystone-bootstrap-mrvrw\" (UID: \"72a6a702-fda5-4a79-8a0b-10b9cae6d9c7\") " pod="openstack/keystone-bootstrap-mrvrw" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.371148 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-dm94g"] Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.384012 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/72a6a702-fda5-4a79-8a0b-10b9cae6d9c7-credential-keys\") pod \"keystone-bootstrap-mrvrw\" (UID: \"72a6a702-fda5-4a79-8a0b-10b9cae6d9c7\") " pod="openstack/keystone-bootstrap-mrvrw" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.384551 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72a6a702-fda5-4a79-8a0b-10b9cae6d9c7-config-data\") pod \"keystone-bootstrap-mrvrw\" (UID: \"72a6a702-fda5-4a79-8a0b-10b9cae6d9c7\") " pod="openstack/keystone-bootstrap-mrvrw" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.385229 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/72a6a702-fda5-4a79-8a0b-10b9cae6d9c7-fernet-keys\") pod \"keystone-bootstrap-mrvrw\" (UID: \"72a6a702-fda5-4a79-8a0b-10b9cae6d9c7\") " pod="openstack/keystone-bootstrap-mrvrw" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.393903 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-dm94g" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.398952 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-8p2dh" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.399164 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.402378 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-chhx7\" (UniqueName: \"kubernetes.io/projected/72a6a702-fda5-4a79-8a0b-10b9cae6d9c7-kube-api-access-chhx7\") pod \"keystone-bootstrap-mrvrw\" (UID: \"72a6a702-fda5-4a79-8a0b-10b9cae6d9c7\") " pod="openstack/keystone-bootstrap-mrvrw" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.433765 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-dm94g"] Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.457825 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59fd52b7-f25b-4a81-a961-41e6cb526c89-combined-ca-bundle\") pod \"heat-db-sync-tbtcb\" (UID: \"59fd52b7-f25b-4a81-a961-41e6cb526c89\") " pod="openstack/heat-db-sync-tbtcb" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.457987 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/4a683589-a940-405b-afe1-1838f34dac9a-config\") pod \"neutron-db-sync-mrctr\" (UID: \"4a683589-a940-405b-afe1-1838f34dac9a\") " pod="openstack/neutron-db-sync-mrctr" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.458010 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p4csc\" (UniqueName: \"kubernetes.io/projected/59fd52b7-f25b-4a81-a961-41e6cb526c89-kube-api-access-p4csc\") pod \"heat-db-sync-tbtcb\" (UID: \"59fd52b7-f25b-4a81-a961-41e6cb526c89\") " pod="openstack/heat-db-sync-tbtcb" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.458036 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a683589-a940-405b-afe1-1838f34dac9a-combined-ca-bundle\") pod \"neutron-db-sync-mrctr\" (UID: \"4a683589-a940-405b-afe1-1838f34dac9a\") " pod="openstack/neutron-db-sync-mrctr" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.458200 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59fd52b7-f25b-4a81-a961-41e6cb526c89-config-data\") pod \"heat-db-sync-tbtcb\" (UID: \"59fd52b7-f25b-4a81-a961-41e6cb526c89\") " pod="openstack/heat-db-sync-tbtcb" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.458392 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t7rdp\" (UniqueName: \"kubernetes.io/projected/4a683589-a940-405b-afe1-1838f34dac9a-kube-api-access-t7rdp\") pod \"neutron-db-sync-mrctr\" (UID: \"4a683589-a940-405b-afe1-1838f34dac9a\") " pod="openstack/neutron-db-sync-mrctr" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.463982 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59fd52b7-f25b-4a81-a961-41e6cb526c89-config-data\") pod \"heat-db-sync-tbtcb\" (UID: \"59fd52b7-f25b-4a81-a961-41e6cb526c89\") " pod="openstack/heat-db-sync-tbtcb" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.464668 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-xj48w"] Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.473416 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-xj48w" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.484014 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.484198 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-fzrth" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.484357 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.484876 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59fd52b7-f25b-4a81-a961-41e6cb526c89-combined-ca-bundle\") pod \"heat-db-sync-tbtcb\" (UID: \"59fd52b7-f25b-4a81-a961-41e6cb526c89\") " pod="openstack/heat-db-sync-tbtcb" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.500917 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p4csc\" (UniqueName: \"kubernetes.io/projected/59fd52b7-f25b-4a81-a961-41e6cb526c89-kube-api-access-p4csc\") pod \"heat-db-sync-tbtcb\" (UID: \"59fd52b7-f25b-4a81-a961-41e6cb526c89\") " pod="openstack/heat-db-sync-tbtcb" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.505566 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-7dvt9"] Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.507084 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-7dvt9" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.512237 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.512445 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.512609 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-pr8nf" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.513058 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-tbtcb" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.546207 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6f8c45789f-lgbrf"] Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.560762 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/4a683589-a940-405b-afe1-1838f34dac9a-config\") pod \"neutron-db-sync-mrctr\" (UID: \"4a683589-a940-405b-afe1-1838f34dac9a\") " pod="openstack/neutron-db-sync-mrctr" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.561245 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a683589-a940-405b-afe1-1838f34dac9a-combined-ca-bundle\") pod \"neutron-db-sync-mrctr\" (UID: \"4a683589-a940-405b-afe1-1838f34dac9a\") " pod="openstack/neutron-db-sync-mrctr" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.561441 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rn26s\" (UniqueName: \"kubernetes.io/projected/de087e28-b2d8-4686-b758-f4cd38443f25-kube-api-access-rn26s\") pod \"barbican-db-sync-dm94g\" (UID: \"de087e28-b2d8-4686-b758-f4cd38443f25\") " pod="openstack/barbican-db-sync-dm94g" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.561593 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/de087e28-b2d8-4686-b758-f4cd38443f25-db-sync-config-data\") pod \"barbican-db-sync-dm94g\" (UID: \"de087e28-b2d8-4686-b758-f4cd38443f25\") " pod="openstack/barbican-db-sync-dm94g" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.561716 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t7rdp\" (UniqueName: \"kubernetes.io/projected/4a683589-a940-405b-afe1-1838f34dac9a-kube-api-access-t7rdp\") pod \"neutron-db-sync-mrctr\" (UID: \"4a683589-a940-405b-afe1-1838f34dac9a\") " pod="openstack/neutron-db-sync-mrctr" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.561833 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de087e28-b2d8-4686-b758-f4cd38443f25-combined-ca-bundle\") pod \"barbican-db-sync-dm94g\" (UID: \"de087e28-b2d8-4686-b758-f4cd38443f25\") " pod="openstack/barbican-db-sync-dm94g" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.565446 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a683589-a940-405b-afe1-1838f34dac9a-combined-ca-bundle\") pod \"neutron-db-sync-mrctr\" (UID: \"4a683589-a940-405b-afe1-1838f34dac9a\") " pod="openstack/neutron-db-sync-mrctr" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.565939 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/4a683589-a940-405b-afe1-1838f34dac9a-config\") pod \"neutron-db-sync-mrctr\" (UID: \"4a683589-a940-405b-afe1-1838f34dac9a\") " pod="openstack/neutron-db-sync-mrctr" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.582927 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-xj48w"] Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.593187 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t7rdp\" (UniqueName: \"kubernetes.io/projected/4a683589-a940-405b-afe1-1838f34dac9a-kube-api-access-t7rdp\") pod \"neutron-db-sync-mrctr\" (UID: \"4a683589-a940-405b-afe1-1838f34dac9a\") " pod="openstack/neutron-db-sync-mrctr" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.605193 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-7dvt9"] Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.664351 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rn26s\" (UniqueName: \"kubernetes.io/projected/de087e28-b2d8-4686-b758-f4cd38443f25-kube-api-access-rn26s\") pod \"barbican-db-sync-dm94g\" (UID: \"de087e28-b2d8-4686-b758-f4cd38443f25\") " pod="openstack/barbican-db-sync-dm94g" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.664438 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/55644d8b-b8a3-463e-8045-b7391e9c2419-etc-machine-id\") pod \"cinder-db-sync-7dvt9\" (UID: \"55644d8b-b8a3-463e-8045-b7391e9c2419\") " pod="openstack/cinder-db-sync-7dvt9" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.664477 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/de087e28-b2d8-4686-b758-f4cd38443f25-db-sync-config-data\") pod \"barbican-db-sync-dm94g\" (UID: \"de087e28-b2d8-4686-b758-f4cd38443f25\") " pod="openstack/barbican-db-sync-dm94g" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.664521 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55644d8b-b8a3-463e-8045-b7391e9c2419-config-data\") pod \"cinder-db-sync-7dvt9\" (UID: \"55644d8b-b8a3-463e-8045-b7391e9c2419\") " pod="openstack/cinder-db-sync-7dvt9" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.664595 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55644d8b-b8a3-463e-8045-b7391e9c2419-combined-ca-bundle\") pod \"cinder-db-sync-7dvt9\" (UID: \"55644d8b-b8a3-463e-8045-b7391e9c2419\") " pod="openstack/cinder-db-sync-7dvt9" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.664655 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b0c8cf90-d5b4-43b0-9037-4b579f26c74b-logs\") pod \"placement-db-sync-xj48w\" (UID: \"b0c8cf90-d5b4-43b0-9037-4b579f26c74b\") " pod="openstack/placement-db-sync-xj48w" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.664674 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/55644d8b-b8a3-463e-8045-b7391e9c2419-scripts\") pod \"cinder-db-sync-7dvt9\" (UID: \"55644d8b-b8a3-463e-8045-b7391e9c2419\") " pod="openstack/cinder-db-sync-7dvt9" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.664708 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/55644d8b-b8a3-463e-8045-b7391e9c2419-db-sync-config-data\") pod \"cinder-db-sync-7dvt9\" (UID: \"55644d8b-b8a3-463e-8045-b7391e9c2419\") " pod="openstack/cinder-db-sync-7dvt9" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.664747 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dzppc\" (UniqueName: \"kubernetes.io/projected/55644d8b-b8a3-463e-8045-b7391e9c2419-kube-api-access-dzppc\") pod \"cinder-db-sync-7dvt9\" (UID: \"55644d8b-b8a3-463e-8045-b7391e9c2419\") " pod="openstack/cinder-db-sync-7dvt9" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.664800 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0c8cf90-d5b4-43b0-9037-4b579f26c74b-combined-ca-bundle\") pod \"placement-db-sync-xj48w\" (UID: \"b0c8cf90-d5b4-43b0-9037-4b579f26c74b\") " pod="openstack/placement-db-sync-xj48w" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.664836 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b0c8cf90-d5b4-43b0-9037-4b579f26c74b-scripts\") pod \"placement-db-sync-xj48w\" (UID: \"b0c8cf90-d5b4-43b0-9037-4b579f26c74b\") " pod="openstack/placement-db-sync-xj48w" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.664867 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7wxh\" (UniqueName: \"kubernetes.io/projected/b0c8cf90-d5b4-43b0-9037-4b579f26c74b-kube-api-access-v7wxh\") pod \"placement-db-sync-xj48w\" (UID: \"b0c8cf90-d5b4-43b0-9037-4b579f26c74b\") " pod="openstack/placement-db-sync-xj48w" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.664922 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de087e28-b2d8-4686-b758-f4cd38443f25-combined-ca-bundle\") pod \"barbican-db-sync-dm94g\" (UID: \"de087e28-b2d8-4686-b758-f4cd38443f25\") " pod="openstack/barbican-db-sync-dm94g" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.665788 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0c8cf90-d5b4-43b0-9037-4b579f26c74b-config-data\") pod \"placement-db-sync-xj48w\" (UID: \"b0c8cf90-d5b4-43b0-9037-4b579f26c74b\") " pod="openstack/placement-db-sync-xj48w" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.677277 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-fcfdd6f9f-6l4ws"] Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.677549 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/de087e28-b2d8-4686-b758-f4cd38443f25-db-sync-config-data\") pod \"barbican-db-sync-dm94g\" (UID: \"de087e28-b2d8-4686-b758-f4cd38443f25\") " pod="openstack/barbican-db-sync-dm94g" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.678060 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-mrvrw" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.689394 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de087e28-b2d8-4686-b758-f4cd38443f25-combined-ca-bundle\") pod \"barbican-db-sync-dm94g\" (UID: \"de087e28-b2d8-4686-b758-f4cd38443f25\") " pod="openstack/barbican-db-sync-dm94g" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.690681 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-fcfdd6f9f-6l4ws"] Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.690798 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-fcfdd6f9f-6l4ws" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.693335 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rn26s\" (UniqueName: \"kubernetes.io/projected/de087e28-b2d8-4686-b758-f4cd38443f25-kube-api-access-rn26s\") pod \"barbican-db-sync-dm94g\" (UID: \"de087e28-b2d8-4686-b758-f4cd38443f25\") " pod="openstack/barbican-db-sync-dm94g" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.765952 4897 generic.go:334] "Generic (PLEG): container finished" podID="fbc040d4-3c4c-4b07-8fd4-1eb61db44e5d" containerID="81c07f5180c87605c31b38047cf5d4bc886014b4de5781764c005b00293f6cde" exitCode=0 Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.766032 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"fbc040d4-3c4c-4b07-8fd4-1eb61db44e5d","Type":"ContainerDied","Data":"81c07f5180c87605c31b38047cf5d4bc886014b4de5781764c005b00293f6cde"} Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.767306 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0c8cf90-d5b4-43b0-9037-4b579f26c74b-config-data\") pod \"placement-db-sync-xj48w\" (UID: \"b0c8cf90-d5b4-43b0-9037-4b579f26c74b\") " pod="openstack/placement-db-sync-xj48w" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.767400 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/55644d8b-b8a3-463e-8045-b7391e9c2419-etc-machine-id\") pod \"cinder-db-sync-7dvt9\" (UID: \"55644d8b-b8a3-463e-8045-b7391e9c2419\") " pod="openstack/cinder-db-sync-7dvt9" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.767430 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55644d8b-b8a3-463e-8045-b7391e9c2419-config-data\") pod \"cinder-db-sync-7dvt9\" (UID: \"55644d8b-b8a3-463e-8045-b7391e9c2419\") " pod="openstack/cinder-db-sync-7dvt9" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.767458 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55644d8b-b8a3-463e-8045-b7391e9c2419-combined-ca-bundle\") pod \"cinder-db-sync-7dvt9\" (UID: \"55644d8b-b8a3-463e-8045-b7391e9c2419\") " pod="openstack/cinder-db-sync-7dvt9" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.767512 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b0c8cf90-d5b4-43b0-9037-4b579f26c74b-logs\") pod \"placement-db-sync-xj48w\" (UID: \"b0c8cf90-d5b4-43b0-9037-4b579f26c74b\") " pod="openstack/placement-db-sync-xj48w" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.767529 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/55644d8b-b8a3-463e-8045-b7391e9c2419-scripts\") pod \"cinder-db-sync-7dvt9\" (UID: \"55644d8b-b8a3-463e-8045-b7391e9c2419\") " pod="openstack/cinder-db-sync-7dvt9" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.767549 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/55644d8b-b8a3-463e-8045-b7391e9c2419-db-sync-config-data\") pod \"cinder-db-sync-7dvt9\" (UID: \"55644d8b-b8a3-463e-8045-b7391e9c2419\") " pod="openstack/cinder-db-sync-7dvt9" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.767581 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dzppc\" (UniqueName: \"kubernetes.io/projected/55644d8b-b8a3-463e-8045-b7391e9c2419-kube-api-access-dzppc\") pod \"cinder-db-sync-7dvt9\" (UID: \"55644d8b-b8a3-463e-8045-b7391e9c2419\") " pod="openstack/cinder-db-sync-7dvt9" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.767621 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0c8cf90-d5b4-43b0-9037-4b579f26c74b-combined-ca-bundle\") pod \"placement-db-sync-xj48w\" (UID: \"b0c8cf90-d5b4-43b0-9037-4b579f26c74b\") " pod="openstack/placement-db-sync-xj48w" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.767644 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b0c8cf90-d5b4-43b0-9037-4b579f26c74b-scripts\") pod \"placement-db-sync-xj48w\" (UID: \"b0c8cf90-d5b4-43b0-9037-4b579f26c74b\") " pod="openstack/placement-db-sync-xj48w" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.767667 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v7wxh\" (UniqueName: \"kubernetes.io/projected/b0c8cf90-d5b4-43b0-9037-4b579f26c74b-kube-api-access-v7wxh\") pod \"placement-db-sync-xj48w\" (UID: \"b0c8cf90-d5b4-43b0-9037-4b579f26c74b\") " pod="openstack/placement-db-sync-xj48w" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.768062 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b0c8cf90-d5b4-43b0-9037-4b579f26c74b-logs\") pod \"placement-db-sync-xj48w\" (UID: \"b0c8cf90-d5b4-43b0-9037-4b579f26c74b\") " pod="openstack/placement-db-sync-xj48w" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.769853 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/55644d8b-b8a3-463e-8045-b7391e9c2419-etc-machine-id\") pod \"cinder-db-sync-7dvt9\" (UID: \"55644d8b-b8a3-463e-8045-b7391e9c2419\") " pod="openstack/cinder-db-sync-7dvt9" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.783269 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0c8cf90-d5b4-43b0-9037-4b579f26c74b-combined-ca-bundle\") pod \"placement-db-sync-xj48w\" (UID: \"b0c8cf90-d5b4-43b0-9037-4b579f26c74b\") " pod="openstack/placement-db-sync-xj48w" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.783644 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55644d8b-b8a3-463e-8045-b7391e9c2419-config-data\") pod \"cinder-db-sync-7dvt9\" (UID: \"55644d8b-b8a3-463e-8045-b7391e9c2419\") " pod="openstack/cinder-db-sync-7dvt9" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.784024 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55644d8b-b8a3-463e-8045-b7391e9c2419-combined-ca-bundle\") pod \"cinder-db-sync-7dvt9\" (UID: \"55644d8b-b8a3-463e-8045-b7391e9c2419\") " pod="openstack/cinder-db-sync-7dvt9" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.784416 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/55644d8b-b8a3-463e-8045-b7391e9c2419-scripts\") pod \"cinder-db-sync-7dvt9\" (UID: \"55644d8b-b8a3-463e-8045-b7391e9c2419\") " pod="openstack/cinder-db-sync-7dvt9" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.788397 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0c8cf90-d5b4-43b0-9037-4b579f26c74b-config-data\") pod \"placement-db-sync-xj48w\" (UID: \"b0c8cf90-d5b4-43b0-9037-4b579f26c74b\") " pod="openstack/placement-db-sync-xj48w" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.790647 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b0c8cf90-d5b4-43b0-9037-4b579f26c74b-scripts\") pod \"placement-db-sync-xj48w\" (UID: \"b0c8cf90-d5b4-43b0-9037-4b579f26c74b\") " pod="openstack/placement-db-sync-xj48w" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.794762 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dzppc\" (UniqueName: \"kubernetes.io/projected/55644d8b-b8a3-463e-8045-b7391e9c2419-kube-api-access-dzppc\") pod \"cinder-db-sync-7dvt9\" (UID: \"55644d8b-b8a3-463e-8045-b7391e9c2419\") " pod="openstack/cinder-db-sync-7dvt9" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.795324 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v7wxh\" (UniqueName: \"kubernetes.io/projected/b0c8cf90-d5b4-43b0-9037-4b579f26c74b-kube-api-access-v7wxh\") pod \"placement-db-sync-xj48w\" (UID: \"b0c8cf90-d5b4-43b0-9037-4b579f26c74b\") " pod="openstack/placement-db-sync-xj48w" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.811960 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/55644d8b-b8a3-463e-8045-b7391e9c2419-db-sync-config-data\") pod \"cinder-db-sync-7dvt9\" (UID: \"55644d8b-b8a3-463e-8045-b7391e9c2419\") " pod="openstack/cinder-db-sync-7dvt9" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.841044 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-mrctr" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.874221 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-dm94g" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.874794 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/11f3e22f-051b-42ab-b70e-7813ecbbc8f5-dns-swift-storage-0\") pod \"dnsmasq-dns-fcfdd6f9f-6l4ws\" (UID: \"11f3e22f-051b-42ab-b70e-7813ecbbc8f5\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-6l4ws" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.874865 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/11f3e22f-051b-42ab-b70e-7813ecbbc8f5-dns-svc\") pod \"dnsmasq-dns-fcfdd6f9f-6l4ws\" (UID: \"11f3e22f-051b-42ab-b70e-7813ecbbc8f5\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-6l4ws" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.875088 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/11f3e22f-051b-42ab-b70e-7813ecbbc8f5-ovsdbserver-sb\") pod \"dnsmasq-dns-fcfdd6f9f-6l4ws\" (UID: \"11f3e22f-051b-42ab-b70e-7813ecbbc8f5\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-6l4ws" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.875312 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11f3e22f-051b-42ab-b70e-7813ecbbc8f5-config\") pod \"dnsmasq-dns-fcfdd6f9f-6l4ws\" (UID: \"11f3e22f-051b-42ab-b70e-7813ecbbc8f5\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-6l4ws" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.875482 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mv89f\" (UniqueName: \"kubernetes.io/projected/11f3e22f-051b-42ab-b70e-7813ecbbc8f5-kube-api-access-mv89f\") pod \"dnsmasq-dns-fcfdd6f9f-6l4ws\" (UID: \"11f3e22f-051b-42ab-b70e-7813ecbbc8f5\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-6l4ws" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.875542 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/11f3e22f-051b-42ab-b70e-7813ecbbc8f5-ovsdbserver-nb\") pod \"dnsmasq-dns-fcfdd6f9f-6l4ws\" (UID: \"11f3e22f-051b-42ab-b70e-7813ecbbc8f5\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-6l4ws" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.884810 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.893528 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.896365 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-xj48w" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.897650 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.897690 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.900411 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.912246 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-7dvt9" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.981857 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3696f1d-db0b-4910-a65c-31410e2316df-config-data\") pod \"ceilometer-0\" (UID: \"c3696f1d-db0b-4910-a65c-31410e2316df\") " pod="openstack/ceilometer-0" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.981961 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c3696f1d-db0b-4910-a65c-31410e2316df-run-httpd\") pod \"ceilometer-0\" (UID: \"c3696f1d-db0b-4910-a65c-31410e2316df\") " pod="openstack/ceilometer-0" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.982013 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c3696f1d-db0b-4910-a65c-31410e2316df-log-httpd\") pod \"ceilometer-0\" (UID: \"c3696f1d-db0b-4910-a65c-31410e2316df\") " pod="openstack/ceilometer-0" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.982077 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mv89f\" (UniqueName: \"kubernetes.io/projected/11f3e22f-051b-42ab-b70e-7813ecbbc8f5-kube-api-access-mv89f\") pod \"dnsmasq-dns-fcfdd6f9f-6l4ws\" (UID: \"11f3e22f-051b-42ab-b70e-7813ecbbc8f5\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-6l4ws" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.982111 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/11f3e22f-051b-42ab-b70e-7813ecbbc8f5-ovsdbserver-nb\") pod \"dnsmasq-dns-fcfdd6f9f-6l4ws\" (UID: \"11f3e22f-051b-42ab-b70e-7813ecbbc8f5\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-6l4ws" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.982185 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lx66x\" (UniqueName: \"kubernetes.io/projected/c3696f1d-db0b-4910-a65c-31410e2316df-kube-api-access-lx66x\") pod \"ceilometer-0\" (UID: \"c3696f1d-db0b-4910-a65c-31410e2316df\") " pod="openstack/ceilometer-0" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.982262 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c3696f1d-db0b-4910-a65c-31410e2316df-scripts\") pod \"ceilometer-0\" (UID: \"c3696f1d-db0b-4910-a65c-31410e2316df\") " pod="openstack/ceilometer-0" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.982313 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/11f3e22f-051b-42ab-b70e-7813ecbbc8f5-dns-swift-storage-0\") pod \"dnsmasq-dns-fcfdd6f9f-6l4ws\" (UID: \"11f3e22f-051b-42ab-b70e-7813ecbbc8f5\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-6l4ws" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.982364 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/11f3e22f-051b-42ab-b70e-7813ecbbc8f5-dns-svc\") pod \"dnsmasq-dns-fcfdd6f9f-6l4ws\" (UID: \"11f3e22f-051b-42ab-b70e-7813ecbbc8f5\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-6l4ws" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.982534 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c3696f1d-db0b-4910-a65c-31410e2316df-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c3696f1d-db0b-4910-a65c-31410e2316df\") " pod="openstack/ceilometer-0" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.982592 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/11f3e22f-051b-42ab-b70e-7813ecbbc8f5-ovsdbserver-sb\") pod \"dnsmasq-dns-fcfdd6f9f-6l4ws\" (UID: \"11f3e22f-051b-42ab-b70e-7813ecbbc8f5\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-6l4ws" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.982620 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3696f1d-db0b-4910-a65c-31410e2316df-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c3696f1d-db0b-4910-a65c-31410e2316df\") " pod="openstack/ceilometer-0" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.982784 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11f3e22f-051b-42ab-b70e-7813ecbbc8f5-config\") pod \"dnsmasq-dns-fcfdd6f9f-6l4ws\" (UID: \"11f3e22f-051b-42ab-b70e-7813ecbbc8f5\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-6l4ws" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.988352 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/11f3e22f-051b-42ab-b70e-7813ecbbc8f5-dns-swift-storage-0\") pod \"dnsmasq-dns-fcfdd6f9f-6l4ws\" (UID: \"11f3e22f-051b-42ab-b70e-7813ecbbc8f5\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-6l4ws" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.989410 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/11f3e22f-051b-42ab-b70e-7813ecbbc8f5-ovsdbserver-sb\") pod \"dnsmasq-dns-fcfdd6f9f-6l4ws\" (UID: \"11f3e22f-051b-42ab-b70e-7813ecbbc8f5\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-6l4ws" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.990649 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/11f3e22f-051b-42ab-b70e-7813ecbbc8f5-ovsdbserver-nb\") pod \"dnsmasq-dns-fcfdd6f9f-6l4ws\" (UID: \"11f3e22f-051b-42ab-b70e-7813ecbbc8f5\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-6l4ws" Nov 21 14:31:02 crc kubenswrapper[4897]: I1121 14:31:02.994816 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/11f3e22f-051b-42ab-b70e-7813ecbbc8f5-dns-svc\") pod \"dnsmasq-dns-fcfdd6f9f-6l4ws\" (UID: \"11f3e22f-051b-42ab-b70e-7813ecbbc8f5\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-6l4ws" Nov 21 14:31:03 crc kubenswrapper[4897]: I1121 14:31:03.025925 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11f3e22f-051b-42ab-b70e-7813ecbbc8f5-config\") pod \"dnsmasq-dns-fcfdd6f9f-6l4ws\" (UID: \"11f3e22f-051b-42ab-b70e-7813ecbbc8f5\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-6l4ws" Nov 21 14:31:03 crc kubenswrapper[4897]: I1121 14:31:03.036376 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mv89f\" (UniqueName: \"kubernetes.io/projected/11f3e22f-051b-42ab-b70e-7813ecbbc8f5-kube-api-access-mv89f\") pod \"dnsmasq-dns-fcfdd6f9f-6l4ws\" (UID: \"11f3e22f-051b-42ab-b70e-7813ecbbc8f5\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-6l4ws" Nov 21 14:31:03 crc kubenswrapper[4897]: I1121 14:31:03.086727 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lx66x\" (UniqueName: \"kubernetes.io/projected/c3696f1d-db0b-4910-a65c-31410e2316df-kube-api-access-lx66x\") pod \"ceilometer-0\" (UID: \"c3696f1d-db0b-4910-a65c-31410e2316df\") " pod="openstack/ceilometer-0" Nov 21 14:31:03 crc kubenswrapper[4897]: I1121 14:31:03.086812 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c3696f1d-db0b-4910-a65c-31410e2316df-scripts\") pod \"ceilometer-0\" (UID: \"c3696f1d-db0b-4910-a65c-31410e2316df\") " pod="openstack/ceilometer-0" Nov 21 14:31:03 crc kubenswrapper[4897]: I1121 14:31:03.086930 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c3696f1d-db0b-4910-a65c-31410e2316df-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c3696f1d-db0b-4910-a65c-31410e2316df\") " pod="openstack/ceilometer-0" Nov 21 14:31:03 crc kubenswrapper[4897]: I1121 14:31:03.086985 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3696f1d-db0b-4910-a65c-31410e2316df-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c3696f1d-db0b-4910-a65c-31410e2316df\") " pod="openstack/ceilometer-0" Nov 21 14:31:03 crc kubenswrapper[4897]: I1121 14:31:03.092682 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3696f1d-db0b-4910-a65c-31410e2316df-config-data\") pod \"ceilometer-0\" (UID: \"c3696f1d-db0b-4910-a65c-31410e2316df\") " pod="openstack/ceilometer-0" Nov 21 14:31:03 crc kubenswrapper[4897]: I1121 14:31:03.100019 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c3696f1d-db0b-4910-a65c-31410e2316df-run-httpd\") pod \"ceilometer-0\" (UID: \"c3696f1d-db0b-4910-a65c-31410e2316df\") " pod="openstack/ceilometer-0" Nov 21 14:31:03 crc kubenswrapper[4897]: I1121 14:31:03.100170 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c3696f1d-db0b-4910-a65c-31410e2316df-log-httpd\") pod \"ceilometer-0\" (UID: \"c3696f1d-db0b-4910-a65c-31410e2316df\") " pod="openstack/ceilometer-0" Nov 21 14:31:03 crc kubenswrapper[4897]: I1121 14:31:03.130898 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c3696f1d-db0b-4910-a65c-31410e2316df-log-httpd\") pod \"ceilometer-0\" (UID: \"c3696f1d-db0b-4910-a65c-31410e2316df\") " pod="openstack/ceilometer-0" Nov 21 14:31:03 crc kubenswrapper[4897]: I1121 14:31:03.132326 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6f8c45789f-lgbrf"] Nov 21 14:31:03 crc kubenswrapper[4897]: I1121 14:31:03.134198 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c3696f1d-db0b-4910-a65c-31410e2316df-run-httpd\") pod \"ceilometer-0\" (UID: \"c3696f1d-db0b-4910-a65c-31410e2316df\") " pod="openstack/ceilometer-0" Nov 21 14:31:03 crc kubenswrapper[4897]: I1121 14:31:03.138470 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c3696f1d-db0b-4910-a65c-31410e2316df-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c3696f1d-db0b-4910-a65c-31410e2316df\") " pod="openstack/ceilometer-0" Nov 21 14:31:03 crc kubenswrapper[4897]: I1121 14:31:03.139051 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c3696f1d-db0b-4910-a65c-31410e2316df-scripts\") pod \"ceilometer-0\" (UID: \"c3696f1d-db0b-4910-a65c-31410e2316df\") " pod="openstack/ceilometer-0" Nov 21 14:31:03 crc kubenswrapper[4897]: I1121 14:31:03.144214 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3696f1d-db0b-4910-a65c-31410e2316df-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c3696f1d-db0b-4910-a65c-31410e2316df\") " pod="openstack/ceilometer-0" Nov 21 14:31:03 crc kubenswrapper[4897]: I1121 14:31:03.147026 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lx66x\" (UniqueName: \"kubernetes.io/projected/c3696f1d-db0b-4910-a65c-31410e2316df-kube-api-access-lx66x\") pod \"ceilometer-0\" (UID: \"c3696f1d-db0b-4910-a65c-31410e2316df\") " pod="openstack/ceilometer-0" Nov 21 14:31:03 crc kubenswrapper[4897]: I1121 14:31:03.149346 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3696f1d-db0b-4910-a65c-31410e2316df-config-data\") pod \"ceilometer-0\" (UID: \"c3696f1d-db0b-4910-a65c-31410e2316df\") " pod="openstack/ceilometer-0" Nov 21 14:31:03 crc kubenswrapper[4897]: I1121 14:31:03.271083 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-fcfdd6f9f-6l4ws" Nov 21 14:31:03 crc kubenswrapper[4897]: I1121 14:31:03.280805 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 21 14:31:03 crc kubenswrapper[4897]: I1121 14:31:03.328997 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-tbtcb"] Nov 21 14:31:03 crc kubenswrapper[4897]: I1121 14:31:03.476961 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-mrvrw"] Nov 21 14:31:03 crc kubenswrapper[4897]: I1121 14:31:03.781391 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f8c45789f-lgbrf" event={"ID":"6b7418e5-f3f0-46a8-9c0d-b84b922763be","Type":"ContainerStarted","Data":"16baafb7e27aa21a06283dbd216bd65c8bbc65f24fdff08533e7b8dca53a5342"} Nov 21 14:31:03 crc kubenswrapper[4897]: I1121 14:31:03.784361 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-tbtcb" event={"ID":"59fd52b7-f25b-4a81-a961-41e6cb526c89","Type":"ContainerStarted","Data":"cc0f36880c4fc19ca7daede88714edef43c228c9382b7a3bd03a7c5b9e7808fa"} Nov 21 14:31:03 crc kubenswrapper[4897]: I1121 14:31:03.795012 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-mrvrw" event={"ID":"72a6a702-fda5-4a79-8a0b-10b9cae6d9c7","Type":"ContainerStarted","Data":"0a4c04fd3d6f3e822c208737b3d64d33d89ddd30ee98aeb96442367b0fbea2b5"} Nov 21 14:31:03 crc kubenswrapper[4897]: I1121 14:31:03.801671 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"fbc040d4-3c4c-4b07-8fd4-1eb61db44e5d","Type":"ContainerStarted","Data":"4351927c2bb2b5a4e6ba4e281b29c1ff3fea5381403d052e7e8989345fd87d51"} Nov 21 14:31:03 crc kubenswrapper[4897]: I1121 14:31:03.953044 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-mrctr"] Nov 21 14:31:04 crc kubenswrapper[4897]: I1121 14:31:04.006578 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-xj48w"] Nov 21 14:31:04 crc kubenswrapper[4897]: I1121 14:31:04.030479 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-dm94g"] Nov 21 14:31:04 crc kubenswrapper[4897]: I1121 14:31:04.066555 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-7dvt9"] Nov 21 14:31:04 crc kubenswrapper[4897]: I1121 14:31:04.200637 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 21 14:31:04 crc kubenswrapper[4897]: I1121 14:31:04.295643 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-fcfdd6f9f-6l4ws"] Nov 21 14:31:04 crc kubenswrapper[4897]: I1121 14:31:04.373100 4897 patch_prober.go:28] interesting pod/machine-config-daemon-krv5b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 21 14:31:04 crc kubenswrapper[4897]: I1121 14:31:04.373821 4897 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 21 14:31:04 crc kubenswrapper[4897]: I1121 14:31:04.833763 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-fcfdd6f9f-6l4ws" event={"ID":"11f3e22f-051b-42ab-b70e-7813ecbbc8f5","Type":"ContainerStarted","Data":"888737d2076501df7e2c323741a7989bc6466db72ef1a795ceedff5203a50639"} Nov 21 14:31:04 crc kubenswrapper[4897]: I1121 14:31:04.834023 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-fcfdd6f9f-6l4ws" event={"ID":"11f3e22f-051b-42ab-b70e-7813ecbbc8f5","Type":"ContainerStarted","Data":"5ad095be28aed48a8934c0083373f12f0f4b4bea7b99c14921bb6920995de6b8"} Nov 21 14:31:04 crc kubenswrapper[4897]: I1121 14:31:04.837405 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-xj48w" event={"ID":"b0c8cf90-d5b4-43b0-9037-4b579f26c74b","Type":"ContainerStarted","Data":"fa7d015e268d0c7617ac51bf6139eda75898180fc42a864c1868f3d23855691b"} Nov 21 14:31:04 crc kubenswrapper[4897]: I1121 14:31:04.841314 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-mrctr" event={"ID":"4a683589-a940-405b-afe1-1838f34dac9a","Type":"ContainerStarted","Data":"b3b649aa1cd6febd3c35ff713d0112d349fcbe8d37fb13176de414dfeca6b26c"} Nov 21 14:31:04 crc kubenswrapper[4897]: I1121 14:31:04.841344 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-mrctr" event={"ID":"4a683589-a940-405b-afe1-1838f34dac9a","Type":"ContainerStarted","Data":"350ee59368631067e4b119b860a3b0e5f8fb23b051990ae4884466a22966dd83"} Nov 21 14:31:04 crc kubenswrapper[4897]: I1121 14:31:04.846466 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-mrvrw" event={"ID":"72a6a702-fda5-4a79-8a0b-10b9cae6d9c7","Type":"ContainerStarted","Data":"c3dd08ed558f92e886e50c86f9564fa7a7936c723ad747d6a561c7c2c73adc0b"} Nov 21 14:31:04 crc kubenswrapper[4897]: I1121 14:31:04.855781 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-dm94g" event={"ID":"de087e28-b2d8-4686-b758-f4cd38443f25","Type":"ContainerStarted","Data":"58066df22ae21154428c28012cb0b305024745237e2c2d5ff301b4f066770a1d"} Nov 21 14:31:04 crc kubenswrapper[4897]: I1121 14:31:04.863945 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c3696f1d-db0b-4910-a65c-31410e2316df","Type":"ContainerStarted","Data":"f77bbf271263dc44422e9ab3ef764bb5d530727524987b2b69ac5f01e86fd312"} Nov 21 14:31:04 crc kubenswrapper[4897]: I1121 14:31:04.873362 4897 generic.go:334] "Generic (PLEG): container finished" podID="6b7418e5-f3f0-46a8-9c0d-b84b922763be" containerID="a1d0673b5427cb2e2dde4a309febc186f8f2c640bfcc09817694f3e05854ed66" exitCode=0 Nov 21 14:31:04 crc kubenswrapper[4897]: I1121 14:31:04.873784 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f8c45789f-lgbrf" event={"ID":"6b7418e5-f3f0-46a8-9c0d-b84b922763be","Type":"ContainerDied","Data":"a1d0673b5427cb2e2dde4a309febc186f8f2c640bfcc09817694f3e05854ed66"} Nov 21 14:31:04 crc kubenswrapper[4897]: I1121 14:31:04.878677 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-7dvt9" event={"ID":"55644d8b-b8a3-463e-8045-b7391e9c2419","Type":"ContainerStarted","Data":"33c7a02a10a9950715b35226e715a310a1019d82d1c2fdd396e29be1f520e9d4"} Nov 21 14:31:04 crc kubenswrapper[4897]: I1121 14:31:04.893032 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-mrvrw" podStartSLOduration=2.893013012 podStartE2EDuration="2.893013012s" podCreationTimestamp="2025-11-21 14:31:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:31:04.879439238 +0000 UTC m=+1342.164032713" watchObservedRunningTime="2025-11-21 14:31:04.893013012 +0000 UTC m=+1342.177606487" Nov 21 14:31:04 crc kubenswrapper[4897]: I1121 14:31:04.903865 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-mrctr" podStartSLOduration=2.903852023 podStartE2EDuration="2.903852023s" podCreationTimestamp="2025-11-21 14:31:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:31:04.902700662 +0000 UTC m=+1342.187294147" watchObservedRunningTime="2025-11-21 14:31:04.903852023 +0000 UTC m=+1342.188445498" Nov 21 14:31:05 crc kubenswrapper[4897]: I1121 14:31:05.368124 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 21 14:31:05 crc kubenswrapper[4897]: I1121 14:31:05.480001 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f8c45789f-lgbrf" Nov 21 14:31:05 crc kubenswrapper[4897]: I1121 14:31:05.589957 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6b7418e5-f3f0-46a8-9c0d-b84b922763be-ovsdbserver-sb\") pod \"6b7418e5-f3f0-46a8-9c0d-b84b922763be\" (UID: \"6b7418e5-f3f0-46a8-9c0d-b84b922763be\") " Nov 21 14:31:05 crc kubenswrapper[4897]: I1121 14:31:05.590415 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cpfg6\" (UniqueName: \"kubernetes.io/projected/6b7418e5-f3f0-46a8-9c0d-b84b922763be-kube-api-access-cpfg6\") pod \"6b7418e5-f3f0-46a8-9c0d-b84b922763be\" (UID: \"6b7418e5-f3f0-46a8-9c0d-b84b922763be\") " Nov 21 14:31:05 crc kubenswrapper[4897]: I1121 14:31:05.590519 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6b7418e5-f3f0-46a8-9c0d-b84b922763be-dns-svc\") pod \"6b7418e5-f3f0-46a8-9c0d-b84b922763be\" (UID: \"6b7418e5-f3f0-46a8-9c0d-b84b922763be\") " Nov 21 14:31:05 crc kubenswrapper[4897]: I1121 14:31:05.590614 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6b7418e5-f3f0-46a8-9c0d-b84b922763be-ovsdbserver-nb\") pod \"6b7418e5-f3f0-46a8-9c0d-b84b922763be\" (UID: \"6b7418e5-f3f0-46a8-9c0d-b84b922763be\") " Nov 21 14:31:05 crc kubenswrapper[4897]: I1121 14:31:05.590696 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6b7418e5-f3f0-46a8-9c0d-b84b922763be-dns-swift-storage-0\") pod \"6b7418e5-f3f0-46a8-9c0d-b84b922763be\" (UID: \"6b7418e5-f3f0-46a8-9c0d-b84b922763be\") " Nov 21 14:31:05 crc kubenswrapper[4897]: I1121 14:31:05.590747 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b7418e5-f3f0-46a8-9c0d-b84b922763be-config\") pod \"6b7418e5-f3f0-46a8-9c0d-b84b922763be\" (UID: \"6b7418e5-f3f0-46a8-9c0d-b84b922763be\") " Nov 21 14:31:05 crc kubenswrapper[4897]: I1121 14:31:05.678129 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6b7418e5-f3f0-46a8-9c0d-b84b922763be-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "6b7418e5-f3f0-46a8-9c0d-b84b922763be" (UID: "6b7418e5-f3f0-46a8-9c0d-b84b922763be"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:31:05 crc kubenswrapper[4897]: I1121 14:31:05.694860 4897 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6b7418e5-f3f0-46a8-9c0d-b84b922763be-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 21 14:31:05 crc kubenswrapper[4897]: I1121 14:31:05.717450 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b7418e5-f3f0-46a8-9c0d-b84b922763be-kube-api-access-cpfg6" (OuterVolumeSpecName: "kube-api-access-cpfg6") pod "6b7418e5-f3f0-46a8-9c0d-b84b922763be" (UID: "6b7418e5-f3f0-46a8-9c0d-b84b922763be"). InnerVolumeSpecName "kube-api-access-cpfg6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:31:05 crc kubenswrapper[4897]: I1121 14:31:05.729747 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6b7418e5-f3f0-46a8-9c0d-b84b922763be-config" (OuterVolumeSpecName: "config") pod "6b7418e5-f3f0-46a8-9c0d-b84b922763be" (UID: "6b7418e5-f3f0-46a8-9c0d-b84b922763be"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:31:05 crc kubenswrapper[4897]: I1121 14:31:05.797321 4897 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b7418e5-f3f0-46a8-9c0d-b84b922763be-config\") on node \"crc\" DevicePath \"\"" Nov 21 14:31:05 crc kubenswrapper[4897]: I1121 14:31:05.797352 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cpfg6\" (UniqueName: \"kubernetes.io/projected/6b7418e5-f3f0-46a8-9c0d-b84b922763be-kube-api-access-cpfg6\") on node \"crc\" DevicePath \"\"" Nov 21 14:31:05 crc kubenswrapper[4897]: I1121 14:31:05.824713 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6b7418e5-f3f0-46a8-9c0d-b84b922763be-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6b7418e5-f3f0-46a8-9c0d-b84b922763be" (UID: "6b7418e5-f3f0-46a8-9c0d-b84b922763be"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:31:05 crc kubenswrapper[4897]: I1121 14:31:05.899020 4897 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6b7418e5-f3f0-46a8-9c0d-b84b922763be-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 21 14:31:05 crc kubenswrapper[4897]: I1121 14:31:05.899407 4897 generic.go:334] "Generic (PLEG): container finished" podID="11f3e22f-051b-42ab-b70e-7813ecbbc8f5" containerID="888737d2076501df7e2c323741a7989bc6466db72ef1a795ceedff5203a50639" exitCode=0 Nov 21 14:31:05 crc kubenswrapper[4897]: I1121 14:31:05.899476 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-fcfdd6f9f-6l4ws" event={"ID":"11f3e22f-051b-42ab-b70e-7813ecbbc8f5","Type":"ContainerDied","Data":"888737d2076501df7e2c323741a7989bc6466db72ef1a795ceedff5203a50639"} Nov 21 14:31:05 crc kubenswrapper[4897]: I1121 14:31:05.901991 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f8c45789f-lgbrf" Nov 21 14:31:05 crc kubenswrapper[4897]: I1121 14:31:05.902041 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f8c45789f-lgbrf" event={"ID":"6b7418e5-f3f0-46a8-9c0d-b84b922763be","Type":"ContainerDied","Data":"16baafb7e27aa21a06283dbd216bd65c8bbc65f24fdff08533e7b8dca53a5342"} Nov 21 14:31:05 crc kubenswrapper[4897]: I1121 14:31:05.902080 4897 scope.go:117] "RemoveContainer" containerID="a1d0673b5427cb2e2dde4a309febc186f8f2c640bfcc09817694f3e05854ed66" Nov 21 14:31:05 crc kubenswrapper[4897]: I1121 14:31:05.945664 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6b7418e5-f3f0-46a8-9c0d-b84b922763be-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "6b7418e5-f3f0-46a8-9c0d-b84b922763be" (UID: "6b7418e5-f3f0-46a8-9c0d-b84b922763be"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:31:06 crc kubenswrapper[4897]: I1121 14:31:06.002238 4897 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6b7418e5-f3f0-46a8-9c0d-b84b922763be-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 21 14:31:06 crc kubenswrapper[4897]: I1121 14:31:06.433770 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6b7418e5-f3f0-46a8-9c0d-b84b922763be-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "6b7418e5-f3f0-46a8-9c0d-b84b922763be" (UID: "6b7418e5-f3f0-46a8-9c0d-b84b922763be"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:31:06 crc kubenswrapper[4897]: I1121 14:31:06.513545 4897 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6b7418e5-f3f0-46a8-9c0d-b84b922763be-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 21 14:31:06 crc kubenswrapper[4897]: I1121 14:31:06.580991 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6f8c45789f-lgbrf"] Nov 21 14:31:06 crc kubenswrapper[4897]: I1121 14:31:06.602827 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6f8c45789f-lgbrf"] Nov 21 14:31:07 crc kubenswrapper[4897]: I1121 14:31:07.959625 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-fcfdd6f9f-6l4ws" event={"ID":"11f3e22f-051b-42ab-b70e-7813ecbbc8f5","Type":"ContainerStarted","Data":"41766958aa12b4ccf753cc8c11168d6cd59f891bf4c5f073c6e21504d016ade5"} Nov 21 14:31:07 crc kubenswrapper[4897]: I1121 14:31:07.960055 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-fcfdd6f9f-6l4ws" Nov 21 14:31:07 crc kubenswrapper[4897]: I1121 14:31:07.965702 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"fbc040d4-3c4c-4b07-8fd4-1eb61db44e5d","Type":"ContainerStarted","Data":"67f5ac871108b0cc21ccda0a5d73684ea744fa745649c656bd5fd43c6b2d9c03"} Nov 21 14:31:07 crc kubenswrapper[4897]: I1121 14:31:07.965743 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"fbc040d4-3c4c-4b07-8fd4-1eb61db44e5d","Type":"ContainerStarted","Data":"8146639a369a4b00717cd116e1ed8a55fde2e6381b0acbc27def6e8a6f1e4cf2"} Nov 21 14:31:07 crc kubenswrapper[4897]: I1121 14:31:07.984396 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-fcfdd6f9f-6l4ws" podStartSLOduration=5.984379253 podStartE2EDuration="5.984379253s" podCreationTimestamp="2025-11-21 14:31:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:31:07.977856587 +0000 UTC m=+1345.262450062" watchObservedRunningTime="2025-11-21 14:31:07.984379253 +0000 UTC m=+1345.268972728" Nov 21 14:31:08 crc kubenswrapper[4897]: I1121 14:31:08.106015 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b7418e5-f3f0-46a8-9c0d-b84b922763be" path="/var/lib/kubelet/pods/6b7418e5-f3f0-46a8-9c0d-b84b922763be/volumes" Nov 21 14:31:09 crc kubenswrapper[4897]: I1121 14:31:09.011485 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=21.011465823 podStartE2EDuration="21.011465823s" podCreationTimestamp="2025-11-21 14:30:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:31:08.998228627 +0000 UTC m=+1346.282822102" watchObservedRunningTime="2025-11-21 14:31:09.011465823 +0000 UTC m=+1346.296059298" Nov 21 14:31:09 crc kubenswrapper[4897]: I1121 14:31:09.349868 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Nov 21 14:31:13 crc kubenswrapper[4897]: I1121 14:31:13.272654 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-fcfdd6f9f-6l4ws" Nov 21 14:31:13 crc kubenswrapper[4897]: I1121 14:31:13.358394 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d5b6d6b67-qc7fs"] Nov 21 14:31:13 crc kubenswrapper[4897]: I1121 14:31:13.358719 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6d5b6d6b67-qc7fs" podUID="17a6c4a6-4493-4531-8f5c-814c7333a6ae" containerName="dnsmasq-dns" containerID="cri-o://afe2abe49994bcaffbe92412ab26a0240d4102be177179d0e806cea09b185b65" gracePeriod=10 Nov 21 14:31:14 crc kubenswrapper[4897]: I1121 14:31:14.051963 4897 generic.go:334] "Generic (PLEG): container finished" podID="17a6c4a6-4493-4531-8f5c-814c7333a6ae" containerID="afe2abe49994bcaffbe92412ab26a0240d4102be177179d0e806cea09b185b65" exitCode=0 Nov 21 14:31:14 crc kubenswrapper[4897]: I1121 14:31:14.052003 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d5b6d6b67-qc7fs" event={"ID":"17a6c4a6-4493-4531-8f5c-814c7333a6ae","Type":"ContainerDied","Data":"afe2abe49994bcaffbe92412ab26a0240d4102be177179d0e806cea09b185b65"} Nov 21 14:31:19 crc kubenswrapper[4897]: I1121 14:31:19.350393 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Nov 21 14:31:19 crc kubenswrapper[4897]: I1121 14:31:19.357614 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Nov 21 14:31:20 crc kubenswrapper[4897]: I1121 14:31:20.134886 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Nov 21 14:31:23 crc kubenswrapper[4897]: I1121 14:31:23.058239 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6d5b6d6b67-qc7fs" podUID="17a6c4a6-4493-4531-8f5c-814c7333a6ae" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.164:5353: i/o timeout" Nov 21 14:31:28 crc kubenswrapper[4897]: I1121 14:31:28.059535 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6d5b6d6b67-qc7fs" podUID="17a6c4a6-4493-4531-8f5c-814c7333a6ae" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.164:5353: i/o timeout" Nov 21 14:31:30 crc kubenswrapper[4897]: I1121 14:31:30.227981 4897 generic.go:334] "Generic (PLEG): container finished" podID="72a6a702-fda5-4a79-8a0b-10b9cae6d9c7" containerID="c3dd08ed558f92e886e50c86f9564fa7a7936c723ad747d6a561c7c2c73adc0b" exitCode=0 Nov 21 14:31:30 crc kubenswrapper[4897]: I1121 14:31:30.228212 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-mrvrw" event={"ID":"72a6a702-fda5-4a79-8a0b-10b9cae6d9c7","Type":"ContainerDied","Data":"c3dd08ed558f92e886e50c86f9564fa7a7936c723ad747d6a561c7c2c73adc0b"} Nov 21 14:31:30 crc kubenswrapper[4897]: E1121 14:31:30.304392 4897 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified" Nov 21 14:31:30 crc kubenswrapper[4897]: E1121 14:31:30.304552 4897 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rn26s,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-dm94g_openstack(de087e28-b2d8-4686-b758-f4cd38443f25): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 21 14:31:30 crc kubenswrapper[4897]: E1121 14:31:30.305726 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-dm94g" podUID="de087e28-b2d8-4686-b758-f4cd38443f25" Nov 21 14:31:31 crc kubenswrapper[4897]: E1121 14:31:31.245768 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified\\\"\"" pod="openstack/barbican-db-sync-dm94g" podUID="de087e28-b2d8-4686-b758-f4cd38443f25" Nov 21 14:31:33 crc kubenswrapper[4897]: I1121 14:31:33.060642 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6d5b6d6b67-qc7fs" podUID="17a6c4a6-4493-4531-8f5c-814c7333a6ae" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.164:5353: i/o timeout" Nov 21 14:31:33 crc kubenswrapper[4897]: I1121 14:31:33.060951 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6d5b6d6b67-qc7fs" Nov 21 14:31:34 crc kubenswrapper[4897]: I1121 14:31:34.370673 4897 patch_prober.go:28] interesting pod/machine-config-daemon-krv5b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 21 14:31:34 crc kubenswrapper[4897]: I1121 14:31:34.371119 4897 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 21 14:31:34 crc kubenswrapper[4897]: I1121 14:31:34.371183 4897 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" Nov 21 14:31:34 crc kubenswrapper[4897]: I1121 14:31:34.372321 4897 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fbff5a4c19b1cfbe0c7e57eb05f0093aa3829fbabee820a190f916e0ce4fe6b9"} pod="openshift-machine-config-operator/machine-config-daemon-krv5b" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 21 14:31:34 crc kubenswrapper[4897]: I1121 14:31:34.372431 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" containerName="machine-config-daemon" containerID="cri-o://fbff5a4c19b1cfbe0c7e57eb05f0093aa3829fbabee820a190f916e0ce4fe6b9" gracePeriod=600 Nov 21 14:31:35 crc kubenswrapper[4897]: I1121 14:31:35.283948 4897 generic.go:334] "Generic (PLEG): container finished" podID="e7670227-d280-4847-b882-754429f56b0f" containerID="fbff5a4c19b1cfbe0c7e57eb05f0093aa3829fbabee820a190f916e0ce4fe6b9" exitCode=0 Nov 21 14:31:35 crc kubenswrapper[4897]: I1121 14:31:35.284043 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" event={"ID":"e7670227-d280-4847-b882-754429f56b0f","Type":"ContainerDied","Data":"fbff5a4c19b1cfbe0c7e57eb05f0093aa3829fbabee820a190f916e0ce4fe6b9"} Nov 21 14:31:35 crc kubenswrapper[4897]: I1121 14:31:35.284431 4897 scope.go:117] "RemoveContainer" containerID="cdc84dcb39834c826868fb54c0e5d0aa88dbf0ae2b0ceea30db993577fc2bded" Nov 21 14:31:36 crc kubenswrapper[4897]: E1121 14:31:36.393531 4897 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-placement-api:current-podified" Nov 21 14:31:36 crc kubenswrapper[4897]: E1121 14:31:36.394047 4897 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:placement-db-sync,Image:quay.io/podified-antelope-centos9/openstack-placement-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/placement,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:placement-dbsync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-v7wxh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42482,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-db-sync-xj48w_openstack(b0c8cf90-d5b4-43b0-9037-4b579f26c74b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 21 14:31:36 crc kubenswrapper[4897]: E1121 14:31:36.395306 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/placement-db-sync-xj48w" podUID="b0c8cf90-d5b4-43b0-9037-4b579f26c74b" Nov 21 14:31:37 crc kubenswrapper[4897]: E1121 14:31:37.325353 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-placement-api:current-podified\\\"\"" pod="openstack/placement-db-sync-xj48w" podUID="b0c8cf90-d5b4-43b0-9037-4b579f26c74b" Nov 21 14:31:38 crc kubenswrapper[4897]: I1121 14:31:38.061449 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6d5b6d6b67-qc7fs" podUID="17a6c4a6-4493-4531-8f5c-814c7333a6ae" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.164:5353: i/o timeout" Nov 21 14:31:43 crc kubenswrapper[4897]: I1121 14:31:43.062456 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6d5b6d6b67-qc7fs" podUID="17a6c4a6-4493-4531-8f5c-814c7333a6ae" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.164:5353: i/o timeout" Nov 21 14:31:48 crc kubenswrapper[4897]: I1121 14:31:48.064218 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6d5b6d6b67-qc7fs" podUID="17a6c4a6-4493-4531-8f5c-814c7333a6ae" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.164:5353: i/o timeout" Nov 21 14:31:53 crc kubenswrapper[4897]: I1121 14:31:53.065812 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6d5b6d6b67-qc7fs" podUID="17a6c4a6-4493-4531-8f5c-814c7333a6ae" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.164:5353: i/o timeout" Nov 21 14:31:54 crc kubenswrapper[4897]: I1121 14:31:54.938453 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d5b6d6b67-qc7fs" Nov 21 14:31:54 crc kubenswrapper[4897]: I1121 14:31:54.945945 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-mrvrw" Nov 21 14:31:55 crc kubenswrapper[4897]: I1121 14:31:55.035157 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/72a6a702-fda5-4a79-8a0b-10b9cae6d9c7-credential-keys\") pod \"72a6a702-fda5-4a79-8a0b-10b9cae6d9c7\" (UID: \"72a6a702-fda5-4a79-8a0b-10b9cae6d9c7\") " Nov 21 14:31:55 crc kubenswrapper[4897]: I1121 14:31:55.035220 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/17a6c4a6-4493-4531-8f5c-814c7333a6ae-ovsdbserver-nb\") pod \"17a6c4a6-4493-4531-8f5c-814c7333a6ae\" (UID: \"17a6c4a6-4493-4531-8f5c-814c7333a6ae\") " Nov 21 14:31:55 crc kubenswrapper[4897]: I1121 14:31:55.035264 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2blq5\" (UniqueName: \"kubernetes.io/projected/17a6c4a6-4493-4531-8f5c-814c7333a6ae-kube-api-access-2blq5\") pod \"17a6c4a6-4493-4531-8f5c-814c7333a6ae\" (UID: \"17a6c4a6-4493-4531-8f5c-814c7333a6ae\") " Nov 21 14:31:55 crc kubenswrapper[4897]: I1121 14:31:55.035314 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/17a6c4a6-4493-4531-8f5c-814c7333a6ae-dns-svc\") pod \"17a6c4a6-4493-4531-8f5c-814c7333a6ae\" (UID: \"17a6c4a6-4493-4531-8f5c-814c7333a6ae\") " Nov 21 14:31:55 crc kubenswrapper[4897]: I1121 14:31:55.035341 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72a6a702-fda5-4a79-8a0b-10b9cae6d9c7-config-data\") pod \"72a6a702-fda5-4a79-8a0b-10b9cae6d9c7\" (UID: \"72a6a702-fda5-4a79-8a0b-10b9cae6d9c7\") " Nov 21 14:31:55 crc kubenswrapper[4897]: I1121 14:31:55.035377 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72a6a702-fda5-4a79-8a0b-10b9cae6d9c7-combined-ca-bundle\") pod \"72a6a702-fda5-4a79-8a0b-10b9cae6d9c7\" (UID: \"72a6a702-fda5-4a79-8a0b-10b9cae6d9c7\") " Nov 21 14:31:55 crc kubenswrapper[4897]: I1121 14:31:55.035415 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/17a6c4a6-4493-4531-8f5c-814c7333a6ae-dns-swift-storage-0\") pod \"17a6c4a6-4493-4531-8f5c-814c7333a6ae\" (UID: \"17a6c4a6-4493-4531-8f5c-814c7333a6ae\") " Nov 21 14:31:55 crc kubenswrapper[4897]: I1121 14:31:55.035461 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/17a6c4a6-4493-4531-8f5c-814c7333a6ae-ovsdbserver-sb\") pod \"17a6c4a6-4493-4531-8f5c-814c7333a6ae\" (UID: \"17a6c4a6-4493-4531-8f5c-814c7333a6ae\") " Nov 21 14:31:55 crc kubenswrapper[4897]: I1121 14:31:55.035490 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17a6c4a6-4493-4531-8f5c-814c7333a6ae-config\") pod \"17a6c4a6-4493-4531-8f5c-814c7333a6ae\" (UID: \"17a6c4a6-4493-4531-8f5c-814c7333a6ae\") " Nov 21 14:31:55 crc kubenswrapper[4897]: I1121 14:31:55.035525 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-chhx7\" (UniqueName: \"kubernetes.io/projected/72a6a702-fda5-4a79-8a0b-10b9cae6d9c7-kube-api-access-chhx7\") pod \"72a6a702-fda5-4a79-8a0b-10b9cae6d9c7\" (UID: \"72a6a702-fda5-4a79-8a0b-10b9cae6d9c7\") " Nov 21 14:31:55 crc kubenswrapper[4897]: I1121 14:31:55.035559 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/72a6a702-fda5-4a79-8a0b-10b9cae6d9c7-fernet-keys\") pod \"72a6a702-fda5-4a79-8a0b-10b9cae6d9c7\" (UID: \"72a6a702-fda5-4a79-8a0b-10b9cae6d9c7\") " Nov 21 14:31:55 crc kubenswrapper[4897]: I1121 14:31:55.035574 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/72a6a702-fda5-4a79-8a0b-10b9cae6d9c7-scripts\") pod \"72a6a702-fda5-4a79-8a0b-10b9cae6d9c7\" (UID: \"72a6a702-fda5-4a79-8a0b-10b9cae6d9c7\") " Nov 21 14:31:55 crc kubenswrapper[4897]: I1121 14:31:55.087983 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72a6a702-fda5-4a79-8a0b-10b9cae6d9c7-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "72a6a702-fda5-4a79-8a0b-10b9cae6d9c7" (UID: "72a6a702-fda5-4a79-8a0b-10b9cae6d9c7"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:31:55 crc kubenswrapper[4897]: I1121 14:31:55.100705 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72a6a702-fda5-4a79-8a0b-10b9cae6d9c7-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "72a6a702-fda5-4a79-8a0b-10b9cae6d9c7" (UID: "72a6a702-fda5-4a79-8a0b-10b9cae6d9c7"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:31:55 crc kubenswrapper[4897]: I1121 14:31:55.100798 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72a6a702-fda5-4a79-8a0b-10b9cae6d9c7-scripts" (OuterVolumeSpecName: "scripts") pod "72a6a702-fda5-4a79-8a0b-10b9cae6d9c7" (UID: "72a6a702-fda5-4a79-8a0b-10b9cae6d9c7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:31:55 crc kubenswrapper[4897]: I1121 14:31:55.111728 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17a6c4a6-4493-4531-8f5c-814c7333a6ae-kube-api-access-2blq5" (OuterVolumeSpecName: "kube-api-access-2blq5") pod "17a6c4a6-4493-4531-8f5c-814c7333a6ae" (UID: "17a6c4a6-4493-4531-8f5c-814c7333a6ae"). InnerVolumeSpecName "kube-api-access-2blq5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:31:55 crc kubenswrapper[4897]: I1121 14:31:55.137963 4897 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/72a6a702-fda5-4a79-8a0b-10b9cae6d9c7-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 21 14:31:55 crc kubenswrapper[4897]: I1121 14:31:55.138000 4897 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/72a6a702-fda5-4a79-8a0b-10b9cae6d9c7-scripts\") on node \"crc\" DevicePath \"\"" Nov 21 14:31:55 crc kubenswrapper[4897]: I1121 14:31:55.138009 4897 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/72a6a702-fda5-4a79-8a0b-10b9cae6d9c7-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 21 14:31:55 crc kubenswrapper[4897]: I1121 14:31:55.138018 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2blq5\" (UniqueName: \"kubernetes.io/projected/17a6c4a6-4493-4531-8f5c-814c7333a6ae-kube-api-access-2blq5\") on node \"crc\" DevicePath \"\"" Nov 21 14:31:55 crc kubenswrapper[4897]: I1121 14:31:55.160276 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72a6a702-fda5-4a79-8a0b-10b9cae6d9c7-kube-api-access-chhx7" (OuterVolumeSpecName: "kube-api-access-chhx7") pod "72a6a702-fda5-4a79-8a0b-10b9cae6d9c7" (UID: "72a6a702-fda5-4a79-8a0b-10b9cae6d9c7"). InnerVolumeSpecName "kube-api-access-chhx7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:31:55 crc kubenswrapper[4897]: I1121 14:31:55.219013 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72a6a702-fda5-4a79-8a0b-10b9cae6d9c7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "72a6a702-fda5-4a79-8a0b-10b9cae6d9c7" (UID: "72a6a702-fda5-4a79-8a0b-10b9cae6d9c7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:31:55 crc kubenswrapper[4897]: I1121 14:31:55.234665 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72a6a702-fda5-4a79-8a0b-10b9cae6d9c7-config-data" (OuterVolumeSpecName: "config-data") pod "72a6a702-fda5-4a79-8a0b-10b9cae6d9c7" (UID: "72a6a702-fda5-4a79-8a0b-10b9cae6d9c7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:31:55 crc kubenswrapper[4897]: I1121 14:31:55.239146 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/17a6c4a6-4493-4531-8f5c-814c7333a6ae-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "17a6c4a6-4493-4531-8f5c-814c7333a6ae" (UID: "17a6c4a6-4493-4531-8f5c-814c7333a6ae"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:31:55 crc kubenswrapper[4897]: I1121 14:31:55.239637 4897 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/17a6c4a6-4493-4531-8f5c-814c7333a6ae-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 21 14:31:55 crc kubenswrapper[4897]: I1121 14:31:55.239662 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-chhx7\" (UniqueName: \"kubernetes.io/projected/72a6a702-fda5-4a79-8a0b-10b9cae6d9c7-kube-api-access-chhx7\") on node \"crc\" DevicePath \"\"" Nov 21 14:31:55 crc kubenswrapper[4897]: I1121 14:31:55.239671 4897 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72a6a702-fda5-4a79-8a0b-10b9cae6d9c7-config-data\") on node \"crc\" DevicePath \"\"" Nov 21 14:31:55 crc kubenswrapper[4897]: I1121 14:31:55.239680 4897 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72a6a702-fda5-4a79-8a0b-10b9cae6d9c7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 21 14:31:55 crc kubenswrapper[4897]: I1121 14:31:55.248804 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/17a6c4a6-4493-4531-8f5c-814c7333a6ae-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "17a6c4a6-4493-4531-8f5c-814c7333a6ae" (UID: "17a6c4a6-4493-4531-8f5c-814c7333a6ae"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:31:55 crc kubenswrapper[4897]: I1121 14:31:55.253840 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/17a6c4a6-4493-4531-8f5c-814c7333a6ae-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "17a6c4a6-4493-4531-8f5c-814c7333a6ae" (UID: "17a6c4a6-4493-4531-8f5c-814c7333a6ae"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:31:55 crc kubenswrapper[4897]: I1121 14:31:55.270131 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/17a6c4a6-4493-4531-8f5c-814c7333a6ae-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "17a6c4a6-4493-4531-8f5c-814c7333a6ae" (UID: "17a6c4a6-4493-4531-8f5c-814c7333a6ae"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:31:55 crc kubenswrapper[4897]: I1121 14:31:55.276256 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/17a6c4a6-4493-4531-8f5c-814c7333a6ae-config" (OuterVolumeSpecName: "config") pod "17a6c4a6-4493-4531-8f5c-814c7333a6ae" (UID: "17a6c4a6-4493-4531-8f5c-814c7333a6ae"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:31:55 crc kubenswrapper[4897]: I1121 14:31:55.341636 4897 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/17a6c4a6-4493-4531-8f5c-814c7333a6ae-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 21 14:31:55 crc kubenswrapper[4897]: I1121 14:31:55.341691 4897 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/17a6c4a6-4493-4531-8f5c-814c7333a6ae-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 21 14:31:55 crc kubenswrapper[4897]: I1121 14:31:55.341705 4897 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/17a6c4a6-4493-4531-8f5c-814c7333a6ae-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 21 14:31:55 crc kubenswrapper[4897]: I1121 14:31:55.341715 4897 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17a6c4a6-4493-4531-8f5c-814c7333a6ae-config\") on node \"crc\" DevicePath \"\"" Nov 21 14:31:55 crc kubenswrapper[4897]: I1121 14:31:55.549480 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-mrvrw" event={"ID":"72a6a702-fda5-4a79-8a0b-10b9cae6d9c7","Type":"ContainerDied","Data":"0a4c04fd3d6f3e822c208737b3d64d33d89ddd30ee98aeb96442367b0fbea2b5"} Nov 21 14:31:55 crc kubenswrapper[4897]: I1121 14:31:55.549536 4897 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0a4c04fd3d6f3e822c208737b3d64d33d89ddd30ee98aeb96442367b0fbea2b5" Nov 21 14:31:55 crc kubenswrapper[4897]: I1121 14:31:55.549593 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-mrvrw" Nov 21 14:31:55 crc kubenswrapper[4897]: I1121 14:31:55.554418 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d5b6d6b67-qc7fs" event={"ID":"17a6c4a6-4493-4531-8f5c-814c7333a6ae","Type":"ContainerDied","Data":"7c030fe110868de3a3e43d878600b726357395b387e561ca836f1ad54225b6f0"} Nov 21 14:31:55 crc kubenswrapper[4897]: I1121 14:31:55.554471 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d5b6d6b67-qc7fs" Nov 21 14:31:55 crc kubenswrapper[4897]: I1121 14:31:55.596023 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d5b6d6b67-qc7fs"] Nov 21 14:31:55 crc kubenswrapper[4897]: I1121 14:31:55.614066 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6d5b6d6b67-qc7fs"] Nov 21 14:31:56 crc kubenswrapper[4897]: I1121 14:31:56.077640 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-mrvrw"] Nov 21 14:31:56 crc kubenswrapper[4897]: I1121 14:31:56.102893 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="17a6c4a6-4493-4531-8f5c-814c7333a6ae" path="/var/lib/kubelet/pods/17a6c4a6-4493-4531-8f5c-814c7333a6ae/volumes" Nov 21 14:31:56 crc kubenswrapper[4897]: I1121 14:31:56.104031 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-mrvrw"] Nov 21 14:31:56 crc kubenswrapper[4897]: I1121 14:31:56.168685 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-bgs55"] Nov 21 14:31:56 crc kubenswrapper[4897]: E1121 14:31:56.169077 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17a6c4a6-4493-4531-8f5c-814c7333a6ae" containerName="dnsmasq-dns" Nov 21 14:31:56 crc kubenswrapper[4897]: I1121 14:31:56.169094 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="17a6c4a6-4493-4531-8f5c-814c7333a6ae" containerName="dnsmasq-dns" Nov 21 14:31:56 crc kubenswrapper[4897]: E1121 14:31:56.169110 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17a6c4a6-4493-4531-8f5c-814c7333a6ae" containerName="init" Nov 21 14:31:56 crc kubenswrapper[4897]: I1121 14:31:56.169117 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="17a6c4a6-4493-4531-8f5c-814c7333a6ae" containerName="init" Nov 21 14:31:56 crc kubenswrapper[4897]: E1121 14:31:56.169136 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72a6a702-fda5-4a79-8a0b-10b9cae6d9c7" containerName="keystone-bootstrap" Nov 21 14:31:56 crc kubenswrapper[4897]: I1121 14:31:56.169142 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="72a6a702-fda5-4a79-8a0b-10b9cae6d9c7" containerName="keystone-bootstrap" Nov 21 14:31:56 crc kubenswrapper[4897]: E1121 14:31:56.169157 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b7418e5-f3f0-46a8-9c0d-b84b922763be" containerName="init" Nov 21 14:31:56 crc kubenswrapper[4897]: I1121 14:31:56.169162 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b7418e5-f3f0-46a8-9c0d-b84b922763be" containerName="init" Nov 21 14:31:56 crc kubenswrapper[4897]: I1121 14:31:56.169352 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="72a6a702-fda5-4a79-8a0b-10b9cae6d9c7" containerName="keystone-bootstrap" Nov 21 14:31:56 crc kubenswrapper[4897]: I1121 14:31:56.169372 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="17a6c4a6-4493-4531-8f5c-814c7333a6ae" containerName="dnsmasq-dns" Nov 21 14:31:56 crc kubenswrapper[4897]: I1121 14:31:56.169386 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b7418e5-f3f0-46a8-9c0d-b84b922763be" containerName="init" Nov 21 14:31:56 crc kubenswrapper[4897]: I1121 14:31:56.170071 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-bgs55" Nov 21 14:31:56 crc kubenswrapper[4897]: I1121 14:31:56.174332 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-wbfsb" Nov 21 14:31:56 crc kubenswrapper[4897]: I1121 14:31:56.174520 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 21 14:31:56 crc kubenswrapper[4897]: I1121 14:31:56.174529 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 21 14:31:56 crc kubenswrapper[4897]: I1121 14:31:56.174558 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 21 14:31:56 crc kubenswrapper[4897]: I1121 14:31:56.176562 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 21 14:31:56 crc kubenswrapper[4897]: I1121 14:31:56.181260 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-bgs55"] Nov 21 14:31:56 crc kubenswrapper[4897]: I1121 14:31:56.260868 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f7487cb-deab-4a72-ad73-5d805b4e25cd-combined-ca-bundle\") pod \"keystone-bootstrap-bgs55\" (UID: \"7f7487cb-deab-4a72-ad73-5d805b4e25cd\") " pod="openstack/keystone-bootstrap-bgs55" Nov 21 14:31:56 crc kubenswrapper[4897]: I1121 14:31:56.260941 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/7f7487cb-deab-4a72-ad73-5d805b4e25cd-credential-keys\") pod \"keystone-bootstrap-bgs55\" (UID: \"7f7487cb-deab-4a72-ad73-5d805b4e25cd\") " pod="openstack/keystone-bootstrap-bgs55" Nov 21 14:31:56 crc kubenswrapper[4897]: I1121 14:31:56.261046 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7f7487cb-deab-4a72-ad73-5d805b4e25cd-fernet-keys\") pod \"keystone-bootstrap-bgs55\" (UID: \"7f7487cb-deab-4a72-ad73-5d805b4e25cd\") " pod="openstack/keystone-bootstrap-bgs55" Nov 21 14:31:56 crc kubenswrapper[4897]: I1121 14:31:56.261075 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ppj2b\" (UniqueName: \"kubernetes.io/projected/7f7487cb-deab-4a72-ad73-5d805b4e25cd-kube-api-access-ppj2b\") pod \"keystone-bootstrap-bgs55\" (UID: \"7f7487cb-deab-4a72-ad73-5d805b4e25cd\") " pod="openstack/keystone-bootstrap-bgs55" Nov 21 14:31:56 crc kubenswrapper[4897]: I1121 14:31:56.261122 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f7487cb-deab-4a72-ad73-5d805b4e25cd-config-data\") pod \"keystone-bootstrap-bgs55\" (UID: \"7f7487cb-deab-4a72-ad73-5d805b4e25cd\") " pod="openstack/keystone-bootstrap-bgs55" Nov 21 14:31:56 crc kubenswrapper[4897]: I1121 14:31:56.261195 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7f7487cb-deab-4a72-ad73-5d805b4e25cd-scripts\") pod \"keystone-bootstrap-bgs55\" (UID: \"7f7487cb-deab-4a72-ad73-5d805b4e25cd\") " pod="openstack/keystone-bootstrap-bgs55" Nov 21 14:31:56 crc kubenswrapper[4897]: I1121 14:31:56.362452 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f7487cb-deab-4a72-ad73-5d805b4e25cd-combined-ca-bundle\") pod \"keystone-bootstrap-bgs55\" (UID: \"7f7487cb-deab-4a72-ad73-5d805b4e25cd\") " pod="openstack/keystone-bootstrap-bgs55" Nov 21 14:31:56 crc kubenswrapper[4897]: I1121 14:31:56.362583 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/7f7487cb-deab-4a72-ad73-5d805b4e25cd-credential-keys\") pod \"keystone-bootstrap-bgs55\" (UID: \"7f7487cb-deab-4a72-ad73-5d805b4e25cd\") " pod="openstack/keystone-bootstrap-bgs55" Nov 21 14:31:56 crc kubenswrapper[4897]: I1121 14:31:56.362707 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7f7487cb-deab-4a72-ad73-5d805b4e25cd-fernet-keys\") pod \"keystone-bootstrap-bgs55\" (UID: \"7f7487cb-deab-4a72-ad73-5d805b4e25cd\") " pod="openstack/keystone-bootstrap-bgs55" Nov 21 14:31:56 crc kubenswrapper[4897]: I1121 14:31:56.362750 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ppj2b\" (UniqueName: \"kubernetes.io/projected/7f7487cb-deab-4a72-ad73-5d805b4e25cd-kube-api-access-ppj2b\") pod \"keystone-bootstrap-bgs55\" (UID: \"7f7487cb-deab-4a72-ad73-5d805b4e25cd\") " pod="openstack/keystone-bootstrap-bgs55" Nov 21 14:31:56 crc kubenswrapper[4897]: I1121 14:31:56.362918 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f7487cb-deab-4a72-ad73-5d805b4e25cd-config-data\") pod \"keystone-bootstrap-bgs55\" (UID: \"7f7487cb-deab-4a72-ad73-5d805b4e25cd\") " pod="openstack/keystone-bootstrap-bgs55" Nov 21 14:31:56 crc kubenswrapper[4897]: I1121 14:31:56.363071 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7f7487cb-deab-4a72-ad73-5d805b4e25cd-scripts\") pod \"keystone-bootstrap-bgs55\" (UID: \"7f7487cb-deab-4a72-ad73-5d805b4e25cd\") " pod="openstack/keystone-bootstrap-bgs55" Nov 21 14:31:56 crc kubenswrapper[4897]: I1121 14:31:56.366885 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7f7487cb-deab-4a72-ad73-5d805b4e25cd-scripts\") pod \"keystone-bootstrap-bgs55\" (UID: \"7f7487cb-deab-4a72-ad73-5d805b4e25cd\") " pod="openstack/keystone-bootstrap-bgs55" Nov 21 14:31:56 crc kubenswrapper[4897]: I1121 14:31:56.367232 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/7f7487cb-deab-4a72-ad73-5d805b4e25cd-credential-keys\") pod \"keystone-bootstrap-bgs55\" (UID: \"7f7487cb-deab-4a72-ad73-5d805b4e25cd\") " pod="openstack/keystone-bootstrap-bgs55" Nov 21 14:31:56 crc kubenswrapper[4897]: I1121 14:31:56.367259 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f7487cb-deab-4a72-ad73-5d805b4e25cd-config-data\") pod \"keystone-bootstrap-bgs55\" (UID: \"7f7487cb-deab-4a72-ad73-5d805b4e25cd\") " pod="openstack/keystone-bootstrap-bgs55" Nov 21 14:31:56 crc kubenswrapper[4897]: I1121 14:31:56.367825 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f7487cb-deab-4a72-ad73-5d805b4e25cd-combined-ca-bundle\") pod \"keystone-bootstrap-bgs55\" (UID: \"7f7487cb-deab-4a72-ad73-5d805b4e25cd\") " pod="openstack/keystone-bootstrap-bgs55" Nov 21 14:31:56 crc kubenswrapper[4897]: I1121 14:31:56.371065 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7f7487cb-deab-4a72-ad73-5d805b4e25cd-fernet-keys\") pod \"keystone-bootstrap-bgs55\" (UID: \"7f7487cb-deab-4a72-ad73-5d805b4e25cd\") " pod="openstack/keystone-bootstrap-bgs55" Nov 21 14:31:56 crc kubenswrapper[4897]: I1121 14:31:56.382262 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ppj2b\" (UniqueName: \"kubernetes.io/projected/7f7487cb-deab-4a72-ad73-5d805b4e25cd-kube-api-access-ppj2b\") pod \"keystone-bootstrap-bgs55\" (UID: \"7f7487cb-deab-4a72-ad73-5d805b4e25cd\") " pod="openstack/keystone-bootstrap-bgs55" Nov 21 14:31:56 crc kubenswrapper[4897]: I1121 14:31:56.518125 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-bgs55" Nov 21 14:31:58 crc kubenswrapper[4897]: I1121 14:31:58.067903 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6d5b6d6b67-qc7fs" podUID="17a6c4a6-4493-4531-8f5c-814c7333a6ae" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.164:5353: i/o timeout" Nov 21 14:31:58 crc kubenswrapper[4897]: I1121 14:31:58.113283 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="72a6a702-fda5-4a79-8a0b-10b9cae6d9c7" path="/var/lib/kubelet/pods/72a6a702-fda5-4a79-8a0b-10b9cae6d9c7/volumes" Nov 21 14:31:59 crc kubenswrapper[4897]: E1121 14:31:59.142092 4897 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified" Nov 21 14:31:59 crc kubenswrapper[4897]: E1121 14:31:59.143566 4897 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:heat-db-sync,Image:quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified,Command:[/bin/bash],Args:[-c /usr/bin/heat-manage --config-dir /etc/heat/heat.conf.d db_sync],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/heat/heat.conf.d/00-default.conf,SubPath:00-default.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/heat/heat.conf.d/01-custom.conf,SubPath:01-custom.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-p4csc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42418,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*42418,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-db-sync-tbtcb_openstack(59fd52b7-f25b-4a81-a961-41e6cb526c89): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 21 14:31:59 crc kubenswrapper[4897]: E1121 14:31:59.144796 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/heat-db-sync-tbtcb" podUID="59fd52b7-f25b-4a81-a961-41e6cb526c89" Nov 21 14:31:59 crc kubenswrapper[4897]: E1121 14:31:59.603952 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified\\\"\"" pod="openstack/heat-db-sync-tbtcb" podUID="59fd52b7-f25b-4a81-a961-41e6cb526c89" Nov 21 14:32:00 crc kubenswrapper[4897]: I1121 14:32:00.174865 4897 scope.go:117] "RemoveContainer" containerID="afe2abe49994bcaffbe92412ab26a0240d4102be177179d0e806cea09b185b65" Nov 21 14:32:00 crc kubenswrapper[4897]: E1121 14:32:00.215319 4897 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Nov 21 14:32:00 crc kubenswrapper[4897]: E1121 14:32:00.215534 4897 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dzppc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-7dvt9_openstack(55644d8b-b8a3-463e-8045-b7391e9c2419): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 21 14:32:00 crc kubenswrapper[4897]: E1121 14:32:00.217285 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-7dvt9" podUID="55644d8b-b8a3-463e-8045-b7391e9c2419" Nov 21 14:32:00 crc kubenswrapper[4897]: I1121 14:32:00.405492 4897 scope.go:117] "RemoveContainer" containerID="164ea673a2b950fa6284d85e4eb1fe8e329dd37706927daf4eaa21b953f48acb" Nov 21 14:32:00 crc kubenswrapper[4897]: E1121 14:32:00.620819 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-7dvt9" podUID="55644d8b-b8a3-463e-8045-b7391e9c2419" Nov 21 14:32:00 crc kubenswrapper[4897]: I1121 14:32:00.842518 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-bgs55"] Nov 21 14:32:00 crc kubenswrapper[4897]: W1121 14:32:00.858106 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7f7487cb_deab_4a72_ad73_5d805b4e25cd.slice/crio-cb4ab73c3755461228638e8f997fc71e305376d51cbc3d99b9327a3e121b4542 WatchSource:0}: Error finding container cb4ab73c3755461228638e8f997fc71e305376d51cbc3d99b9327a3e121b4542: Status 404 returned error can't find the container with id cb4ab73c3755461228638e8f997fc71e305376d51cbc3d99b9327a3e121b4542 Nov 21 14:32:01 crc kubenswrapper[4897]: I1121 14:32:01.667405 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-bgs55" event={"ID":"7f7487cb-deab-4a72-ad73-5d805b4e25cd","Type":"ContainerStarted","Data":"442b3e7ff2a0cf9fd50378ea8e061b6f8a89f463d1baacbad2a4188c2f931c88"} Nov 21 14:32:01 crc kubenswrapper[4897]: I1121 14:32:01.668168 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-bgs55" event={"ID":"7f7487cb-deab-4a72-ad73-5d805b4e25cd","Type":"ContainerStarted","Data":"cb4ab73c3755461228638e8f997fc71e305376d51cbc3d99b9327a3e121b4542"} Nov 21 14:32:01 crc kubenswrapper[4897]: I1121 14:32:01.678151 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-dm94g" event={"ID":"de087e28-b2d8-4686-b758-f4cd38443f25","Type":"ContainerStarted","Data":"9d873a09028f196e71827f91a3a2045d109b5c5614da0e1247c853f713bf9839"} Nov 21 14:32:01 crc kubenswrapper[4897]: I1121 14:32:01.681267 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c3696f1d-db0b-4910-a65c-31410e2316df","Type":"ContainerStarted","Data":"4e82c69b36fc692afc0ff6d360048a4a9871804fa3ae5fb182d5567f08ec50d1"} Nov 21 14:32:01 crc kubenswrapper[4897]: I1121 14:32:01.684008 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" event={"ID":"e7670227-d280-4847-b882-754429f56b0f","Type":"ContainerStarted","Data":"c68f1b003d12b4a2964169a85a04e3f0272d86c891f52e6bebd52f3781520d1a"} Nov 21 14:32:01 crc kubenswrapper[4897]: I1121 14:32:01.686388 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-xj48w" event={"ID":"b0c8cf90-d5b4-43b0-9037-4b579f26c74b","Type":"ContainerStarted","Data":"f494f8ffa1945123f3045599a6855553d8ce35e55d6ce7171e3c0d5f9ed36b69"} Nov 21 14:32:01 crc kubenswrapper[4897]: I1121 14:32:01.698303 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-bgs55" podStartSLOduration=5.698283545 podStartE2EDuration="5.698283545s" podCreationTimestamp="2025-11-21 14:31:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:32:01.691838351 +0000 UTC m=+1398.976431836" watchObservedRunningTime="2025-11-21 14:32:01.698283545 +0000 UTC m=+1398.982877020" Nov 21 14:32:01 crc kubenswrapper[4897]: I1121 14:32:01.714807 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-xj48w" podStartSLOduration=3.071211861 podStartE2EDuration="59.714785597s" podCreationTimestamp="2025-11-21 14:31:02 +0000 UTC" firstStartedPulling="2025-11-21 14:31:03.968896604 +0000 UTC m=+1341.253490079" lastFinishedPulling="2025-11-21 14:32:00.61247035 +0000 UTC m=+1397.897063815" observedRunningTime="2025-11-21 14:32:01.712058534 +0000 UTC m=+1398.996652009" watchObservedRunningTime="2025-11-21 14:32:01.714785597 +0000 UTC m=+1398.999379082" Nov 21 14:32:01 crc kubenswrapper[4897]: I1121 14:32:01.752800 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-dm94g" podStartSLOduration=3.09022858 podStartE2EDuration="59.752777226s" podCreationTimestamp="2025-11-21 14:31:02 +0000 UTC" firstStartedPulling="2025-11-21 14:31:03.955423262 +0000 UTC m=+1341.240016747" lastFinishedPulling="2025-11-21 14:32:00.617971928 +0000 UTC m=+1397.902565393" observedRunningTime="2025-11-21 14:32:01.743086546 +0000 UTC m=+1399.027680021" watchObservedRunningTime="2025-11-21 14:32:01.752777226 +0000 UTC m=+1399.037370711" Nov 21 14:32:03 crc kubenswrapper[4897]: I1121 14:32:03.711413 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c3696f1d-db0b-4910-a65c-31410e2316df","Type":"ContainerStarted","Data":"b676d00cf16995ea43993b65d8f5b05023450400f02e4e944e5a3777a2e7e0a5"} Nov 21 14:32:04 crc kubenswrapper[4897]: I1121 14:32:04.737181 4897 generic.go:334] "Generic (PLEG): container finished" podID="7f7487cb-deab-4a72-ad73-5d805b4e25cd" containerID="442b3e7ff2a0cf9fd50378ea8e061b6f8a89f463d1baacbad2a4188c2f931c88" exitCode=0 Nov 21 14:32:04 crc kubenswrapper[4897]: I1121 14:32:04.737279 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-bgs55" event={"ID":"7f7487cb-deab-4a72-ad73-5d805b4e25cd","Type":"ContainerDied","Data":"442b3e7ff2a0cf9fd50378ea8e061b6f8a89f463d1baacbad2a4188c2f931c88"} Nov 21 14:32:05 crc kubenswrapper[4897]: I1121 14:32:05.748792 4897 generic.go:334] "Generic (PLEG): container finished" podID="b0c8cf90-d5b4-43b0-9037-4b579f26c74b" containerID="f494f8ffa1945123f3045599a6855553d8ce35e55d6ce7171e3c0d5f9ed36b69" exitCode=0 Nov 21 14:32:05 crc kubenswrapper[4897]: I1121 14:32:05.748879 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-xj48w" event={"ID":"b0c8cf90-d5b4-43b0-9037-4b579f26c74b","Type":"ContainerDied","Data":"f494f8ffa1945123f3045599a6855553d8ce35e55d6ce7171e3c0d5f9ed36b69"} Nov 21 14:32:06 crc kubenswrapper[4897]: I1121 14:32:06.760752 4897 generic.go:334] "Generic (PLEG): container finished" podID="de087e28-b2d8-4686-b758-f4cd38443f25" containerID="9d873a09028f196e71827f91a3a2045d109b5c5614da0e1247c853f713bf9839" exitCode=0 Nov 21 14:32:06 crc kubenswrapper[4897]: I1121 14:32:06.760885 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-dm94g" event={"ID":"de087e28-b2d8-4686-b758-f4cd38443f25","Type":"ContainerDied","Data":"9d873a09028f196e71827f91a3a2045d109b5c5614da0e1247c853f713bf9839"} Nov 21 14:32:08 crc kubenswrapper[4897]: I1121 14:32:08.035472 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-bgs55" Nov 21 14:32:08 crc kubenswrapper[4897]: I1121 14:32:08.137681 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7f7487cb-deab-4a72-ad73-5d805b4e25cd-fernet-keys\") pod \"7f7487cb-deab-4a72-ad73-5d805b4e25cd\" (UID: \"7f7487cb-deab-4a72-ad73-5d805b4e25cd\") " Nov 21 14:32:08 crc kubenswrapper[4897]: I1121 14:32:08.138614 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f7487cb-deab-4a72-ad73-5d805b4e25cd-combined-ca-bundle\") pod \"7f7487cb-deab-4a72-ad73-5d805b4e25cd\" (UID: \"7f7487cb-deab-4a72-ad73-5d805b4e25cd\") " Nov 21 14:32:08 crc kubenswrapper[4897]: I1121 14:32:08.138813 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ppj2b\" (UniqueName: \"kubernetes.io/projected/7f7487cb-deab-4a72-ad73-5d805b4e25cd-kube-api-access-ppj2b\") pod \"7f7487cb-deab-4a72-ad73-5d805b4e25cd\" (UID: \"7f7487cb-deab-4a72-ad73-5d805b4e25cd\") " Nov 21 14:32:08 crc kubenswrapper[4897]: I1121 14:32:08.138978 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f7487cb-deab-4a72-ad73-5d805b4e25cd-config-data\") pod \"7f7487cb-deab-4a72-ad73-5d805b4e25cd\" (UID: \"7f7487cb-deab-4a72-ad73-5d805b4e25cd\") " Nov 21 14:32:08 crc kubenswrapper[4897]: I1121 14:32:08.139108 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7f7487cb-deab-4a72-ad73-5d805b4e25cd-scripts\") pod \"7f7487cb-deab-4a72-ad73-5d805b4e25cd\" (UID: \"7f7487cb-deab-4a72-ad73-5d805b4e25cd\") " Nov 21 14:32:08 crc kubenswrapper[4897]: I1121 14:32:08.139234 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/7f7487cb-deab-4a72-ad73-5d805b4e25cd-credential-keys\") pod \"7f7487cb-deab-4a72-ad73-5d805b4e25cd\" (UID: \"7f7487cb-deab-4a72-ad73-5d805b4e25cd\") " Nov 21 14:32:08 crc kubenswrapper[4897]: I1121 14:32:08.143593 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f7487cb-deab-4a72-ad73-5d805b4e25cd-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "7f7487cb-deab-4a72-ad73-5d805b4e25cd" (UID: "7f7487cb-deab-4a72-ad73-5d805b4e25cd"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:32:08 crc kubenswrapper[4897]: I1121 14:32:08.143626 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f7487cb-deab-4a72-ad73-5d805b4e25cd-kube-api-access-ppj2b" (OuterVolumeSpecName: "kube-api-access-ppj2b") pod "7f7487cb-deab-4a72-ad73-5d805b4e25cd" (UID: "7f7487cb-deab-4a72-ad73-5d805b4e25cd"). InnerVolumeSpecName "kube-api-access-ppj2b". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:32:08 crc kubenswrapper[4897]: I1121 14:32:08.144600 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f7487cb-deab-4a72-ad73-5d805b4e25cd-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "7f7487cb-deab-4a72-ad73-5d805b4e25cd" (UID: "7f7487cb-deab-4a72-ad73-5d805b4e25cd"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:32:08 crc kubenswrapper[4897]: I1121 14:32:08.146929 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f7487cb-deab-4a72-ad73-5d805b4e25cd-scripts" (OuterVolumeSpecName: "scripts") pod "7f7487cb-deab-4a72-ad73-5d805b4e25cd" (UID: "7f7487cb-deab-4a72-ad73-5d805b4e25cd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:32:08 crc kubenswrapper[4897]: I1121 14:32:08.168809 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f7487cb-deab-4a72-ad73-5d805b4e25cd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7f7487cb-deab-4a72-ad73-5d805b4e25cd" (UID: "7f7487cb-deab-4a72-ad73-5d805b4e25cd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:32:08 crc kubenswrapper[4897]: I1121 14:32:08.172981 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f7487cb-deab-4a72-ad73-5d805b4e25cd-config-data" (OuterVolumeSpecName: "config-data") pod "7f7487cb-deab-4a72-ad73-5d805b4e25cd" (UID: "7f7487cb-deab-4a72-ad73-5d805b4e25cd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:32:08 crc kubenswrapper[4897]: I1121 14:32:08.183712 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-xj48w" Nov 21 14:32:08 crc kubenswrapper[4897]: I1121 14:32:08.211332 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-dm94g" Nov 21 14:32:08 crc kubenswrapper[4897]: I1121 14:32:08.241329 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b0c8cf90-d5b4-43b0-9037-4b579f26c74b-logs\") pod \"b0c8cf90-d5b4-43b0-9037-4b579f26c74b\" (UID: \"b0c8cf90-d5b4-43b0-9037-4b579f26c74b\") " Nov 21 14:32:08 crc kubenswrapper[4897]: I1121 14:32:08.241432 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b0c8cf90-d5b4-43b0-9037-4b579f26c74b-scripts\") pod \"b0c8cf90-d5b4-43b0-9037-4b579f26c74b\" (UID: \"b0c8cf90-d5b4-43b0-9037-4b579f26c74b\") " Nov 21 14:32:08 crc kubenswrapper[4897]: I1121 14:32:08.241658 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0c8cf90-d5b4-43b0-9037-4b579f26c74b-config-data\") pod \"b0c8cf90-d5b4-43b0-9037-4b579f26c74b\" (UID: \"b0c8cf90-d5b4-43b0-9037-4b579f26c74b\") " Nov 21 14:32:08 crc kubenswrapper[4897]: I1121 14:32:08.242386 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0c8cf90-d5b4-43b0-9037-4b579f26c74b-combined-ca-bundle\") pod \"b0c8cf90-d5b4-43b0-9037-4b579f26c74b\" (UID: \"b0c8cf90-d5b4-43b0-9037-4b579f26c74b\") " Nov 21 14:32:08 crc kubenswrapper[4897]: I1121 14:32:08.242488 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v7wxh\" (UniqueName: \"kubernetes.io/projected/b0c8cf90-d5b4-43b0-9037-4b579f26c74b-kube-api-access-v7wxh\") pod \"b0c8cf90-d5b4-43b0-9037-4b579f26c74b\" (UID: \"b0c8cf90-d5b4-43b0-9037-4b579f26c74b\") " Nov 21 14:32:08 crc kubenswrapper[4897]: I1121 14:32:08.243180 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b0c8cf90-d5b4-43b0-9037-4b579f26c74b-logs" (OuterVolumeSpecName: "logs") pod "b0c8cf90-d5b4-43b0-9037-4b579f26c74b" (UID: "b0c8cf90-d5b4-43b0-9037-4b579f26c74b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:32:08 crc kubenswrapper[4897]: I1121 14:32:08.243771 4897 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7f7487cb-deab-4a72-ad73-5d805b4e25cd-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 21 14:32:08 crc kubenswrapper[4897]: I1121 14:32:08.243800 4897 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f7487cb-deab-4a72-ad73-5d805b4e25cd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 21 14:32:08 crc kubenswrapper[4897]: I1121 14:32:08.243812 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ppj2b\" (UniqueName: \"kubernetes.io/projected/7f7487cb-deab-4a72-ad73-5d805b4e25cd-kube-api-access-ppj2b\") on node \"crc\" DevicePath \"\"" Nov 21 14:32:08 crc kubenswrapper[4897]: I1121 14:32:08.243821 4897 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f7487cb-deab-4a72-ad73-5d805b4e25cd-config-data\") on node \"crc\" DevicePath \"\"" Nov 21 14:32:08 crc kubenswrapper[4897]: I1121 14:32:08.243833 4897 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7f7487cb-deab-4a72-ad73-5d805b4e25cd-scripts\") on node \"crc\" DevicePath \"\"" Nov 21 14:32:08 crc kubenswrapper[4897]: I1121 14:32:08.243841 4897 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/7f7487cb-deab-4a72-ad73-5d805b4e25cd-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 21 14:32:08 crc kubenswrapper[4897]: I1121 14:32:08.243848 4897 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b0c8cf90-d5b4-43b0-9037-4b579f26c74b-logs\") on node \"crc\" DevicePath \"\"" Nov 21 14:32:08 crc kubenswrapper[4897]: I1121 14:32:08.251833 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b0c8cf90-d5b4-43b0-9037-4b579f26c74b-scripts" (OuterVolumeSpecName: "scripts") pod "b0c8cf90-d5b4-43b0-9037-4b579f26c74b" (UID: "b0c8cf90-d5b4-43b0-9037-4b579f26c74b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:32:08 crc kubenswrapper[4897]: I1121 14:32:08.251980 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b0c8cf90-d5b4-43b0-9037-4b579f26c74b-kube-api-access-v7wxh" (OuterVolumeSpecName: "kube-api-access-v7wxh") pod "b0c8cf90-d5b4-43b0-9037-4b579f26c74b" (UID: "b0c8cf90-d5b4-43b0-9037-4b579f26c74b"). InnerVolumeSpecName "kube-api-access-v7wxh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:32:08 crc kubenswrapper[4897]: I1121 14:32:08.281865 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b0c8cf90-d5b4-43b0-9037-4b579f26c74b-config-data" (OuterVolumeSpecName: "config-data") pod "b0c8cf90-d5b4-43b0-9037-4b579f26c74b" (UID: "b0c8cf90-d5b4-43b0-9037-4b579f26c74b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:32:08 crc kubenswrapper[4897]: I1121 14:32:08.296386 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b0c8cf90-d5b4-43b0-9037-4b579f26c74b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b0c8cf90-d5b4-43b0-9037-4b579f26c74b" (UID: "b0c8cf90-d5b4-43b0-9037-4b579f26c74b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:32:08 crc kubenswrapper[4897]: I1121 14:32:08.345478 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de087e28-b2d8-4686-b758-f4cd38443f25-combined-ca-bundle\") pod \"de087e28-b2d8-4686-b758-f4cd38443f25\" (UID: \"de087e28-b2d8-4686-b758-f4cd38443f25\") " Nov 21 14:32:08 crc kubenswrapper[4897]: I1121 14:32:08.345674 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/de087e28-b2d8-4686-b758-f4cd38443f25-db-sync-config-data\") pod \"de087e28-b2d8-4686-b758-f4cd38443f25\" (UID: \"de087e28-b2d8-4686-b758-f4cd38443f25\") " Nov 21 14:32:08 crc kubenswrapper[4897]: I1121 14:32:08.345896 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rn26s\" (UniqueName: \"kubernetes.io/projected/de087e28-b2d8-4686-b758-f4cd38443f25-kube-api-access-rn26s\") pod \"de087e28-b2d8-4686-b758-f4cd38443f25\" (UID: \"de087e28-b2d8-4686-b758-f4cd38443f25\") " Nov 21 14:32:08 crc kubenswrapper[4897]: I1121 14:32:08.346441 4897 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0c8cf90-d5b4-43b0-9037-4b579f26c74b-config-data\") on node \"crc\" DevicePath \"\"" Nov 21 14:32:08 crc kubenswrapper[4897]: I1121 14:32:08.346466 4897 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0c8cf90-d5b4-43b0-9037-4b579f26c74b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 21 14:32:08 crc kubenswrapper[4897]: I1121 14:32:08.346480 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v7wxh\" (UniqueName: \"kubernetes.io/projected/b0c8cf90-d5b4-43b0-9037-4b579f26c74b-kube-api-access-v7wxh\") on node \"crc\" DevicePath \"\"" Nov 21 14:32:08 crc kubenswrapper[4897]: I1121 14:32:08.346493 4897 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b0c8cf90-d5b4-43b0-9037-4b579f26c74b-scripts\") on node \"crc\" DevicePath \"\"" Nov 21 14:32:08 crc kubenswrapper[4897]: I1121 14:32:08.348603 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de087e28-b2d8-4686-b758-f4cd38443f25-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "de087e28-b2d8-4686-b758-f4cd38443f25" (UID: "de087e28-b2d8-4686-b758-f4cd38443f25"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:32:08 crc kubenswrapper[4897]: I1121 14:32:08.350927 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de087e28-b2d8-4686-b758-f4cd38443f25-kube-api-access-rn26s" (OuterVolumeSpecName: "kube-api-access-rn26s") pod "de087e28-b2d8-4686-b758-f4cd38443f25" (UID: "de087e28-b2d8-4686-b758-f4cd38443f25"). InnerVolumeSpecName "kube-api-access-rn26s". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:32:08 crc kubenswrapper[4897]: I1121 14:32:08.371821 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de087e28-b2d8-4686-b758-f4cd38443f25-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "de087e28-b2d8-4686-b758-f4cd38443f25" (UID: "de087e28-b2d8-4686-b758-f4cd38443f25"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:32:08 crc kubenswrapper[4897]: I1121 14:32:08.448933 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rn26s\" (UniqueName: \"kubernetes.io/projected/de087e28-b2d8-4686-b758-f4cd38443f25-kube-api-access-rn26s\") on node \"crc\" DevicePath \"\"" Nov 21 14:32:08 crc kubenswrapper[4897]: I1121 14:32:08.449005 4897 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de087e28-b2d8-4686-b758-f4cd38443f25-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 21 14:32:08 crc kubenswrapper[4897]: I1121 14:32:08.449028 4897 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/de087e28-b2d8-4686-b758-f4cd38443f25-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 21 14:32:08 crc kubenswrapper[4897]: I1121 14:32:08.784234 4897 generic.go:334] "Generic (PLEG): container finished" podID="9c08d8b3-d052-47b7-ad42-7b0f9fbe1920" containerID="f0b5d22717e8bf9ca37f66f5b7cbefcb84e12420967595ff9e141eebb59aaa33" exitCode=0 Nov 21 14:32:08 crc kubenswrapper[4897]: I1121 14:32:08.784351 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-zz6kx" event={"ID":"9c08d8b3-d052-47b7-ad42-7b0f9fbe1920","Type":"ContainerDied","Data":"f0b5d22717e8bf9ca37f66f5b7cbefcb84e12420967595ff9e141eebb59aaa33"} Nov 21 14:32:08 crc kubenswrapper[4897]: I1121 14:32:08.787556 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c3696f1d-db0b-4910-a65c-31410e2316df","Type":"ContainerStarted","Data":"3c6c1e27ec72c0edf905c8a623e78bf432a92f75bf1c931a7707cf9faca2b35d"} Nov 21 14:32:08 crc kubenswrapper[4897]: I1121 14:32:08.789823 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-xj48w" event={"ID":"b0c8cf90-d5b4-43b0-9037-4b579f26c74b","Type":"ContainerDied","Data":"fa7d015e268d0c7617ac51bf6139eda75898180fc42a864c1868f3d23855691b"} Nov 21 14:32:08 crc kubenswrapper[4897]: I1121 14:32:08.789853 4897 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fa7d015e268d0c7617ac51bf6139eda75898180fc42a864c1868f3d23855691b" Nov 21 14:32:08 crc kubenswrapper[4897]: I1121 14:32:08.789875 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-xj48w" Nov 21 14:32:08 crc kubenswrapper[4897]: I1121 14:32:08.791948 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-bgs55" event={"ID":"7f7487cb-deab-4a72-ad73-5d805b4e25cd","Type":"ContainerDied","Data":"cb4ab73c3755461228638e8f997fc71e305376d51cbc3d99b9327a3e121b4542"} Nov 21 14:32:08 crc kubenswrapper[4897]: I1121 14:32:08.791980 4897 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cb4ab73c3755461228638e8f997fc71e305376d51cbc3d99b9327a3e121b4542" Nov 21 14:32:08 crc kubenswrapper[4897]: I1121 14:32:08.791953 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-bgs55" Nov 21 14:32:08 crc kubenswrapper[4897]: I1121 14:32:08.793641 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-dm94g" event={"ID":"de087e28-b2d8-4686-b758-f4cd38443f25","Type":"ContainerDied","Data":"58066df22ae21154428c28012cb0b305024745237e2c2d5ff301b4f066770a1d"} Nov 21 14:32:08 crc kubenswrapper[4897]: I1121 14:32:08.793662 4897 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="58066df22ae21154428c28012cb0b305024745237e2c2d5ff301b4f066770a1d" Nov 21 14:32:08 crc kubenswrapper[4897]: I1121 14:32:08.793711 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-dm94g" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.122670 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-7c944bd776-fnlxn"] Nov 21 14:32:09 crc kubenswrapper[4897]: E1121 14:32:09.123097 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f7487cb-deab-4a72-ad73-5d805b4e25cd" containerName="keystone-bootstrap" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.123110 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f7487cb-deab-4a72-ad73-5d805b4e25cd" containerName="keystone-bootstrap" Nov 21 14:32:09 crc kubenswrapper[4897]: E1121 14:32:09.123130 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de087e28-b2d8-4686-b758-f4cd38443f25" containerName="barbican-db-sync" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.123136 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="de087e28-b2d8-4686-b758-f4cd38443f25" containerName="barbican-db-sync" Nov 21 14:32:09 crc kubenswrapper[4897]: E1121 14:32:09.123151 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0c8cf90-d5b4-43b0-9037-4b579f26c74b" containerName="placement-db-sync" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.123156 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0c8cf90-d5b4-43b0-9037-4b579f26c74b" containerName="placement-db-sync" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.123348 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="de087e28-b2d8-4686-b758-f4cd38443f25" containerName="barbican-db-sync" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.123357 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f7487cb-deab-4a72-ad73-5d805b4e25cd" containerName="keystone-bootstrap" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.123382 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="b0c8cf90-d5b4-43b0-9037-4b579f26c74b" containerName="placement-db-sync" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.124403 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-7c944bd776-fnlxn" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.133068 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.133112 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.133192 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-8p2dh" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.134750 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-78dc44f895-5stt7"] Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.136467 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-78dc44f895-5stt7" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.138579 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.143968 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-7c944bd776-fnlxn"] Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.158618 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-78dc44f895-5stt7"] Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.169004 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p5khp\" (UniqueName: \"kubernetes.io/projected/672ba474-ca76-40df-a93b-fa8d3cc91e94-kube-api-access-p5khp\") pod \"barbican-keystone-listener-7c944bd776-fnlxn\" (UID: \"672ba474-ca76-40df-a93b-fa8d3cc91e94\") " pod="openstack/barbican-keystone-listener-7c944bd776-fnlxn" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.169049 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/672ba474-ca76-40df-a93b-fa8d3cc91e94-config-data-custom\") pod \"barbican-keystone-listener-7c944bd776-fnlxn\" (UID: \"672ba474-ca76-40df-a93b-fa8d3cc91e94\") " pod="openstack/barbican-keystone-listener-7c944bd776-fnlxn" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.169106 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/672ba474-ca76-40df-a93b-fa8d3cc91e94-logs\") pod \"barbican-keystone-listener-7c944bd776-fnlxn\" (UID: \"672ba474-ca76-40df-a93b-fa8d3cc91e94\") " pod="openstack/barbican-keystone-listener-7c944bd776-fnlxn" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.169139 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/672ba474-ca76-40df-a93b-fa8d3cc91e94-combined-ca-bundle\") pod \"barbican-keystone-listener-7c944bd776-fnlxn\" (UID: \"672ba474-ca76-40df-a93b-fa8d3cc91e94\") " pod="openstack/barbican-keystone-listener-7c944bd776-fnlxn" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.169161 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/672ba474-ca76-40df-a93b-fa8d3cc91e94-config-data\") pod \"barbican-keystone-listener-7c944bd776-fnlxn\" (UID: \"672ba474-ca76-40df-a93b-fa8d3cc91e94\") " pod="openstack/barbican-keystone-listener-7c944bd776-fnlxn" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.207910 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-f4f4768cf-drjj2"] Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.210377 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f4f4768cf-drjj2" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.227723 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-f4f4768cf-drjj2"] Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.270964 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1c554928-41d4-4b52-a59b-83411d77596c-config-data-custom\") pod \"barbican-worker-78dc44f895-5stt7\" (UID: \"1c554928-41d4-4b52-a59b-83411d77596c\") " pod="openstack/barbican-worker-78dc44f895-5stt7" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.271262 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9bc19a19-de48-43b0-bcba-12e1a44f354d-dns-svc\") pod \"dnsmasq-dns-f4f4768cf-drjj2\" (UID: \"9bc19a19-de48-43b0-bcba-12e1a44f354d\") " pod="openstack/dnsmasq-dns-f4f4768cf-drjj2" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.271300 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9bc19a19-de48-43b0-bcba-12e1a44f354d-dns-swift-storage-0\") pod \"dnsmasq-dns-f4f4768cf-drjj2\" (UID: \"9bc19a19-de48-43b0-bcba-12e1a44f354d\") " pod="openstack/dnsmasq-dns-f4f4768cf-drjj2" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.271339 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p5khp\" (UniqueName: \"kubernetes.io/projected/672ba474-ca76-40df-a93b-fa8d3cc91e94-kube-api-access-p5khp\") pod \"barbican-keystone-listener-7c944bd776-fnlxn\" (UID: \"672ba474-ca76-40df-a93b-fa8d3cc91e94\") " pod="openstack/barbican-keystone-listener-7c944bd776-fnlxn" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.271360 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/672ba474-ca76-40df-a93b-fa8d3cc91e94-config-data-custom\") pod \"barbican-keystone-listener-7c944bd776-fnlxn\" (UID: \"672ba474-ca76-40df-a93b-fa8d3cc91e94\") " pod="openstack/barbican-keystone-listener-7c944bd776-fnlxn" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.271400 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9bc19a19-de48-43b0-bcba-12e1a44f354d-config\") pod \"dnsmasq-dns-f4f4768cf-drjj2\" (UID: \"9bc19a19-de48-43b0-bcba-12e1a44f354d\") " pod="openstack/dnsmasq-dns-f4f4768cf-drjj2" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.271423 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c554928-41d4-4b52-a59b-83411d77596c-combined-ca-bundle\") pod \"barbican-worker-78dc44f895-5stt7\" (UID: \"1c554928-41d4-4b52-a59b-83411d77596c\") " pod="openstack/barbican-worker-78dc44f895-5stt7" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.271447 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/672ba474-ca76-40df-a93b-fa8d3cc91e94-logs\") pod \"barbican-keystone-listener-7c944bd776-fnlxn\" (UID: \"672ba474-ca76-40df-a93b-fa8d3cc91e94\") " pod="openstack/barbican-keystone-listener-7c944bd776-fnlxn" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.271476 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5jhn\" (UniqueName: \"kubernetes.io/projected/1c554928-41d4-4b52-a59b-83411d77596c-kube-api-access-x5jhn\") pod \"barbican-worker-78dc44f895-5stt7\" (UID: \"1c554928-41d4-4b52-a59b-83411d77596c\") " pod="openstack/barbican-worker-78dc44f895-5stt7" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.271495 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/672ba474-ca76-40df-a93b-fa8d3cc91e94-combined-ca-bundle\") pod \"barbican-keystone-listener-7c944bd776-fnlxn\" (UID: \"672ba474-ca76-40df-a93b-fa8d3cc91e94\") " pod="openstack/barbican-keystone-listener-7c944bd776-fnlxn" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.271527 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/672ba474-ca76-40df-a93b-fa8d3cc91e94-config-data\") pod \"barbican-keystone-listener-7c944bd776-fnlxn\" (UID: \"672ba474-ca76-40df-a93b-fa8d3cc91e94\") " pod="openstack/barbican-keystone-listener-7c944bd776-fnlxn" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.271555 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c554928-41d4-4b52-a59b-83411d77596c-config-data\") pod \"barbican-worker-78dc44f895-5stt7\" (UID: \"1c554928-41d4-4b52-a59b-83411d77596c\") " pod="openstack/barbican-worker-78dc44f895-5stt7" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.271577 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1c554928-41d4-4b52-a59b-83411d77596c-logs\") pod \"barbican-worker-78dc44f895-5stt7\" (UID: \"1c554928-41d4-4b52-a59b-83411d77596c\") " pod="openstack/barbican-worker-78dc44f895-5stt7" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.271592 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9bc19a19-de48-43b0-bcba-12e1a44f354d-ovsdbserver-nb\") pod \"dnsmasq-dns-f4f4768cf-drjj2\" (UID: \"9bc19a19-de48-43b0-bcba-12e1a44f354d\") " pod="openstack/dnsmasq-dns-f4f4768cf-drjj2" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.271620 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9bc19a19-de48-43b0-bcba-12e1a44f354d-ovsdbserver-sb\") pod \"dnsmasq-dns-f4f4768cf-drjj2\" (UID: \"9bc19a19-de48-43b0-bcba-12e1a44f354d\") " pod="openstack/dnsmasq-dns-f4f4768cf-drjj2" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.271661 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h7ksh\" (UniqueName: \"kubernetes.io/projected/9bc19a19-de48-43b0-bcba-12e1a44f354d-kube-api-access-h7ksh\") pod \"dnsmasq-dns-f4f4768cf-drjj2\" (UID: \"9bc19a19-de48-43b0-bcba-12e1a44f354d\") " pod="openstack/dnsmasq-dns-f4f4768cf-drjj2" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.281627 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/672ba474-ca76-40df-a93b-fa8d3cc91e94-config-data-custom\") pod \"barbican-keystone-listener-7c944bd776-fnlxn\" (UID: \"672ba474-ca76-40df-a93b-fa8d3cc91e94\") " pod="openstack/barbican-keystone-listener-7c944bd776-fnlxn" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.281904 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/672ba474-ca76-40df-a93b-fa8d3cc91e94-logs\") pod \"barbican-keystone-listener-7c944bd776-fnlxn\" (UID: \"672ba474-ca76-40df-a93b-fa8d3cc91e94\") " pod="openstack/barbican-keystone-listener-7c944bd776-fnlxn" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.293375 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p5khp\" (UniqueName: \"kubernetes.io/projected/672ba474-ca76-40df-a93b-fa8d3cc91e94-kube-api-access-p5khp\") pod \"barbican-keystone-listener-7c944bd776-fnlxn\" (UID: \"672ba474-ca76-40df-a93b-fa8d3cc91e94\") " pod="openstack/barbican-keystone-listener-7c944bd776-fnlxn" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.303232 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-78cfd6c7f4-mzcx9"] Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.310199 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/672ba474-ca76-40df-a93b-fa8d3cc91e94-config-data\") pod \"barbican-keystone-listener-7c944bd776-fnlxn\" (UID: \"672ba474-ca76-40df-a93b-fa8d3cc91e94\") " pod="openstack/barbican-keystone-listener-7c944bd776-fnlxn" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.312822 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/672ba474-ca76-40df-a93b-fa8d3cc91e94-combined-ca-bundle\") pod \"barbican-keystone-listener-7c944bd776-fnlxn\" (UID: \"672ba474-ca76-40df-a93b-fa8d3cc91e94\") " pod="openstack/barbican-keystone-listener-7c944bd776-fnlxn" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.322101 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-78cfd6c7f4-mzcx9"] Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.322222 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-78cfd6c7f4-mzcx9" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.327958 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.373938 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9bc19a19-de48-43b0-bcba-12e1a44f354d-config\") pod \"dnsmasq-dns-f4f4768cf-drjj2\" (UID: \"9bc19a19-de48-43b0-bcba-12e1a44f354d\") " pod="openstack/dnsmasq-dns-f4f4768cf-drjj2" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.373993 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c554928-41d4-4b52-a59b-83411d77596c-combined-ca-bundle\") pod \"barbican-worker-78dc44f895-5stt7\" (UID: \"1c554928-41d4-4b52-a59b-83411d77596c\") " pod="openstack/barbican-worker-78dc44f895-5stt7" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.374031 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2fda0597-e34c-4e8b-a600-c919f01be4b1-config-data\") pod \"barbican-api-78cfd6c7f4-mzcx9\" (UID: \"2fda0597-e34c-4e8b-a600-c919f01be4b1\") " pod="openstack/barbican-api-78cfd6c7f4-mzcx9" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.374060 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x5jhn\" (UniqueName: \"kubernetes.io/projected/1c554928-41d4-4b52-a59b-83411d77596c-kube-api-access-x5jhn\") pod \"barbican-worker-78dc44f895-5stt7\" (UID: \"1c554928-41d4-4b52-a59b-83411d77596c\") " pod="openstack/barbican-worker-78dc44f895-5stt7" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.374110 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c554928-41d4-4b52-a59b-83411d77596c-config-data\") pod \"barbican-worker-78dc44f895-5stt7\" (UID: \"1c554928-41d4-4b52-a59b-83411d77596c\") " pod="openstack/barbican-worker-78dc44f895-5stt7" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.374129 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1c554928-41d4-4b52-a59b-83411d77596c-logs\") pod \"barbican-worker-78dc44f895-5stt7\" (UID: \"1c554928-41d4-4b52-a59b-83411d77596c\") " pod="openstack/barbican-worker-78dc44f895-5stt7" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.374145 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9bc19a19-de48-43b0-bcba-12e1a44f354d-ovsdbserver-nb\") pod \"dnsmasq-dns-f4f4768cf-drjj2\" (UID: \"9bc19a19-de48-43b0-bcba-12e1a44f354d\") " pod="openstack/dnsmasq-dns-f4f4768cf-drjj2" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.374164 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9bc19a19-de48-43b0-bcba-12e1a44f354d-ovsdbserver-sb\") pod \"dnsmasq-dns-f4f4768cf-drjj2\" (UID: \"9bc19a19-de48-43b0-bcba-12e1a44f354d\") " pod="openstack/dnsmasq-dns-f4f4768cf-drjj2" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.374200 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h7ksh\" (UniqueName: \"kubernetes.io/projected/9bc19a19-de48-43b0-bcba-12e1a44f354d-kube-api-access-h7ksh\") pod \"dnsmasq-dns-f4f4768cf-drjj2\" (UID: \"9bc19a19-de48-43b0-bcba-12e1a44f354d\") " pod="openstack/dnsmasq-dns-f4f4768cf-drjj2" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.374252 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2lhqk\" (UniqueName: \"kubernetes.io/projected/2fda0597-e34c-4e8b-a600-c919f01be4b1-kube-api-access-2lhqk\") pod \"barbican-api-78cfd6c7f4-mzcx9\" (UID: \"2fda0597-e34c-4e8b-a600-c919f01be4b1\") " pod="openstack/barbican-api-78cfd6c7f4-mzcx9" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.374268 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2fda0597-e34c-4e8b-a600-c919f01be4b1-logs\") pod \"barbican-api-78cfd6c7f4-mzcx9\" (UID: \"2fda0597-e34c-4e8b-a600-c919f01be4b1\") " pod="openstack/barbican-api-78cfd6c7f4-mzcx9" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.374296 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1c554928-41d4-4b52-a59b-83411d77596c-config-data-custom\") pod \"barbican-worker-78dc44f895-5stt7\" (UID: \"1c554928-41d4-4b52-a59b-83411d77596c\") " pod="openstack/barbican-worker-78dc44f895-5stt7" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.374317 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2fda0597-e34c-4e8b-a600-c919f01be4b1-config-data-custom\") pod \"barbican-api-78cfd6c7f4-mzcx9\" (UID: \"2fda0597-e34c-4e8b-a600-c919f01be4b1\") " pod="openstack/barbican-api-78cfd6c7f4-mzcx9" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.374334 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fda0597-e34c-4e8b-a600-c919f01be4b1-combined-ca-bundle\") pod \"barbican-api-78cfd6c7f4-mzcx9\" (UID: \"2fda0597-e34c-4e8b-a600-c919f01be4b1\") " pod="openstack/barbican-api-78cfd6c7f4-mzcx9" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.374359 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9bc19a19-de48-43b0-bcba-12e1a44f354d-dns-svc\") pod \"dnsmasq-dns-f4f4768cf-drjj2\" (UID: \"9bc19a19-de48-43b0-bcba-12e1a44f354d\") " pod="openstack/dnsmasq-dns-f4f4768cf-drjj2" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.374389 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9bc19a19-de48-43b0-bcba-12e1a44f354d-dns-swift-storage-0\") pod \"dnsmasq-dns-f4f4768cf-drjj2\" (UID: \"9bc19a19-de48-43b0-bcba-12e1a44f354d\") " pod="openstack/dnsmasq-dns-f4f4768cf-drjj2" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.375277 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9bc19a19-de48-43b0-bcba-12e1a44f354d-dns-swift-storage-0\") pod \"dnsmasq-dns-f4f4768cf-drjj2\" (UID: \"9bc19a19-de48-43b0-bcba-12e1a44f354d\") " pod="openstack/dnsmasq-dns-f4f4768cf-drjj2" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.375845 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9bc19a19-de48-43b0-bcba-12e1a44f354d-config\") pod \"dnsmasq-dns-f4f4768cf-drjj2\" (UID: \"9bc19a19-de48-43b0-bcba-12e1a44f354d\") " pod="openstack/dnsmasq-dns-f4f4768cf-drjj2" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.385612 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1c554928-41d4-4b52-a59b-83411d77596c-config-data-custom\") pod \"barbican-worker-78dc44f895-5stt7\" (UID: \"1c554928-41d4-4b52-a59b-83411d77596c\") " pod="openstack/barbican-worker-78dc44f895-5stt7" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.387573 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-8974ffc64-c68b5"] Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.387959 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1c554928-41d4-4b52-a59b-83411d77596c-logs\") pod \"barbican-worker-78dc44f895-5stt7\" (UID: \"1c554928-41d4-4b52-a59b-83411d77596c\") " pod="openstack/barbican-worker-78dc44f895-5stt7" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.388667 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9bc19a19-de48-43b0-bcba-12e1a44f354d-dns-svc\") pod \"dnsmasq-dns-f4f4768cf-drjj2\" (UID: \"9bc19a19-de48-43b0-bcba-12e1a44f354d\") " pod="openstack/dnsmasq-dns-f4f4768cf-drjj2" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.389498 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-8974ffc64-c68b5" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.391207 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9bc19a19-de48-43b0-bcba-12e1a44f354d-ovsdbserver-sb\") pod \"dnsmasq-dns-f4f4768cf-drjj2\" (UID: \"9bc19a19-de48-43b0-bcba-12e1a44f354d\") " pod="openstack/dnsmasq-dns-f4f4768cf-drjj2" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.391472 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c554928-41d4-4b52-a59b-83411d77596c-config-data\") pod \"barbican-worker-78dc44f895-5stt7\" (UID: \"1c554928-41d4-4b52-a59b-83411d77596c\") " pod="openstack/barbican-worker-78dc44f895-5stt7" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.391492 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.391778 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.391958 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.392296 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-fzrth" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.392416 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.393041 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9bc19a19-de48-43b0-bcba-12e1a44f354d-ovsdbserver-nb\") pod \"dnsmasq-dns-f4f4768cf-drjj2\" (UID: \"9bc19a19-de48-43b0-bcba-12e1a44f354d\") " pod="openstack/dnsmasq-dns-f4f4768cf-drjj2" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.402635 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c554928-41d4-4b52-a59b-83411d77596c-combined-ca-bundle\") pod \"barbican-worker-78dc44f895-5stt7\" (UID: \"1c554928-41d4-4b52-a59b-83411d77596c\") " pod="openstack/barbican-worker-78dc44f895-5stt7" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.406820 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h7ksh\" (UniqueName: \"kubernetes.io/projected/9bc19a19-de48-43b0-bcba-12e1a44f354d-kube-api-access-h7ksh\") pod \"dnsmasq-dns-f4f4768cf-drjj2\" (UID: \"9bc19a19-de48-43b0-bcba-12e1a44f354d\") " pod="openstack/dnsmasq-dns-f4f4768cf-drjj2" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.416868 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x5jhn\" (UniqueName: \"kubernetes.io/projected/1c554928-41d4-4b52-a59b-83411d77596c-kube-api-access-x5jhn\") pod \"barbican-worker-78dc44f895-5stt7\" (UID: \"1c554928-41d4-4b52-a59b-83411d77596c\") " pod="openstack/barbican-worker-78dc44f895-5stt7" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.448815 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-7c944bd776-fnlxn" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.473634 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-77b5d6c768-mnz58"] Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.476339 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-77b5d6c768-mnz58" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.478279 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-78dc44f895-5stt7" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.478772 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.481699 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.482197 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.482301 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.482803 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.483377 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-wbfsb" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.486852 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7fpbx\" (UniqueName: \"kubernetes.io/projected/129e5e44-512d-4d11-930d-a1026e28e61c-kube-api-access-7fpbx\") pod \"placement-8974ffc64-c68b5\" (UID: \"129e5e44-512d-4d11-930d-a1026e28e61c\") " pod="openstack/placement-8974ffc64-c68b5" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.486925 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2fda0597-e34c-4e8b-a600-c919f01be4b1-config-data-custom\") pod \"barbican-api-78cfd6c7f4-mzcx9\" (UID: \"2fda0597-e34c-4e8b-a600-c919f01be4b1\") " pod="openstack/barbican-api-78cfd6c7f4-mzcx9" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.486953 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fda0597-e34c-4e8b-a600-c919f01be4b1-combined-ca-bundle\") pod \"barbican-api-78cfd6c7f4-mzcx9\" (UID: \"2fda0597-e34c-4e8b-a600-c919f01be4b1\") " pod="openstack/barbican-api-78cfd6c7f4-mzcx9" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.487746 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/129e5e44-512d-4d11-930d-a1026e28e61c-internal-tls-certs\") pod \"placement-8974ffc64-c68b5\" (UID: \"129e5e44-512d-4d11-930d-a1026e28e61c\") " pod="openstack/placement-8974ffc64-c68b5" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.487879 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2fda0597-e34c-4e8b-a600-c919f01be4b1-config-data\") pod \"barbican-api-78cfd6c7f4-mzcx9\" (UID: \"2fda0597-e34c-4e8b-a600-c919f01be4b1\") " pod="openstack/barbican-api-78cfd6c7f4-mzcx9" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.487915 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/129e5e44-512d-4d11-930d-a1026e28e61c-combined-ca-bundle\") pod \"placement-8974ffc64-c68b5\" (UID: \"129e5e44-512d-4d11-930d-a1026e28e61c\") " pod="openstack/placement-8974ffc64-c68b5" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.488059 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/129e5e44-512d-4d11-930d-a1026e28e61c-public-tls-certs\") pod \"placement-8974ffc64-c68b5\" (UID: \"129e5e44-512d-4d11-930d-a1026e28e61c\") " pod="openstack/placement-8974ffc64-c68b5" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.488096 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/129e5e44-512d-4d11-930d-a1026e28e61c-scripts\") pod \"placement-8974ffc64-c68b5\" (UID: \"129e5e44-512d-4d11-930d-a1026e28e61c\") " pod="openstack/placement-8974ffc64-c68b5" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.488145 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/129e5e44-512d-4d11-930d-a1026e28e61c-config-data\") pod \"placement-8974ffc64-c68b5\" (UID: \"129e5e44-512d-4d11-930d-a1026e28e61c\") " pod="openstack/placement-8974ffc64-c68b5" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.488186 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2lhqk\" (UniqueName: \"kubernetes.io/projected/2fda0597-e34c-4e8b-a600-c919f01be4b1-kube-api-access-2lhqk\") pod \"barbican-api-78cfd6c7f4-mzcx9\" (UID: \"2fda0597-e34c-4e8b-a600-c919f01be4b1\") " pod="openstack/barbican-api-78cfd6c7f4-mzcx9" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.488204 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2fda0597-e34c-4e8b-a600-c919f01be4b1-logs\") pod \"barbican-api-78cfd6c7f4-mzcx9\" (UID: \"2fda0597-e34c-4e8b-a600-c919f01be4b1\") " pod="openstack/barbican-api-78cfd6c7f4-mzcx9" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.488241 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/129e5e44-512d-4d11-930d-a1026e28e61c-logs\") pod \"placement-8974ffc64-c68b5\" (UID: \"129e5e44-512d-4d11-930d-a1026e28e61c\") " pod="openstack/placement-8974ffc64-c68b5" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.490812 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2fda0597-e34c-4e8b-a600-c919f01be4b1-logs\") pod \"barbican-api-78cfd6c7f4-mzcx9\" (UID: \"2fda0597-e34c-4e8b-a600-c919f01be4b1\") " pod="openstack/barbican-api-78cfd6c7f4-mzcx9" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.497628 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fda0597-e34c-4e8b-a600-c919f01be4b1-combined-ca-bundle\") pod \"barbican-api-78cfd6c7f4-mzcx9\" (UID: \"2fda0597-e34c-4e8b-a600-c919f01be4b1\") " pod="openstack/barbican-api-78cfd6c7f4-mzcx9" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.497800 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2fda0597-e34c-4e8b-a600-c919f01be4b1-config-data-custom\") pod \"barbican-api-78cfd6c7f4-mzcx9\" (UID: \"2fda0597-e34c-4e8b-a600-c919f01be4b1\") " pod="openstack/barbican-api-78cfd6c7f4-mzcx9" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.506924 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2fda0597-e34c-4e8b-a600-c919f01be4b1-config-data\") pod \"barbican-api-78cfd6c7f4-mzcx9\" (UID: \"2fda0597-e34c-4e8b-a600-c919f01be4b1\") " pod="openstack/barbican-api-78cfd6c7f4-mzcx9" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.507003 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-77b5d6c768-mnz58"] Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.510282 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2lhqk\" (UniqueName: \"kubernetes.io/projected/2fda0597-e34c-4e8b-a600-c919f01be4b1-kube-api-access-2lhqk\") pod \"barbican-api-78cfd6c7f4-mzcx9\" (UID: \"2fda0597-e34c-4e8b-a600-c919f01be4b1\") " pod="openstack/barbican-api-78cfd6c7f4-mzcx9" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.551228 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f4f4768cf-drjj2" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.554938 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-78cfd6c7f4-mzcx9" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.557878 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-8974ffc64-c68b5"] Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.592583 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/129e5e44-512d-4d11-930d-a1026e28e61c-config-data\") pod \"placement-8974ffc64-c68b5\" (UID: \"129e5e44-512d-4d11-930d-a1026e28e61c\") " pod="openstack/placement-8974ffc64-c68b5" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.592649 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/129e5e44-512d-4d11-930d-a1026e28e61c-logs\") pod \"placement-8974ffc64-c68b5\" (UID: \"129e5e44-512d-4d11-930d-a1026e28e61c\") " pod="openstack/placement-8974ffc64-c68b5" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.592678 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7fpbx\" (UniqueName: \"kubernetes.io/projected/129e5e44-512d-4d11-930d-a1026e28e61c-kube-api-access-7fpbx\") pod \"placement-8974ffc64-c68b5\" (UID: \"129e5e44-512d-4d11-930d-a1026e28e61c\") " pod="openstack/placement-8974ffc64-c68b5" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.592716 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e20e1ca8-ff86-4836-8703-40370c65e892-scripts\") pod \"keystone-77b5d6c768-mnz58\" (UID: \"e20e1ca8-ff86-4836-8703-40370c65e892\") " pod="openstack/keystone-77b5d6c768-mnz58" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.592767 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/129e5e44-512d-4d11-930d-a1026e28e61c-internal-tls-certs\") pod \"placement-8974ffc64-c68b5\" (UID: \"129e5e44-512d-4d11-930d-a1026e28e61c\") " pod="openstack/placement-8974ffc64-c68b5" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.592795 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e20e1ca8-ff86-4836-8703-40370c65e892-credential-keys\") pod \"keystone-77b5d6c768-mnz58\" (UID: \"e20e1ca8-ff86-4836-8703-40370c65e892\") " pod="openstack/keystone-77b5d6c768-mnz58" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.592817 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e20e1ca8-ff86-4836-8703-40370c65e892-fernet-keys\") pod \"keystone-77b5d6c768-mnz58\" (UID: \"e20e1ca8-ff86-4836-8703-40370c65e892\") " pod="openstack/keystone-77b5d6c768-mnz58" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.592849 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e20e1ca8-ff86-4836-8703-40370c65e892-internal-tls-certs\") pod \"keystone-77b5d6c768-mnz58\" (UID: \"e20e1ca8-ff86-4836-8703-40370c65e892\") " pod="openstack/keystone-77b5d6c768-mnz58" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.592872 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7v5ch\" (UniqueName: \"kubernetes.io/projected/e20e1ca8-ff86-4836-8703-40370c65e892-kube-api-access-7v5ch\") pod \"keystone-77b5d6c768-mnz58\" (UID: \"e20e1ca8-ff86-4836-8703-40370c65e892\") " pod="openstack/keystone-77b5d6c768-mnz58" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.592890 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/129e5e44-512d-4d11-930d-a1026e28e61c-combined-ca-bundle\") pod \"placement-8974ffc64-c68b5\" (UID: \"129e5e44-512d-4d11-930d-a1026e28e61c\") " pod="openstack/placement-8974ffc64-c68b5" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.592944 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e20e1ca8-ff86-4836-8703-40370c65e892-combined-ca-bundle\") pod \"keystone-77b5d6c768-mnz58\" (UID: \"e20e1ca8-ff86-4836-8703-40370c65e892\") " pod="openstack/keystone-77b5d6c768-mnz58" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.592980 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e20e1ca8-ff86-4836-8703-40370c65e892-config-data\") pod \"keystone-77b5d6c768-mnz58\" (UID: \"e20e1ca8-ff86-4836-8703-40370c65e892\") " pod="openstack/keystone-77b5d6c768-mnz58" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.592999 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e20e1ca8-ff86-4836-8703-40370c65e892-public-tls-certs\") pod \"keystone-77b5d6c768-mnz58\" (UID: \"e20e1ca8-ff86-4836-8703-40370c65e892\") " pod="openstack/keystone-77b5d6c768-mnz58" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.593030 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/129e5e44-512d-4d11-930d-a1026e28e61c-public-tls-certs\") pod \"placement-8974ffc64-c68b5\" (UID: \"129e5e44-512d-4d11-930d-a1026e28e61c\") " pod="openstack/placement-8974ffc64-c68b5" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.593052 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/129e5e44-512d-4d11-930d-a1026e28e61c-scripts\") pod \"placement-8974ffc64-c68b5\" (UID: \"129e5e44-512d-4d11-930d-a1026e28e61c\") " pod="openstack/placement-8974ffc64-c68b5" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.596075 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/129e5e44-512d-4d11-930d-a1026e28e61c-logs\") pod \"placement-8974ffc64-c68b5\" (UID: \"129e5e44-512d-4d11-930d-a1026e28e61c\") " pod="openstack/placement-8974ffc64-c68b5" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.599090 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/129e5e44-512d-4d11-930d-a1026e28e61c-internal-tls-certs\") pod \"placement-8974ffc64-c68b5\" (UID: \"129e5e44-512d-4d11-930d-a1026e28e61c\") " pod="openstack/placement-8974ffc64-c68b5" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.599170 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/129e5e44-512d-4d11-930d-a1026e28e61c-config-data\") pod \"placement-8974ffc64-c68b5\" (UID: \"129e5e44-512d-4d11-930d-a1026e28e61c\") " pod="openstack/placement-8974ffc64-c68b5" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.600601 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/129e5e44-512d-4d11-930d-a1026e28e61c-scripts\") pod \"placement-8974ffc64-c68b5\" (UID: \"129e5e44-512d-4d11-930d-a1026e28e61c\") " pod="openstack/placement-8974ffc64-c68b5" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.600843 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/129e5e44-512d-4d11-930d-a1026e28e61c-public-tls-certs\") pod \"placement-8974ffc64-c68b5\" (UID: \"129e5e44-512d-4d11-930d-a1026e28e61c\") " pod="openstack/placement-8974ffc64-c68b5" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.601186 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/129e5e44-512d-4d11-930d-a1026e28e61c-combined-ca-bundle\") pod \"placement-8974ffc64-c68b5\" (UID: \"129e5e44-512d-4d11-930d-a1026e28e61c\") " pod="openstack/placement-8974ffc64-c68b5" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.611858 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7fpbx\" (UniqueName: \"kubernetes.io/projected/129e5e44-512d-4d11-930d-a1026e28e61c-kube-api-access-7fpbx\") pod \"placement-8974ffc64-c68b5\" (UID: \"129e5e44-512d-4d11-930d-a1026e28e61c\") " pod="openstack/placement-8974ffc64-c68b5" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.696183 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e20e1ca8-ff86-4836-8703-40370c65e892-scripts\") pod \"keystone-77b5d6c768-mnz58\" (UID: \"e20e1ca8-ff86-4836-8703-40370c65e892\") " pod="openstack/keystone-77b5d6c768-mnz58" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.696548 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e20e1ca8-ff86-4836-8703-40370c65e892-credential-keys\") pod \"keystone-77b5d6c768-mnz58\" (UID: \"e20e1ca8-ff86-4836-8703-40370c65e892\") " pod="openstack/keystone-77b5d6c768-mnz58" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.696572 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e20e1ca8-ff86-4836-8703-40370c65e892-fernet-keys\") pod \"keystone-77b5d6c768-mnz58\" (UID: \"e20e1ca8-ff86-4836-8703-40370c65e892\") " pod="openstack/keystone-77b5d6c768-mnz58" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.696606 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e20e1ca8-ff86-4836-8703-40370c65e892-internal-tls-certs\") pod \"keystone-77b5d6c768-mnz58\" (UID: \"e20e1ca8-ff86-4836-8703-40370c65e892\") " pod="openstack/keystone-77b5d6c768-mnz58" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.696635 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7v5ch\" (UniqueName: \"kubernetes.io/projected/e20e1ca8-ff86-4836-8703-40370c65e892-kube-api-access-7v5ch\") pod \"keystone-77b5d6c768-mnz58\" (UID: \"e20e1ca8-ff86-4836-8703-40370c65e892\") " pod="openstack/keystone-77b5d6c768-mnz58" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.696668 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e20e1ca8-ff86-4836-8703-40370c65e892-combined-ca-bundle\") pod \"keystone-77b5d6c768-mnz58\" (UID: \"e20e1ca8-ff86-4836-8703-40370c65e892\") " pod="openstack/keystone-77b5d6c768-mnz58" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.696701 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e20e1ca8-ff86-4836-8703-40370c65e892-config-data\") pod \"keystone-77b5d6c768-mnz58\" (UID: \"e20e1ca8-ff86-4836-8703-40370c65e892\") " pod="openstack/keystone-77b5d6c768-mnz58" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.696726 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e20e1ca8-ff86-4836-8703-40370c65e892-public-tls-certs\") pod \"keystone-77b5d6c768-mnz58\" (UID: \"e20e1ca8-ff86-4836-8703-40370c65e892\") " pod="openstack/keystone-77b5d6c768-mnz58" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.702867 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e20e1ca8-ff86-4836-8703-40370c65e892-scripts\") pod \"keystone-77b5d6c768-mnz58\" (UID: \"e20e1ca8-ff86-4836-8703-40370c65e892\") " pod="openstack/keystone-77b5d6c768-mnz58" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.705657 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e20e1ca8-ff86-4836-8703-40370c65e892-credential-keys\") pod \"keystone-77b5d6c768-mnz58\" (UID: \"e20e1ca8-ff86-4836-8703-40370c65e892\") " pod="openstack/keystone-77b5d6c768-mnz58" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.709308 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e20e1ca8-ff86-4836-8703-40370c65e892-fernet-keys\") pod \"keystone-77b5d6c768-mnz58\" (UID: \"e20e1ca8-ff86-4836-8703-40370c65e892\") " pod="openstack/keystone-77b5d6c768-mnz58" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.711976 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e20e1ca8-ff86-4836-8703-40370c65e892-combined-ca-bundle\") pod \"keystone-77b5d6c768-mnz58\" (UID: \"e20e1ca8-ff86-4836-8703-40370c65e892\") " pod="openstack/keystone-77b5d6c768-mnz58" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.714778 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e20e1ca8-ff86-4836-8703-40370c65e892-internal-tls-certs\") pod \"keystone-77b5d6c768-mnz58\" (UID: \"e20e1ca8-ff86-4836-8703-40370c65e892\") " pod="openstack/keystone-77b5d6c768-mnz58" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.726414 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e20e1ca8-ff86-4836-8703-40370c65e892-public-tls-certs\") pod \"keystone-77b5d6c768-mnz58\" (UID: \"e20e1ca8-ff86-4836-8703-40370c65e892\") " pod="openstack/keystone-77b5d6c768-mnz58" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.727089 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e20e1ca8-ff86-4836-8703-40370c65e892-config-data\") pod \"keystone-77b5d6c768-mnz58\" (UID: \"e20e1ca8-ff86-4836-8703-40370c65e892\") " pod="openstack/keystone-77b5d6c768-mnz58" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.740145 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7v5ch\" (UniqueName: \"kubernetes.io/projected/e20e1ca8-ff86-4836-8703-40370c65e892-kube-api-access-7v5ch\") pod \"keystone-77b5d6c768-mnz58\" (UID: \"e20e1ca8-ff86-4836-8703-40370c65e892\") " pod="openstack/keystone-77b5d6c768-mnz58" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.877304 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-8974ffc64-c68b5" Nov 21 14:32:09 crc kubenswrapper[4897]: I1121 14:32:09.890116 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-77b5d6c768-mnz58" Nov 21 14:32:10 crc kubenswrapper[4897]: I1121 14:32:10.297591 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-78dc44f895-5stt7"] Nov 21 14:32:10 crc kubenswrapper[4897]: W1121 14:32:10.342233 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2fda0597_e34c_4e8b_a600_c919f01be4b1.slice/crio-ab87fdd0c844586389f5eb55925299035e386669a41ac71ec499c5db3bfa9c78 WatchSource:0}: Error finding container ab87fdd0c844586389f5eb55925299035e386669a41ac71ec499c5db3bfa9c78: Status 404 returned error can't find the container with id ab87fdd0c844586389f5eb55925299035e386669a41ac71ec499c5db3bfa9c78 Nov 21 14:32:10 crc kubenswrapper[4897]: I1121 14:32:10.345212 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-78cfd6c7f4-mzcx9"] Nov 21 14:32:10 crc kubenswrapper[4897]: I1121 14:32:10.430384 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-f4f4768cf-drjj2"] Nov 21 14:32:10 crc kubenswrapper[4897]: W1121 14:32:10.501198 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod672ba474_ca76_40df_a93b_fa8d3cc91e94.slice/crio-69caed7470cdce13e1f27cd07a3323dcffdf60220b4da87448fa05ebbb7e8a95 WatchSource:0}: Error finding container 69caed7470cdce13e1f27cd07a3323dcffdf60220b4da87448fa05ebbb7e8a95: Status 404 returned error can't find the container with id 69caed7470cdce13e1f27cd07a3323dcffdf60220b4da87448fa05ebbb7e8a95 Nov 21 14:32:10 crc kubenswrapper[4897]: I1121 14:32:10.520705 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-7c944bd776-fnlxn"] Nov 21 14:32:10 crc kubenswrapper[4897]: I1121 14:32:10.723245 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-zz6kx" Nov 21 14:32:10 crc kubenswrapper[4897]: I1121 14:32:10.854566 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-78cfd6c7f4-mzcx9" event={"ID":"2fda0597-e34c-4e8b-a600-c919f01be4b1","Type":"ContainerStarted","Data":"ab87fdd0c844586389f5eb55925299035e386669a41ac71ec499c5db3bfa9c78"} Nov 21 14:32:10 crc kubenswrapper[4897]: I1121 14:32:10.857219 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-zz6kx" event={"ID":"9c08d8b3-d052-47b7-ad42-7b0f9fbe1920","Type":"ContainerDied","Data":"28c6286341a9a2739d5cbbd8ea2338b86f661821a8bfd60f1c227849a4331e17"} Nov 21 14:32:10 crc kubenswrapper[4897]: I1121 14:32:10.857251 4897 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="28c6286341a9a2739d5cbbd8ea2338b86f661821a8bfd60f1c227849a4331e17" Nov 21 14:32:10 crc kubenswrapper[4897]: I1121 14:32:10.857304 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-zz6kx" Nov 21 14:32:10 crc kubenswrapper[4897]: I1121 14:32:10.858449 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c08d8b3-d052-47b7-ad42-7b0f9fbe1920-config-data\") pod \"9c08d8b3-d052-47b7-ad42-7b0f9fbe1920\" (UID: \"9c08d8b3-d052-47b7-ad42-7b0f9fbe1920\") " Nov 21 14:32:10 crc kubenswrapper[4897]: I1121 14:32:10.858690 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gj9bx\" (UniqueName: \"kubernetes.io/projected/9c08d8b3-d052-47b7-ad42-7b0f9fbe1920-kube-api-access-gj9bx\") pod \"9c08d8b3-d052-47b7-ad42-7b0f9fbe1920\" (UID: \"9c08d8b3-d052-47b7-ad42-7b0f9fbe1920\") " Nov 21 14:32:10 crc kubenswrapper[4897]: I1121 14:32:10.859122 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c08d8b3-d052-47b7-ad42-7b0f9fbe1920-combined-ca-bundle\") pod \"9c08d8b3-d052-47b7-ad42-7b0f9fbe1920\" (UID: \"9c08d8b3-d052-47b7-ad42-7b0f9fbe1920\") " Nov 21 14:32:10 crc kubenswrapper[4897]: I1121 14:32:10.859223 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/9c08d8b3-d052-47b7-ad42-7b0f9fbe1920-db-sync-config-data\") pod \"9c08d8b3-d052-47b7-ad42-7b0f9fbe1920\" (UID: \"9c08d8b3-d052-47b7-ad42-7b0f9fbe1920\") " Nov 21 14:32:10 crc kubenswrapper[4897]: I1121 14:32:10.860404 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-8974ffc64-c68b5"] Nov 21 14:32:10 crc kubenswrapper[4897]: I1121 14:32:10.868873 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-78dc44f895-5stt7" event={"ID":"1c554928-41d4-4b52-a59b-83411d77596c","Type":"ContainerStarted","Data":"674a00fa4dadbd5a75ee031b7dea965559819f4f9235edf66345068d9e966068"} Nov 21 14:32:10 crc kubenswrapper[4897]: I1121 14:32:10.870665 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c08d8b3-d052-47b7-ad42-7b0f9fbe1920-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "9c08d8b3-d052-47b7-ad42-7b0f9fbe1920" (UID: "9c08d8b3-d052-47b7-ad42-7b0f9fbe1920"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:32:10 crc kubenswrapper[4897]: I1121 14:32:10.875045 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c08d8b3-d052-47b7-ad42-7b0f9fbe1920-kube-api-access-gj9bx" (OuterVolumeSpecName: "kube-api-access-gj9bx") pod "9c08d8b3-d052-47b7-ad42-7b0f9fbe1920" (UID: "9c08d8b3-d052-47b7-ad42-7b0f9fbe1920"). InnerVolumeSpecName "kube-api-access-gj9bx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:32:10 crc kubenswrapper[4897]: I1121 14:32:10.876611 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f4f4768cf-drjj2" event={"ID":"9bc19a19-de48-43b0-bcba-12e1a44f354d","Type":"ContainerStarted","Data":"4920c48403baa106b6c967e1feda3a0d011ac6b153c90649f65c740a04533576"} Nov 21 14:32:10 crc kubenswrapper[4897]: I1121 14:32:10.884945 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7c944bd776-fnlxn" event={"ID":"672ba474-ca76-40df-a93b-fa8d3cc91e94","Type":"ContainerStarted","Data":"69caed7470cdce13e1f27cd07a3323dcffdf60220b4da87448fa05ebbb7e8a95"} Nov 21 14:32:10 crc kubenswrapper[4897]: I1121 14:32:10.948974 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c08d8b3-d052-47b7-ad42-7b0f9fbe1920-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9c08d8b3-d052-47b7-ad42-7b0f9fbe1920" (UID: "9c08d8b3-d052-47b7-ad42-7b0f9fbe1920"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:32:10 crc kubenswrapper[4897]: I1121 14:32:10.957822 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c08d8b3-d052-47b7-ad42-7b0f9fbe1920-config-data" (OuterVolumeSpecName: "config-data") pod "9c08d8b3-d052-47b7-ad42-7b0f9fbe1920" (UID: "9c08d8b3-d052-47b7-ad42-7b0f9fbe1920"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:32:10 crc kubenswrapper[4897]: I1121 14:32:10.961764 4897 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c08d8b3-d052-47b7-ad42-7b0f9fbe1920-config-data\") on node \"crc\" DevicePath \"\"" Nov 21 14:32:10 crc kubenswrapper[4897]: I1121 14:32:10.961806 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gj9bx\" (UniqueName: \"kubernetes.io/projected/9c08d8b3-d052-47b7-ad42-7b0f9fbe1920-kube-api-access-gj9bx\") on node \"crc\" DevicePath \"\"" Nov 21 14:32:10 crc kubenswrapper[4897]: I1121 14:32:10.961818 4897 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c08d8b3-d052-47b7-ad42-7b0f9fbe1920-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 21 14:32:10 crc kubenswrapper[4897]: I1121 14:32:10.961828 4897 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/9c08d8b3-d052-47b7-ad42-7b0f9fbe1920-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 21 14:32:11 crc kubenswrapper[4897]: I1121 14:32:11.008066 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-77b5d6c768-mnz58"] Nov 21 14:32:11 crc kubenswrapper[4897]: W1121 14:32:11.067516 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode20e1ca8_ff86_4836_8703_40370c65e892.slice/crio-fb11be6d6eeae1166a1984c853f83671020af95d356aac94337e0f1b2b3bbed9 WatchSource:0}: Error finding container fb11be6d6eeae1166a1984c853f83671020af95d356aac94337e0f1b2b3bbed9: Status 404 returned error can't find the container with id fb11be6d6eeae1166a1984c853f83671020af95d356aac94337e0f1b2b3bbed9 Nov 21 14:32:11 crc kubenswrapper[4897]: I1121 14:32:11.191693 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f4f4768cf-drjj2"] Nov 21 14:32:11 crc kubenswrapper[4897]: I1121 14:32:11.274789 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6d66f584d7-hxs7n"] Nov 21 14:32:11 crc kubenswrapper[4897]: E1121 14:32:11.275223 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c08d8b3-d052-47b7-ad42-7b0f9fbe1920" containerName="glance-db-sync" Nov 21 14:32:11 crc kubenswrapper[4897]: I1121 14:32:11.275235 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c08d8b3-d052-47b7-ad42-7b0f9fbe1920" containerName="glance-db-sync" Nov 21 14:32:11 crc kubenswrapper[4897]: I1121 14:32:11.275472 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c08d8b3-d052-47b7-ad42-7b0f9fbe1920" containerName="glance-db-sync" Nov 21 14:32:11 crc kubenswrapper[4897]: I1121 14:32:11.278636 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d66f584d7-hxs7n" Nov 21 14:32:11 crc kubenswrapper[4897]: I1121 14:32:11.304804 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d66f584d7-hxs7n"] Nov 21 14:32:11 crc kubenswrapper[4897]: I1121 14:32:11.382671 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4b23c32f-2539-41a4-b994-eb4210356a13-dns-swift-storage-0\") pod \"dnsmasq-dns-6d66f584d7-hxs7n\" (UID: \"4b23c32f-2539-41a4-b994-eb4210356a13\") " pod="openstack/dnsmasq-dns-6d66f584d7-hxs7n" Nov 21 14:32:11 crc kubenswrapper[4897]: I1121 14:32:11.382744 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4b23c32f-2539-41a4-b994-eb4210356a13-dns-svc\") pod \"dnsmasq-dns-6d66f584d7-hxs7n\" (UID: \"4b23c32f-2539-41a4-b994-eb4210356a13\") " pod="openstack/dnsmasq-dns-6d66f584d7-hxs7n" Nov 21 14:32:11 crc kubenswrapper[4897]: I1121 14:32:11.382849 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b23c32f-2539-41a4-b994-eb4210356a13-config\") pod \"dnsmasq-dns-6d66f584d7-hxs7n\" (UID: \"4b23c32f-2539-41a4-b994-eb4210356a13\") " pod="openstack/dnsmasq-dns-6d66f584d7-hxs7n" Nov 21 14:32:11 crc kubenswrapper[4897]: I1121 14:32:11.382882 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4b23c32f-2539-41a4-b994-eb4210356a13-ovsdbserver-nb\") pod \"dnsmasq-dns-6d66f584d7-hxs7n\" (UID: \"4b23c32f-2539-41a4-b994-eb4210356a13\") " pod="openstack/dnsmasq-dns-6d66f584d7-hxs7n" Nov 21 14:32:11 crc kubenswrapper[4897]: I1121 14:32:11.382923 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4xtl2\" (UniqueName: \"kubernetes.io/projected/4b23c32f-2539-41a4-b994-eb4210356a13-kube-api-access-4xtl2\") pod \"dnsmasq-dns-6d66f584d7-hxs7n\" (UID: \"4b23c32f-2539-41a4-b994-eb4210356a13\") " pod="openstack/dnsmasq-dns-6d66f584d7-hxs7n" Nov 21 14:32:11 crc kubenswrapper[4897]: I1121 14:32:11.382999 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4b23c32f-2539-41a4-b994-eb4210356a13-ovsdbserver-sb\") pod \"dnsmasq-dns-6d66f584d7-hxs7n\" (UID: \"4b23c32f-2539-41a4-b994-eb4210356a13\") " pod="openstack/dnsmasq-dns-6d66f584d7-hxs7n" Nov 21 14:32:11 crc kubenswrapper[4897]: I1121 14:32:11.484801 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b23c32f-2539-41a4-b994-eb4210356a13-config\") pod \"dnsmasq-dns-6d66f584d7-hxs7n\" (UID: \"4b23c32f-2539-41a4-b994-eb4210356a13\") " pod="openstack/dnsmasq-dns-6d66f584d7-hxs7n" Nov 21 14:32:11 crc kubenswrapper[4897]: I1121 14:32:11.485194 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4b23c32f-2539-41a4-b994-eb4210356a13-ovsdbserver-nb\") pod \"dnsmasq-dns-6d66f584d7-hxs7n\" (UID: \"4b23c32f-2539-41a4-b994-eb4210356a13\") " pod="openstack/dnsmasq-dns-6d66f584d7-hxs7n" Nov 21 14:32:11 crc kubenswrapper[4897]: I1121 14:32:11.485241 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4xtl2\" (UniqueName: \"kubernetes.io/projected/4b23c32f-2539-41a4-b994-eb4210356a13-kube-api-access-4xtl2\") pod \"dnsmasq-dns-6d66f584d7-hxs7n\" (UID: \"4b23c32f-2539-41a4-b994-eb4210356a13\") " pod="openstack/dnsmasq-dns-6d66f584d7-hxs7n" Nov 21 14:32:11 crc kubenswrapper[4897]: I1121 14:32:11.485329 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4b23c32f-2539-41a4-b994-eb4210356a13-ovsdbserver-sb\") pod \"dnsmasq-dns-6d66f584d7-hxs7n\" (UID: \"4b23c32f-2539-41a4-b994-eb4210356a13\") " pod="openstack/dnsmasq-dns-6d66f584d7-hxs7n" Nov 21 14:32:11 crc kubenswrapper[4897]: I1121 14:32:11.485447 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4b23c32f-2539-41a4-b994-eb4210356a13-dns-swift-storage-0\") pod \"dnsmasq-dns-6d66f584d7-hxs7n\" (UID: \"4b23c32f-2539-41a4-b994-eb4210356a13\") " pod="openstack/dnsmasq-dns-6d66f584d7-hxs7n" Nov 21 14:32:11 crc kubenswrapper[4897]: I1121 14:32:11.485473 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4b23c32f-2539-41a4-b994-eb4210356a13-dns-svc\") pod \"dnsmasq-dns-6d66f584d7-hxs7n\" (UID: \"4b23c32f-2539-41a4-b994-eb4210356a13\") " pod="openstack/dnsmasq-dns-6d66f584d7-hxs7n" Nov 21 14:32:11 crc kubenswrapper[4897]: I1121 14:32:11.486913 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4b23c32f-2539-41a4-b994-eb4210356a13-ovsdbserver-nb\") pod \"dnsmasq-dns-6d66f584d7-hxs7n\" (UID: \"4b23c32f-2539-41a4-b994-eb4210356a13\") " pod="openstack/dnsmasq-dns-6d66f584d7-hxs7n" Nov 21 14:32:11 crc kubenswrapper[4897]: I1121 14:32:11.487004 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4b23c32f-2539-41a4-b994-eb4210356a13-dns-svc\") pod \"dnsmasq-dns-6d66f584d7-hxs7n\" (UID: \"4b23c32f-2539-41a4-b994-eb4210356a13\") " pod="openstack/dnsmasq-dns-6d66f584d7-hxs7n" Nov 21 14:32:11 crc kubenswrapper[4897]: I1121 14:32:11.487270 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4b23c32f-2539-41a4-b994-eb4210356a13-dns-swift-storage-0\") pod \"dnsmasq-dns-6d66f584d7-hxs7n\" (UID: \"4b23c32f-2539-41a4-b994-eb4210356a13\") " pod="openstack/dnsmasq-dns-6d66f584d7-hxs7n" Nov 21 14:32:11 crc kubenswrapper[4897]: I1121 14:32:11.487287 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4b23c32f-2539-41a4-b994-eb4210356a13-ovsdbserver-sb\") pod \"dnsmasq-dns-6d66f584d7-hxs7n\" (UID: \"4b23c32f-2539-41a4-b994-eb4210356a13\") " pod="openstack/dnsmasq-dns-6d66f584d7-hxs7n" Nov 21 14:32:11 crc kubenswrapper[4897]: I1121 14:32:11.487467 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b23c32f-2539-41a4-b994-eb4210356a13-config\") pod \"dnsmasq-dns-6d66f584d7-hxs7n\" (UID: \"4b23c32f-2539-41a4-b994-eb4210356a13\") " pod="openstack/dnsmasq-dns-6d66f584d7-hxs7n" Nov 21 14:32:11 crc kubenswrapper[4897]: I1121 14:32:11.513605 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4xtl2\" (UniqueName: \"kubernetes.io/projected/4b23c32f-2539-41a4-b994-eb4210356a13-kube-api-access-4xtl2\") pod \"dnsmasq-dns-6d66f584d7-hxs7n\" (UID: \"4b23c32f-2539-41a4-b994-eb4210356a13\") " pod="openstack/dnsmasq-dns-6d66f584d7-hxs7n" Nov 21 14:32:11 crc kubenswrapper[4897]: I1121 14:32:11.661009 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d66f584d7-hxs7n" Nov 21 14:32:11 crc kubenswrapper[4897]: I1121 14:32:11.933623 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f4f4768cf-drjj2" event={"ID":"9bc19a19-de48-43b0-bcba-12e1a44f354d","Type":"ContainerStarted","Data":"47fc6aa31cde67babb5f2a0c99947a38f69932a7eaad737d93dc34b1dfbcb39d"} Nov 21 14:32:11 crc kubenswrapper[4897]: I1121 14:32:11.938830 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-77b5d6c768-mnz58" event={"ID":"e20e1ca8-ff86-4836-8703-40370c65e892","Type":"ContainerStarted","Data":"fb11be6d6eeae1166a1984c853f83671020af95d356aac94337e0f1b2b3bbed9"} Nov 21 14:32:11 crc kubenswrapper[4897]: I1121 14:32:11.943977 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-78cfd6c7f4-mzcx9" event={"ID":"2fda0597-e34c-4e8b-a600-c919f01be4b1","Type":"ContainerStarted","Data":"3205cf41928a4172b4f181fe69f55ee8dd1a3bc5673e03408997ccab345fb4fb"} Nov 21 14:32:11 crc kubenswrapper[4897]: I1121 14:32:11.948473 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-8974ffc64-c68b5" event={"ID":"129e5e44-512d-4d11-930d-a1026e28e61c","Type":"ContainerStarted","Data":"6ad8d8650c50eed64b88c183ec6ebf731a69e70eedf283c63e320d5eb6d0317a"} Nov 21 14:32:12 crc kubenswrapper[4897]: I1121 14:32:12.189944 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 21 14:32:12 crc kubenswrapper[4897]: I1121 14:32:12.193304 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 21 14:32:12 crc kubenswrapper[4897]: I1121 14:32:12.205945 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 21 14:32:12 crc kubenswrapper[4897]: I1121 14:32:12.225071 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Nov 21 14:32:12 crc kubenswrapper[4897]: I1121 14:32:12.225402 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-wk7cs" Nov 21 14:32:12 crc kubenswrapper[4897]: I1121 14:32:12.225583 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 21 14:32:12 crc kubenswrapper[4897]: I1121 14:32:12.275814 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d66f584d7-hxs7n"] Nov 21 14:32:12 crc kubenswrapper[4897]: I1121 14:32:12.309430 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0b659474-79bd-423b-9f81-1748dfa6682f-logs\") pod \"glance-default-external-api-0\" (UID: \"0b659474-79bd-423b-9f81-1748dfa6682f\") " pod="openstack/glance-default-external-api-0" Nov 21 14:32:12 crc kubenswrapper[4897]: I1121 14:32:12.309549 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0b659474-79bd-423b-9f81-1748dfa6682f-scripts\") pod \"glance-default-external-api-0\" (UID: \"0b659474-79bd-423b-9f81-1748dfa6682f\") " pod="openstack/glance-default-external-api-0" Nov 21 14:32:12 crc kubenswrapper[4897]: I1121 14:32:12.309595 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b659474-79bd-423b-9f81-1748dfa6682f-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"0b659474-79bd-423b-9f81-1748dfa6682f\") " pod="openstack/glance-default-external-api-0" Nov 21 14:32:12 crc kubenswrapper[4897]: I1121 14:32:12.309611 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2t86t\" (UniqueName: \"kubernetes.io/projected/0b659474-79bd-423b-9f81-1748dfa6682f-kube-api-access-2t86t\") pod \"glance-default-external-api-0\" (UID: \"0b659474-79bd-423b-9f81-1748dfa6682f\") " pod="openstack/glance-default-external-api-0" Nov 21 14:32:12 crc kubenswrapper[4897]: I1121 14:32:12.309651 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b659474-79bd-423b-9f81-1748dfa6682f-config-data\") pod \"glance-default-external-api-0\" (UID: \"0b659474-79bd-423b-9f81-1748dfa6682f\") " pod="openstack/glance-default-external-api-0" Nov 21 14:32:12 crc kubenswrapper[4897]: I1121 14:32:12.309667 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"0b659474-79bd-423b-9f81-1748dfa6682f\") " pod="openstack/glance-default-external-api-0" Nov 21 14:32:12 crc kubenswrapper[4897]: I1121 14:32:12.309702 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0b659474-79bd-423b-9f81-1748dfa6682f-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"0b659474-79bd-423b-9f81-1748dfa6682f\") " pod="openstack/glance-default-external-api-0" Nov 21 14:32:12 crc kubenswrapper[4897]: I1121 14:32:12.412736 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0b659474-79bd-423b-9f81-1748dfa6682f-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"0b659474-79bd-423b-9f81-1748dfa6682f\") " pod="openstack/glance-default-external-api-0" Nov 21 14:32:12 crc kubenswrapper[4897]: I1121 14:32:12.412861 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0b659474-79bd-423b-9f81-1748dfa6682f-logs\") pod \"glance-default-external-api-0\" (UID: \"0b659474-79bd-423b-9f81-1748dfa6682f\") " pod="openstack/glance-default-external-api-0" Nov 21 14:32:12 crc kubenswrapper[4897]: I1121 14:32:12.412901 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0b659474-79bd-423b-9f81-1748dfa6682f-scripts\") pod \"glance-default-external-api-0\" (UID: \"0b659474-79bd-423b-9f81-1748dfa6682f\") " pod="openstack/glance-default-external-api-0" Nov 21 14:32:12 crc kubenswrapper[4897]: I1121 14:32:12.412942 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2t86t\" (UniqueName: \"kubernetes.io/projected/0b659474-79bd-423b-9f81-1748dfa6682f-kube-api-access-2t86t\") pod \"glance-default-external-api-0\" (UID: \"0b659474-79bd-423b-9f81-1748dfa6682f\") " pod="openstack/glance-default-external-api-0" Nov 21 14:32:12 crc kubenswrapper[4897]: I1121 14:32:12.412962 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b659474-79bd-423b-9f81-1748dfa6682f-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"0b659474-79bd-423b-9f81-1748dfa6682f\") " pod="openstack/glance-default-external-api-0" Nov 21 14:32:12 crc kubenswrapper[4897]: I1121 14:32:12.413001 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b659474-79bd-423b-9f81-1748dfa6682f-config-data\") pod \"glance-default-external-api-0\" (UID: \"0b659474-79bd-423b-9f81-1748dfa6682f\") " pod="openstack/glance-default-external-api-0" Nov 21 14:32:12 crc kubenswrapper[4897]: I1121 14:32:12.413020 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"0b659474-79bd-423b-9f81-1748dfa6682f\") " pod="openstack/glance-default-external-api-0" Nov 21 14:32:12 crc kubenswrapper[4897]: I1121 14:32:12.413150 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0b659474-79bd-423b-9f81-1748dfa6682f-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"0b659474-79bd-423b-9f81-1748dfa6682f\") " pod="openstack/glance-default-external-api-0" Nov 21 14:32:12 crc kubenswrapper[4897]: I1121 14:32:12.413381 4897 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"0b659474-79bd-423b-9f81-1748dfa6682f\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/glance-default-external-api-0" Nov 21 14:32:12 crc kubenswrapper[4897]: I1121 14:32:12.413792 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0b659474-79bd-423b-9f81-1748dfa6682f-logs\") pod \"glance-default-external-api-0\" (UID: \"0b659474-79bd-423b-9f81-1748dfa6682f\") " pod="openstack/glance-default-external-api-0" Nov 21 14:32:12 crc kubenswrapper[4897]: I1121 14:32:12.418205 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0b659474-79bd-423b-9f81-1748dfa6682f-scripts\") pod \"glance-default-external-api-0\" (UID: \"0b659474-79bd-423b-9f81-1748dfa6682f\") " pod="openstack/glance-default-external-api-0" Nov 21 14:32:12 crc kubenswrapper[4897]: I1121 14:32:12.418556 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b659474-79bd-423b-9f81-1748dfa6682f-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"0b659474-79bd-423b-9f81-1748dfa6682f\") " pod="openstack/glance-default-external-api-0" Nov 21 14:32:12 crc kubenswrapper[4897]: I1121 14:32:12.420117 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b659474-79bd-423b-9f81-1748dfa6682f-config-data\") pod \"glance-default-external-api-0\" (UID: \"0b659474-79bd-423b-9f81-1748dfa6682f\") " pod="openstack/glance-default-external-api-0" Nov 21 14:32:12 crc kubenswrapper[4897]: I1121 14:32:12.433351 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2t86t\" (UniqueName: \"kubernetes.io/projected/0b659474-79bd-423b-9f81-1748dfa6682f-kube-api-access-2t86t\") pod \"glance-default-external-api-0\" (UID: \"0b659474-79bd-423b-9f81-1748dfa6682f\") " pod="openstack/glance-default-external-api-0" Nov 21 14:32:12 crc kubenswrapper[4897]: I1121 14:32:12.446886 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 21 14:32:12 crc kubenswrapper[4897]: I1121 14:32:12.448604 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 21 14:32:12 crc kubenswrapper[4897]: I1121 14:32:12.453821 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 21 14:32:12 crc kubenswrapper[4897]: I1121 14:32:12.477813 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"0b659474-79bd-423b-9f81-1748dfa6682f\") " pod="openstack/glance-default-external-api-0" Nov 21 14:32:12 crc kubenswrapper[4897]: I1121 14:32:12.500573 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 21 14:32:12 crc kubenswrapper[4897]: I1121 14:32:12.523241 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9\") " pod="openstack/glance-default-internal-api-0" Nov 21 14:32:12 crc kubenswrapper[4897]: I1121 14:32:12.525121 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9\") " pod="openstack/glance-default-internal-api-0" Nov 21 14:32:12 crc kubenswrapper[4897]: I1121 14:32:12.525234 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9-logs\") pod \"glance-default-internal-api-0\" (UID: \"3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9\") " pod="openstack/glance-default-internal-api-0" Nov 21 14:32:12 crc kubenswrapper[4897]: I1121 14:32:12.525420 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9-scripts\") pod \"glance-default-internal-api-0\" (UID: \"3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9\") " pod="openstack/glance-default-internal-api-0" Nov 21 14:32:12 crc kubenswrapper[4897]: I1121 14:32:12.525899 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p97h9\" (UniqueName: \"kubernetes.io/projected/3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9-kube-api-access-p97h9\") pod \"glance-default-internal-api-0\" (UID: \"3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9\") " pod="openstack/glance-default-internal-api-0" Nov 21 14:32:12 crc kubenswrapper[4897]: I1121 14:32:12.525983 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9\") " pod="openstack/glance-default-internal-api-0" Nov 21 14:32:12 crc kubenswrapper[4897]: I1121 14:32:12.526054 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9-config-data\") pod \"glance-default-internal-api-0\" (UID: \"3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9\") " pod="openstack/glance-default-internal-api-0" Nov 21 14:32:12 crc kubenswrapper[4897]: I1121 14:32:12.544241 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 21 14:32:12 crc kubenswrapper[4897]: I1121 14:32:12.628733 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p97h9\" (UniqueName: \"kubernetes.io/projected/3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9-kube-api-access-p97h9\") pod \"glance-default-internal-api-0\" (UID: \"3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9\") " pod="openstack/glance-default-internal-api-0" Nov 21 14:32:12 crc kubenswrapper[4897]: I1121 14:32:12.629042 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9\") " pod="openstack/glance-default-internal-api-0" Nov 21 14:32:12 crc kubenswrapper[4897]: I1121 14:32:12.629066 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9-config-data\") pod \"glance-default-internal-api-0\" (UID: \"3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9\") " pod="openstack/glance-default-internal-api-0" Nov 21 14:32:12 crc kubenswrapper[4897]: I1121 14:32:12.629104 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9\") " pod="openstack/glance-default-internal-api-0" Nov 21 14:32:12 crc kubenswrapper[4897]: I1121 14:32:12.629121 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9\") " pod="openstack/glance-default-internal-api-0" Nov 21 14:32:12 crc kubenswrapper[4897]: I1121 14:32:12.629135 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9-logs\") pod \"glance-default-internal-api-0\" (UID: \"3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9\") " pod="openstack/glance-default-internal-api-0" Nov 21 14:32:12 crc kubenswrapper[4897]: I1121 14:32:12.629181 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9-scripts\") pod \"glance-default-internal-api-0\" (UID: \"3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9\") " pod="openstack/glance-default-internal-api-0" Nov 21 14:32:12 crc kubenswrapper[4897]: I1121 14:32:12.629919 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9\") " pod="openstack/glance-default-internal-api-0" Nov 21 14:32:12 crc kubenswrapper[4897]: I1121 14:32:12.630110 4897 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/glance-default-internal-api-0" Nov 21 14:32:12 crc kubenswrapper[4897]: I1121 14:32:12.630161 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9-logs\") pod \"glance-default-internal-api-0\" (UID: \"3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9\") " pod="openstack/glance-default-internal-api-0" Nov 21 14:32:12 crc kubenswrapper[4897]: I1121 14:32:12.677250 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9-scripts\") pod \"glance-default-internal-api-0\" (UID: \"3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9\") " pod="openstack/glance-default-internal-api-0" Nov 21 14:32:12 crc kubenswrapper[4897]: I1121 14:32:12.693478 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p97h9\" (UniqueName: \"kubernetes.io/projected/3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9-kube-api-access-p97h9\") pod \"glance-default-internal-api-0\" (UID: \"3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9\") " pod="openstack/glance-default-internal-api-0" Nov 21 14:32:12 crc kubenswrapper[4897]: I1121 14:32:12.697479 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9\") " pod="openstack/glance-default-internal-api-0" Nov 21 14:32:12 crc kubenswrapper[4897]: I1121 14:32:12.703945 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9-config-data\") pod \"glance-default-internal-api-0\" (UID: \"3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9\") " pod="openstack/glance-default-internal-api-0" Nov 21 14:32:12 crc kubenswrapper[4897]: I1121 14:32:12.744264 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9\") " pod="openstack/glance-default-internal-api-0" Nov 21 14:32:12 crc kubenswrapper[4897]: I1121 14:32:12.872414 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 21 14:32:12 crc kubenswrapper[4897]: I1121 14:32:12.881973 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-58586c84c4-nwhrn"] Nov 21 14:32:12 crc kubenswrapper[4897]: I1121 14:32:12.883858 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-58586c84c4-nwhrn" Nov 21 14:32:12 crc kubenswrapper[4897]: I1121 14:32:12.885930 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Nov 21 14:32:12 crc kubenswrapper[4897]: I1121 14:32:12.886639 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Nov 21 14:32:12 crc kubenswrapper[4897]: I1121 14:32:12.929657 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-58586c84c4-nwhrn"] Nov 21 14:32:12 crc kubenswrapper[4897]: I1121 14:32:12.939364 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b1ff925b-2293-4e8b-b59e-153ae7a518d3-internal-tls-certs\") pod \"barbican-api-58586c84c4-nwhrn\" (UID: \"b1ff925b-2293-4e8b-b59e-153ae7a518d3\") " pod="openstack/barbican-api-58586c84c4-nwhrn" Nov 21 14:32:12 crc kubenswrapper[4897]: I1121 14:32:12.939523 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1ff925b-2293-4e8b-b59e-153ae7a518d3-combined-ca-bundle\") pod \"barbican-api-58586c84c4-nwhrn\" (UID: \"b1ff925b-2293-4e8b-b59e-153ae7a518d3\") " pod="openstack/barbican-api-58586c84c4-nwhrn" Nov 21 14:32:12 crc kubenswrapper[4897]: I1121 14:32:12.939557 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fg2wm\" (UniqueName: \"kubernetes.io/projected/b1ff925b-2293-4e8b-b59e-153ae7a518d3-kube-api-access-fg2wm\") pod \"barbican-api-58586c84c4-nwhrn\" (UID: \"b1ff925b-2293-4e8b-b59e-153ae7a518d3\") " pod="openstack/barbican-api-58586c84c4-nwhrn" Nov 21 14:32:12 crc kubenswrapper[4897]: I1121 14:32:12.939622 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b1ff925b-2293-4e8b-b59e-153ae7a518d3-logs\") pod \"barbican-api-58586c84c4-nwhrn\" (UID: \"b1ff925b-2293-4e8b-b59e-153ae7a518d3\") " pod="openstack/barbican-api-58586c84c4-nwhrn" Nov 21 14:32:12 crc kubenswrapper[4897]: I1121 14:32:12.939657 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1ff925b-2293-4e8b-b59e-153ae7a518d3-config-data\") pod \"barbican-api-58586c84c4-nwhrn\" (UID: \"b1ff925b-2293-4e8b-b59e-153ae7a518d3\") " pod="openstack/barbican-api-58586c84c4-nwhrn" Nov 21 14:32:12 crc kubenswrapper[4897]: I1121 14:32:12.939692 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b1ff925b-2293-4e8b-b59e-153ae7a518d3-config-data-custom\") pod \"barbican-api-58586c84c4-nwhrn\" (UID: \"b1ff925b-2293-4e8b-b59e-153ae7a518d3\") " pod="openstack/barbican-api-58586c84c4-nwhrn" Nov 21 14:32:12 crc kubenswrapper[4897]: I1121 14:32:12.939771 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b1ff925b-2293-4e8b-b59e-153ae7a518d3-public-tls-certs\") pod \"barbican-api-58586c84c4-nwhrn\" (UID: \"b1ff925b-2293-4e8b-b59e-153ae7a518d3\") " pod="openstack/barbican-api-58586c84c4-nwhrn" Nov 21 14:32:12 crc kubenswrapper[4897]: I1121 14:32:12.990358 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d66f584d7-hxs7n" event={"ID":"4b23c32f-2539-41a4-b994-eb4210356a13","Type":"ContainerStarted","Data":"6fd563e868da419d7571f71309d8dd05c6238bbd22da10b375c366fcd7059388"} Nov 21 14:32:13 crc kubenswrapper[4897]: I1121 14:32:13.043993 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b1ff925b-2293-4e8b-b59e-153ae7a518d3-internal-tls-certs\") pod \"barbican-api-58586c84c4-nwhrn\" (UID: \"b1ff925b-2293-4e8b-b59e-153ae7a518d3\") " pod="openstack/barbican-api-58586c84c4-nwhrn" Nov 21 14:32:13 crc kubenswrapper[4897]: I1121 14:32:13.044138 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1ff925b-2293-4e8b-b59e-153ae7a518d3-combined-ca-bundle\") pod \"barbican-api-58586c84c4-nwhrn\" (UID: \"b1ff925b-2293-4e8b-b59e-153ae7a518d3\") " pod="openstack/barbican-api-58586c84c4-nwhrn" Nov 21 14:32:13 crc kubenswrapper[4897]: I1121 14:32:13.044176 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fg2wm\" (UniqueName: \"kubernetes.io/projected/b1ff925b-2293-4e8b-b59e-153ae7a518d3-kube-api-access-fg2wm\") pod \"barbican-api-58586c84c4-nwhrn\" (UID: \"b1ff925b-2293-4e8b-b59e-153ae7a518d3\") " pod="openstack/barbican-api-58586c84c4-nwhrn" Nov 21 14:32:13 crc kubenswrapper[4897]: I1121 14:32:13.044235 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b1ff925b-2293-4e8b-b59e-153ae7a518d3-logs\") pod \"barbican-api-58586c84c4-nwhrn\" (UID: \"b1ff925b-2293-4e8b-b59e-153ae7a518d3\") " pod="openstack/barbican-api-58586c84c4-nwhrn" Nov 21 14:32:13 crc kubenswrapper[4897]: I1121 14:32:13.044306 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1ff925b-2293-4e8b-b59e-153ae7a518d3-config-data\") pod \"barbican-api-58586c84c4-nwhrn\" (UID: \"b1ff925b-2293-4e8b-b59e-153ae7a518d3\") " pod="openstack/barbican-api-58586c84c4-nwhrn" Nov 21 14:32:13 crc kubenswrapper[4897]: I1121 14:32:13.044328 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b1ff925b-2293-4e8b-b59e-153ae7a518d3-config-data-custom\") pod \"barbican-api-58586c84c4-nwhrn\" (UID: \"b1ff925b-2293-4e8b-b59e-153ae7a518d3\") " pod="openstack/barbican-api-58586c84c4-nwhrn" Nov 21 14:32:13 crc kubenswrapper[4897]: I1121 14:32:13.044413 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b1ff925b-2293-4e8b-b59e-153ae7a518d3-public-tls-certs\") pod \"barbican-api-58586c84c4-nwhrn\" (UID: \"b1ff925b-2293-4e8b-b59e-153ae7a518d3\") " pod="openstack/barbican-api-58586c84c4-nwhrn" Nov 21 14:32:13 crc kubenswrapper[4897]: I1121 14:32:13.047049 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b1ff925b-2293-4e8b-b59e-153ae7a518d3-logs\") pod \"barbican-api-58586c84c4-nwhrn\" (UID: \"b1ff925b-2293-4e8b-b59e-153ae7a518d3\") " pod="openstack/barbican-api-58586c84c4-nwhrn" Nov 21 14:32:13 crc kubenswrapper[4897]: I1121 14:32:13.051529 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b1ff925b-2293-4e8b-b59e-153ae7a518d3-internal-tls-certs\") pod \"barbican-api-58586c84c4-nwhrn\" (UID: \"b1ff925b-2293-4e8b-b59e-153ae7a518d3\") " pod="openstack/barbican-api-58586c84c4-nwhrn" Nov 21 14:32:13 crc kubenswrapper[4897]: I1121 14:32:13.051549 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1ff925b-2293-4e8b-b59e-153ae7a518d3-combined-ca-bundle\") pod \"barbican-api-58586c84c4-nwhrn\" (UID: \"b1ff925b-2293-4e8b-b59e-153ae7a518d3\") " pod="openstack/barbican-api-58586c84c4-nwhrn" Nov 21 14:32:13 crc kubenswrapper[4897]: I1121 14:32:13.053436 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b1ff925b-2293-4e8b-b59e-153ae7a518d3-public-tls-certs\") pod \"barbican-api-58586c84c4-nwhrn\" (UID: \"b1ff925b-2293-4e8b-b59e-153ae7a518d3\") " pod="openstack/barbican-api-58586c84c4-nwhrn" Nov 21 14:32:13 crc kubenswrapper[4897]: I1121 14:32:13.055531 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b1ff925b-2293-4e8b-b59e-153ae7a518d3-config-data-custom\") pod \"barbican-api-58586c84c4-nwhrn\" (UID: \"b1ff925b-2293-4e8b-b59e-153ae7a518d3\") " pod="openstack/barbican-api-58586c84c4-nwhrn" Nov 21 14:32:13 crc kubenswrapper[4897]: I1121 14:32:13.055743 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1ff925b-2293-4e8b-b59e-153ae7a518d3-config-data\") pod \"barbican-api-58586c84c4-nwhrn\" (UID: \"b1ff925b-2293-4e8b-b59e-153ae7a518d3\") " pod="openstack/barbican-api-58586c84c4-nwhrn" Nov 21 14:32:13 crc kubenswrapper[4897]: I1121 14:32:13.068738 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fg2wm\" (UniqueName: \"kubernetes.io/projected/b1ff925b-2293-4e8b-b59e-153ae7a518d3-kube-api-access-fg2wm\") pod \"barbican-api-58586c84c4-nwhrn\" (UID: \"b1ff925b-2293-4e8b-b59e-153ae7a518d3\") " pod="openstack/barbican-api-58586c84c4-nwhrn" Nov 21 14:32:13 crc kubenswrapper[4897]: I1121 14:32:13.208186 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-58586c84c4-nwhrn" Nov 21 14:32:13 crc kubenswrapper[4897]: I1121 14:32:13.496963 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 21 14:32:13 crc kubenswrapper[4897]: W1121 14:32:13.500080 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0b659474_79bd_423b_9f81_1748dfa6682f.slice/crio-9d4776eb26e498b60c2e3574a8e1596335873122df1df8b80479b99a7c9f035f WatchSource:0}: Error finding container 9d4776eb26e498b60c2e3574a8e1596335873122df1df8b80479b99a7c9f035f: Status 404 returned error can't find the container with id 9d4776eb26e498b60c2e3574a8e1596335873122df1df8b80479b99a7c9f035f Nov 21 14:32:13 crc kubenswrapper[4897]: I1121 14:32:13.643187 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 21 14:32:13 crc kubenswrapper[4897]: W1121 14:32:13.649637 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3cc0bfd6_5357_416c_ad58_c3f08ca7d2b9.slice/crio-8e790ea119a34e94e015ee943a0b0c94723c96ae1ff86aacef0fe38f7a121ee0 WatchSource:0}: Error finding container 8e790ea119a34e94e015ee943a0b0c94723c96ae1ff86aacef0fe38f7a121ee0: Status 404 returned error can't find the container with id 8e790ea119a34e94e015ee943a0b0c94723c96ae1ff86aacef0fe38f7a121ee0 Nov 21 14:32:13 crc kubenswrapper[4897]: I1121 14:32:13.751294 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-58586c84c4-nwhrn"] Nov 21 14:32:14 crc kubenswrapper[4897]: I1121 14:32:14.011678 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-58586c84c4-nwhrn" event={"ID":"b1ff925b-2293-4e8b-b59e-153ae7a518d3","Type":"ContainerStarted","Data":"866b6292eb32d7797153b61f5da9c02a7f0dc5d79bdf42d807baae7611a9d8f0"} Nov 21 14:32:14 crc kubenswrapper[4897]: I1121 14:32:14.014226 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9","Type":"ContainerStarted","Data":"8e790ea119a34e94e015ee943a0b0c94723c96ae1ff86aacef0fe38f7a121ee0"} Nov 21 14:32:14 crc kubenswrapper[4897]: I1121 14:32:14.015980 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0b659474-79bd-423b-9f81-1748dfa6682f","Type":"ContainerStarted","Data":"9d4776eb26e498b60c2e3574a8e1596335873122df1df8b80479b99a7c9f035f"} Nov 21 14:32:15 crc kubenswrapper[4897]: I1121 14:32:15.603312 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 21 14:32:15 crc kubenswrapper[4897]: I1121 14:32:15.686630 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 21 14:32:16 crc kubenswrapper[4897]: I1121 14:32:16.066863 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-77b5d6c768-mnz58" event={"ID":"e20e1ca8-ff86-4836-8703-40370c65e892","Type":"ContainerStarted","Data":"45f6cc0b7f9f7810671048b3d03bce4dddd6016b4579a323d98b0ef0e7b3319e"} Nov 21 14:32:16 crc kubenswrapper[4897]: I1121 14:32:16.067692 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-77b5d6c768-mnz58" Nov 21 14:32:16 crc kubenswrapper[4897]: I1121 14:32:16.070139 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9","Type":"ContainerStarted","Data":"6e8fbf18d13f5bc74ad04dafdafe72362e9a74b3b5600242b60f2f414a7e99ef"} Nov 21 14:32:16 crc kubenswrapper[4897]: I1121 14:32:16.074680 4897 generic.go:334] "Generic (PLEG): container finished" podID="4b23c32f-2539-41a4-b994-eb4210356a13" containerID="311550bf15cef4574086c7e01f62b0ad80c20ddb0d2e386febbce832e4ff6449" exitCode=0 Nov 21 14:32:16 crc kubenswrapper[4897]: I1121 14:32:16.074842 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d66f584d7-hxs7n" event={"ID":"4b23c32f-2539-41a4-b994-eb4210356a13","Type":"ContainerDied","Data":"311550bf15cef4574086c7e01f62b0ad80c20ddb0d2e386febbce832e4ff6449"} Nov 21 14:32:16 crc kubenswrapper[4897]: I1121 14:32:16.083037 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0b659474-79bd-423b-9f81-1748dfa6682f","Type":"ContainerStarted","Data":"388dbb25a02a13435da1a25cfd5bff61ba7768252158dfca065ed5eb7fdbcfcd"} Nov 21 14:32:16 crc kubenswrapper[4897]: I1121 14:32:16.100361 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-77b5d6c768-mnz58" podStartSLOduration=7.100341704 podStartE2EDuration="7.100341704s" podCreationTimestamp="2025-11-21 14:32:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:32:16.091995271 +0000 UTC m=+1413.376588786" watchObservedRunningTime="2025-11-21 14:32:16.100341704 +0000 UTC m=+1413.384935179" Nov 21 14:32:16 crc kubenswrapper[4897]: I1121 14:32:16.135122 4897 generic.go:334] "Generic (PLEG): container finished" podID="9bc19a19-de48-43b0-bcba-12e1a44f354d" containerID="47fc6aa31cde67babb5f2a0c99947a38f69932a7eaad737d93dc34b1dfbcb39d" exitCode=0 Nov 21 14:32:16 crc kubenswrapper[4897]: I1121 14:32:16.186090 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-8974ffc64-c68b5" event={"ID":"129e5e44-512d-4d11-930d-a1026e28e61c","Type":"ContainerStarted","Data":"b46691fd55d1510a51940b160d010c013edabb2076c3db4ed8152c07616a26cd"} Nov 21 14:32:16 crc kubenswrapper[4897]: I1121 14:32:16.186134 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-58586c84c4-nwhrn" event={"ID":"b1ff925b-2293-4e8b-b59e-153ae7a518d3","Type":"ContainerStarted","Data":"a18902beda2258d02885ecd40a615926387f2e6959d0d4636f4945fae371821f"} Nov 21 14:32:16 crc kubenswrapper[4897]: I1121 14:32:16.186145 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f4f4768cf-drjj2" event={"ID":"9bc19a19-de48-43b0-bcba-12e1a44f354d","Type":"ContainerDied","Data":"47fc6aa31cde67babb5f2a0c99947a38f69932a7eaad737d93dc34b1dfbcb39d"} Nov 21 14:32:16 crc kubenswrapper[4897]: I1121 14:32:16.733726 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f4f4768cf-drjj2" Nov 21 14:32:16 crc kubenswrapper[4897]: I1121 14:32:16.777661 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9bc19a19-de48-43b0-bcba-12e1a44f354d-dns-svc\") pod \"9bc19a19-de48-43b0-bcba-12e1a44f354d\" (UID: \"9bc19a19-de48-43b0-bcba-12e1a44f354d\") " Nov 21 14:32:16 crc kubenswrapper[4897]: I1121 14:32:16.777729 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9bc19a19-de48-43b0-bcba-12e1a44f354d-ovsdbserver-sb\") pod \"9bc19a19-de48-43b0-bcba-12e1a44f354d\" (UID: \"9bc19a19-de48-43b0-bcba-12e1a44f354d\") " Nov 21 14:32:16 crc kubenswrapper[4897]: I1121 14:32:16.777844 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9bc19a19-de48-43b0-bcba-12e1a44f354d-config\") pod \"9bc19a19-de48-43b0-bcba-12e1a44f354d\" (UID: \"9bc19a19-de48-43b0-bcba-12e1a44f354d\") " Nov 21 14:32:16 crc kubenswrapper[4897]: I1121 14:32:16.777898 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9bc19a19-de48-43b0-bcba-12e1a44f354d-ovsdbserver-nb\") pod \"9bc19a19-de48-43b0-bcba-12e1a44f354d\" (UID: \"9bc19a19-de48-43b0-bcba-12e1a44f354d\") " Nov 21 14:32:16 crc kubenswrapper[4897]: I1121 14:32:16.777964 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h7ksh\" (UniqueName: \"kubernetes.io/projected/9bc19a19-de48-43b0-bcba-12e1a44f354d-kube-api-access-h7ksh\") pod \"9bc19a19-de48-43b0-bcba-12e1a44f354d\" (UID: \"9bc19a19-de48-43b0-bcba-12e1a44f354d\") " Nov 21 14:32:16 crc kubenswrapper[4897]: I1121 14:32:16.778017 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9bc19a19-de48-43b0-bcba-12e1a44f354d-dns-swift-storage-0\") pod \"9bc19a19-de48-43b0-bcba-12e1a44f354d\" (UID: \"9bc19a19-de48-43b0-bcba-12e1a44f354d\") " Nov 21 14:32:16 crc kubenswrapper[4897]: I1121 14:32:16.800834 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9bc19a19-de48-43b0-bcba-12e1a44f354d-kube-api-access-h7ksh" (OuterVolumeSpecName: "kube-api-access-h7ksh") pod "9bc19a19-de48-43b0-bcba-12e1a44f354d" (UID: "9bc19a19-de48-43b0-bcba-12e1a44f354d"). InnerVolumeSpecName "kube-api-access-h7ksh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:32:16 crc kubenswrapper[4897]: I1121 14:32:16.810918 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9bc19a19-de48-43b0-bcba-12e1a44f354d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "9bc19a19-de48-43b0-bcba-12e1a44f354d" (UID: "9bc19a19-de48-43b0-bcba-12e1a44f354d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:32:16 crc kubenswrapper[4897]: I1121 14:32:16.820672 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9bc19a19-de48-43b0-bcba-12e1a44f354d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "9bc19a19-de48-43b0-bcba-12e1a44f354d" (UID: "9bc19a19-de48-43b0-bcba-12e1a44f354d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:32:16 crc kubenswrapper[4897]: I1121 14:32:16.852719 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9bc19a19-de48-43b0-bcba-12e1a44f354d-config" (OuterVolumeSpecName: "config") pod "9bc19a19-de48-43b0-bcba-12e1a44f354d" (UID: "9bc19a19-de48-43b0-bcba-12e1a44f354d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:32:16 crc kubenswrapper[4897]: I1121 14:32:16.853023 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9bc19a19-de48-43b0-bcba-12e1a44f354d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9bc19a19-de48-43b0-bcba-12e1a44f354d" (UID: "9bc19a19-de48-43b0-bcba-12e1a44f354d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:32:16 crc kubenswrapper[4897]: I1121 14:32:16.856871 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9bc19a19-de48-43b0-bcba-12e1a44f354d-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "9bc19a19-de48-43b0-bcba-12e1a44f354d" (UID: "9bc19a19-de48-43b0-bcba-12e1a44f354d"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:32:16 crc kubenswrapper[4897]: I1121 14:32:16.880843 4897 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9bc19a19-de48-43b0-bcba-12e1a44f354d-config\") on node \"crc\" DevicePath \"\"" Nov 21 14:32:16 crc kubenswrapper[4897]: I1121 14:32:16.881170 4897 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9bc19a19-de48-43b0-bcba-12e1a44f354d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 21 14:32:16 crc kubenswrapper[4897]: I1121 14:32:16.881184 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h7ksh\" (UniqueName: \"kubernetes.io/projected/9bc19a19-de48-43b0-bcba-12e1a44f354d-kube-api-access-h7ksh\") on node \"crc\" DevicePath \"\"" Nov 21 14:32:16 crc kubenswrapper[4897]: I1121 14:32:16.881192 4897 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9bc19a19-de48-43b0-bcba-12e1a44f354d-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 21 14:32:16 crc kubenswrapper[4897]: I1121 14:32:16.881200 4897 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9bc19a19-de48-43b0-bcba-12e1a44f354d-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 21 14:32:16 crc kubenswrapper[4897]: I1121 14:32:16.881209 4897 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9bc19a19-de48-43b0-bcba-12e1a44f354d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 21 14:32:17 crc kubenswrapper[4897]: I1121 14:32:17.148441 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f4f4768cf-drjj2" Nov 21 14:32:17 crc kubenswrapper[4897]: I1121 14:32:17.148434 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f4f4768cf-drjj2" event={"ID":"9bc19a19-de48-43b0-bcba-12e1a44f354d","Type":"ContainerDied","Data":"4920c48403baa106b6c967e1feda3a0d011ac6b153c90649f65c740a04533576"} Nov 21 14:32:17 crc kubenswrapper[4897]: I1121 14:32:17.148655 4897 scope.go:117] "RemoveContainer" containerID="47fc6aa31cde67babb5f2a0c99947a38f69932a7eaad737d93dc34b1dfbcb39d" Nov 21 14:32:17 crc kubenswrapper[4897]: I1121 14:32:17.152806 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-78cfd6c7f4-mzcx9" event={"ID":"2fda0597-e34c-4e8b-a600-c919f01be4b1","Type":"ContainerStarted","Data":"5f1d19bd9fe7bda7a3e07c2b27a9eb0b6bcd3c942d2b28ab2452ea7c90c9b03c"} Nov 21 14:32:17 crc kubenswrapper[4897]: I1121 14:32:17.216848 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f4f4768cf-drjj2"] Nov 21 14:32:17 crc kubenswrapper[4897]: I1121 14:32:17.226319 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-f4f4768cf-drjj2"] Nov 21 14:32:18 crc kubenswrapper[4897]: I1121 14:32:18.102040 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9bc19a19-de48-43b0-bcba-12e1a44f354d" path="/var/lib/kubelet/pods/9bc19a19-de48-43b0-bcba-12e1a44f354d/volumes" Nov 21 14:32:18 crc kubenswrapper[4897]: I1121 14:32:18.165233 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-58586c84c4-nwhrn" event={"ID":"b1ff925b-2293-4e8b-b59e-153ae7a518d3","Type":"ContainerStarted","Data":"0f71243b3dfe07d87ad7fc963a0437b392a220021754d5c5be89ec70d6b8795e"} Nov 21 14:32:18 crc kubenswrapper[4897]: I1121 14:32:18.165399 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-58586c84c4-nwhrn" Nov 21 14:32:18 crc kubenswrapper[4897]: I1121 14:32:18.165422 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-58586c84c4-nwhrn" Nov 21 14:32:18 crc kubenswrapper[4897]: I1121 14:32:18.167802 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9","Type":"ContainerStarted","Data":"4fa89c8d9ab0c8e911501ea40d04c77dfcf1f6332ae47d538db3cb9f8de12e45"} Nov 21 14:32:18 crc kubenswrapper[4897]: I1121 14:32:18.167863 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9" containerName="glance-log" containerID="cri-o://6e8fbf18d13f5bc74ad04dafdafe72362e9a74b3b5600242b60f2f414a7e99ef" gracePeriod=30 Nov 21 14:32:18 crc kubenswrapper[4897]: I1121 14:32:18.167894 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9" containerName="glance-httpd" containerID="cri-o://4fa89c8d9ab0c8e911501ea40d04c77dfcf1f6332ae47d538db3cb9f8de12e45" gracePeriod=30 Nov 21 14:32:18 crc kubenswrapper[4897]: I1121 14:32:18.170296 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d66f584d7-hxs7n" event={"ID":"4b23c32f-2539-41a4-b994-eb4210356a13","Type":"ContainerStarted","Data":"be5d18a893b03eba20c0f6a73a0c3f89b6a9c8d640286ef155ce37dffd3a7719"} Nov 21 14:32:18 crc kubenswrapper[4897]: I1121 14:32:18.170360 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6d66f584d7-hxs7n" Nov 21 14:32:18 crc kubenswrapper[4897]: I1121 14:32:18.172101 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0b659474-79bd-423b-9f81-1748dfa6682f","Type":"ContainerStarted","Data":"feff9dd89fce21af31b8707e4bd4156615c4e36a387f9a86b269d9ee4cd6c235"} Nov 21 14:32:18 crc kubenswrapper[4897]: I1121 14:32:18.172259 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="0b659474-79bd-423b-9f81-1748dfa6682f" containerName="glance-httpd" containerID="cri-o://feff9dd89fce21af31b8707e4bd4156615c4e36a387f9a86b269d9ee4cd6c235" gracePeriod=30 Nov 21 14:32:18 crc kubenswrapper[4897]: I1121 14:32:18.172222 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="0b659474-79bd-423b-9f81-1748dfa6682f" containerName="glance-log" containerID="cri-o://388dbb25a02a13435da1a25cfd5bff61ba7768252158dfca065ed5eb7fdbcfcd" gracePeriod=30 Nov 21 14:32:18 crc kubenswrapper[4897]: I1121 14:32:18.174820 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-8974ffc64-c68b5" event={"ID":"129e5e44-512d-4d11-930d-a1026e28e61c","Type":"ContainerStarted","Data":"3e36e37950ff801680e9f4a38accd4a0c7fbf204ff9c5c0dea35909d3469f550"} Nov 21 14:32:18 crc kubenswrapper[4897]: I1121 14:32:18.175297 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-8974ffc64-c68b5" Nov 21 14:32:18 crc kubenswrapper[4897]: I1121 14:32:18.175433 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-8974ffc64-c68b5" Nov 21 14:32:18 crc kubenswrapper[4897]: I1121 14:32:18.177131 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-7dvt9" event={"ID":"55644d8b-b8a3-463e-8045-b7391e9c2419","Type":"ContainerStarted","Data":"e7ac0248e2b28f323c61ff8d85c2fd0aaf30199090d2984738889239f670d176"} Nov 21 14:32:18 crc kubenswrapper[4897]: I1121 14:32:18.177306 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-78cfd6c7f4-mzcx9" Nov 21 14:32:18 crc kubenswrapper[4897]: I1121 14:32:18.177327 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-78cfd6c7f4-mzcx9" Nov 21 14:32:18 crc kubenswrapper[4897]: I1121 14:32:18.232300 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6d66f584d7-hxs7n" podStartSLOduration=7.23228086 podStartE2EDuration="7.23228086s" podCreationTimestamp="2025-11-21 14:32:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:32:18.225137259 +0000 UTC m=+1415.509730744" watchObservedRunningTime="2025-11-21 14:32:18.23228086 +0000 UTC m=+1415.516874335" Nov 21 14:32:18 crc kubenswrapper[4897]: I1121 14:32:18.239807 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-58586c84c4-nwhrn" podStartSLOduration=6.239788141 podStartE2EDuration="6.239788141s" podCreationTimestamp="2025-11-21 14:32:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:32:18.203737054 +0000 UTC m=+1415.488330529" watchObservedRunningTime="2025-11-21 14:32:18.239788141 +0000 UTC m=+1415.524381616" Nov 21 14:32:18 crc kubenswrapper[4897]: I1121 14:32:18.486791 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-7dvt9" podStartSLOduration=4.015230584 podStartE2EDuration="1m16.486772417s" podCreationTimestamp="2025-11-21 14:31:02 +0000 UTC" firstStartedPulling="2025-11-21 14:31:04.060603315 +0000 UTC m=+1341.345196790" lastFinishedPulling="2025-11-21 14:32:16.532145148 +0000 UTC m=+1413.816738623" observedRunningTime="2025-11-21 14:32:18.251965518 +0000 UTC m=+1415.536559003" watchObservedRunningTime="2025-11-21 14:32:18.486772417 +0000 UTC m=+1415.771365892" Nov 21 14:32:18 crc kubenswrapper[4897]: I1121 14:32:18.494219 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=7.494202876 podStartE2EDuration="7.494202876s" podCreationTimestamp="2025-11-21 14:32:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:32:18.479484811 +0000 UTC m=+1415.764078286" watchObservedRunningTime="2025-11-21 14:32:18.494202876 +0000 UTC m=+1415.778796351" Nov 21 14:32:18 crc kubenswrapper[4897]: I1121 14:32:18.515114 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-78cfd6c7f4-mzcx9" podStartSLOduration=9.515096346 podStartE2EDuration="9.515096346s" podCreationTimestamp="2025-11-21 14:32:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:32:18.511845119 +0000 UTC m=+1415.796438594" watchObservedRunningTime="2025-11-21 14:32:18.515096346 +0000 UTC m=+1415.799689821" Nov 21 14:32:18 crc kubenswrapper[4897]: I1121 14:32:18.539918 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-8974ffc64-c68b5" podStartSLOduration=9.53984005 podStartE2EDuration="9.53984005s" podCreationTimestamp="2025-11-21 14:32:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:32:18.529055021 +0000 UTC m=+1415.813648496" watchObservedRunningTime="2025-11-21 14:32:18.53984005 +0000 UTC m=+1415.824433525" Nov 21 14:32:18 crc kubenswrapper[4897]: I1121 14:32:18.553912 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=7.553893607 podStartE2EDuration="7.553893607s" podCreationTimestamp="2025-11-21 14:32:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:32:18.546366885 +0000 UTC m=+1415.830960360" watchObservedRunningTime="2025-11-21 14:32:18.553893607 +0000 UTC m=+1415.838487082" Nov 21 14:32:19 crc kubenswrapper[4897]: I1121 14:32:19.188070 4897 generic.go:334] "Generic (PLEG): container finished" podID="0b659474-79bd-423b-9f81-1748dfa6682f" containerID="feff9dd89fce21af31b8707e4bd4156615c4e36a387f9a86b269d9ee4cd6c235" exitCode=143 Nov 21 14:32:19 crc kubenswrapper[4897]: I1121 14:32:19.188107 4897 generic.go:334] "Generic (PLEG): container finished" podID="0b659474-79bd-423b-9f81-1748dfa6682f" containerID="388dbb25a02a13435da1a25cfd5bff61ba7768252158dfca065ed5eb7fdbcfcd" exitCode=143 Nov 21 14:32:19 crc kubenswrapper[4897]: I1121 14:32:19.188120 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0b659474-79bd-423b-9f81-1748dfa6682f","Type":"ContainerDied","Data":"feff9dd89fce21af31b8707e4bd4156615c4e36a387f9a86b269d9ee4cd6c235"} Nov 21 14:32:19 crc kubenswrapper[4897]: I1121 14:32:19.188175 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0b659474-79bd-423b-9f81-1748dfa6682f","Type":"ContainerDied","Data":"388dbb25a02a13435da1a25cfd5bff61ba7768252158dfca065ed5eb7fdbcfcd"} Nov 21 14:32:19 crc kubenswrapper[4897]: I1121 14:32:19.191005 4897 generic.go:334] "Generic (PLEG): container finished" podID="3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9" containerID="4fa89c8d9ab0c8e911501ea40d04c77dfcf1f6332ae47d538db3cb9f8de12e45" exitCode=143 Nov 21 14:32:19 crc kubenswrapper[4897]: I1121 14:32:19.191035 4897 generic.go:334] "Generic (PLEG): container finished" podID="3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9" containerID="6e8fbf18d13f5bc74ad04dafdafe72362e9a74b3b5600242b60f2f414a7e99ef" exitCode=143 Nov 21 14:32:19 crc kubenswrapper[4897]: I1121 14:32:19.191152 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9","Type":"ContainerDied","Data":"4fa89c8d9ab0c8e911501ea40d04c77dfcf1f6332ae47d538db3cb9f8de12e45"} Nov 21 14:32:19 crc kubenswrapper[4897]: I1121 14:32:19.191210 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9","Type":"ContainerDied","Data":"6e8fbf18d13f5bc74ad04dafdafe72362e9a74b3b5600242b60f2f414a7e99ef"} Nov 21 14:32:23 crc kubenswrapper[4897]: I1121 14:32:23.220001 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-78cfd6c7f4-mzcx9" podUID="2fda0597-e34c-4e8b-a600-c919f01be4b1" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.188:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 21 14:32:23 crc kubenswrapper[4897]: I1121 14:32:23.597771 4897 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-78cfd6c7f4-mzcx9" podUID="2fda0597-e34c-4e8b-a600-c919f01be4b1" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.188:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 21 14:32:24 crc kubenswrapper[4897]: I1121 14:32:24.555733 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 21 14:32:24 crc kubenswrapper[4897]: I1121 14:32:24.566268 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 21 14:32:24 crc kubenswrapper[4897]: I1121 14:32:24.675368 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0b659474-79bd-423b-9f81-1748dfa6682f-logs\") pod \"0b659474-79bd-423b-9f81-1748dfa6682f\" (UID: \"0b659474-79bd-423b-9f81-1748dfa6682f\") " Nov 21 14:32:24 crc kubenswrapper[4897]: I1121 14:32:24.675438 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9-logs\") pod \"3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9\" (UID: \"3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9\") " Nov 21 14:32:24 crc kubenswrapper[4897]: I1121 14:32:24.675470 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9\" (UID: \"3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9\") " Nov 21 14:32:24 crc kubenswrapper[4897]: I1121 14:32:24.675692 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p97h9\" (UniqueName: \"kubernetes.io/projected/3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9-kube-api-access-p97h9\") pod \"3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9\" (UID: \"3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9\") " Nov 21 14:32:24 crc kubenswrapper[4897]: I1121 14:32:24.675754 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b659474-79bd-423b-9f81-1748dfa6682f-config-data\") pod \"0b659474-79bd-423b-9f81-1748dfa6682f\" (UID: \"0b659474-79bd-423b-9f81-1748dfa6682f\") " Nov 21 14:32:24 crc kubenswrapper[4897]: I1121 14:32:24.675809 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9-config-data\") pod \"3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9\" (UID: \"3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9\") " Nov 21 14:32:24 crc kubenswrapper[4897]: I1121 14:32:24.675834 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9-httpd-run\") pod \"3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9\" (UID: \"3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9\") " Nov 21 14:32:24 crc kubenswrapper[4897]: I1121 14:32:24.675871 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0b659474-79bd-423b-9f81-1748dfa6682f-scripts\") pod \"0b659474-79bd-423b-9f81-1748dfa6682f\" (UID: \"0b659474-79bd-423b-9f81-1748dfa6682f\") " Nov 21 14:32:24 crc kubenswrapper[4897]: I1121 14:32:24.675913 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2t86t\" (UniqueName: \"kubernetes.io/projected/0b659474-79bd-423b-9f81-1748dfa6682f-kube-api-access-2t86t\") pod \"0b659474-79bd-423b-9f81-1748dfa6682f\" (UID: \"0b659474-79bd-423b-9f81-1748dfa6682f\") " Nov 21 14:32:24 crc kubenswrapper[4897]: I1121 14:32:24.675938 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9-scripts\") pod \"3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9\" (UID: \"3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9\") " Nov 21 14:32:24 crc kubenswrapper[4897]: I1121 14:32:24.675958 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9-combined-ca-bundle\") pod \"3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9\" (UID: \"3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9\") " Nov 21 14:32:24 crc kubenswrapper[4897]: I1121 14:32:24.675986 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b659474-79bd-423b-9f81-1748dfa6682f-combined-ca-bundle\") pod \"0b659474-79bd-423b-9f81-1748dfa6682f\" (UID: \"0b659474-79bd-423b-9f81-1748dfa6682f\") " Nov 21 14:32:24 crc kubenswrapper[4897]: I1121 14:32:24.676015 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0b659474-79bd-423b-9f81-1748dfa6682f-httpd-run\") pod \"0b659474-79bd-423b-9f81-1748dfa6682f\" (UID: \"0b659474-79bd-423b-9f81-1748dfa6682f\") " Nov 21 14:32:24 crc kubenswrapper[4897]: I1121 14:32:24.676043 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"0b659474-79bd-423b-9f81-1748dfa6682f\" (UID: \"0b659474-79bd-423b-9f81-1748dfa6682f\") " Nov 21 14:32:24 crc kubenswrapper[4897]: I1121 14:32:24.677967 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9" (UID: "3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:32:24 crc kubenswrapper[4897]: I1121 14:32:24.679061 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9-logs" (OuterVolumeSpecName: "logs") pod "3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9" (UID: "3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:32:24 crc kubenswrapper[4897]: I1121 14:32:24.682893 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0b659474-79bd-423b-9f81-1748dfa6682f-logs" (OuterVolumeSpecName: "logs") pod "0b659474-79bd-423b-9f81-1748dfa6682f" (UID: "0b659474-79bd-423b-9f81-1748dfa6682f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:32:24 crc kubenswrapper[4897]: I1121 14:32:24.684722 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "glance") pod "3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9" (UID: "3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 21 14:32:24 crc kubenswrapper[4897]: I1121 14:32:24.685829 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0b659474-79bd-423b-9f81-1748dfa6682f-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "0b659474-79bd-423b-9f81-1748dfa6682f" (UID: "0b659474-79bd-423b-9f81-1748dfa6682f"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:32:24 crc kubenswrapper[4897]: I1121 14:32:24.691065 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9-kube-api-access-p97h9" (OuterVolumeSpecName: "kube-api-access-p97h9") pod "3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9" (UID: "3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9"). InnerVolumeSpecName "kube-api-access-p97h9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:32:24 crc kubenswrapper[4897]: I1121 14:32:24.694985 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b659474-79bd-423b-9f81-1748dfa6682f-scripts" (OuterVolumeSpecName: "scripts") pod "0b659474-79bd-423b-9f81-1748dfa6682f" (UID: "0b659474-79bd-423b-9f81-1748dfa6682f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:32:24 crc kubenswrapper[4897]: I1121 14:32:24.703985 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9-scripts" (OuterVolumeSpecName: "scripts") pod "3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9" (UID: "3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:32:24 crc kubenswrapper[4897]: I1121 14:32:24.704036 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b659474-79bd-423b-9f81-1748dfa6682f-kube-api-access-2t86t" (OuterVolumeSpecName: "kube-api-access-2t86t") pod "0b659474-79bd-423b-9f81-1748dfa6682f" (UID: "0b659474-79bd-423b-9f81-1748dfa6682f"). InnerVolumeSpecName "kube-api-access-2t86t". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:32:24 crc kubenswrapper[4897]: I1121 14:32:24.712992 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance") pod "0b659474-79bd-423b-9f81-1748dfa6682f" (UID: "0b659474-79bd-423b-9f81-1748dfa6682f"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 21 14:32:24 crc kubenswrapper[4897]: I1121 14:32:24.754076 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9" (UID: "3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:32:24 crc kubenswrapper[4897]: I1121 14:32:24.770861 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b659474-79bd-423b-9f81-1748dfa6682f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0b659474-79bd-423b-9f81-1748dfa6682f" (UID: "0b659474-79bd-423b-9f81-1748dfa6682f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:32:24 crc kubenswrapper[4897]: I1121 14:32:24.773479 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9-config-data" (OuterVolumeSpecName: "config-data") pod "3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9" (UID: "3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:32:24 crc kubenswrapper[4897]: I1121 14:32:24.780003 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p97h9\" (UniqueName: \"kubernetes.io/projected/3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9-kube-api-access-p97h9\") on node \"crc\" DevicePath \"\"" Nov 21 14:32:24 crc kubenswrapper[4897]: I1121 14:32:24.780035 4897 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9-config-data\") on node \"crc\" DevicePath \"\"" Nov 21 14:32:24 crc kubenswrapper[4897]: I1121 14:32:24.780050 4897 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 21 14:32:24 crc kubenswrapper[4897]: I1121 14:32:24.780061 4897 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0b659474-79bd-423b-9f81-1748dfa6682f-scripts\") on node \"crc\" DevicePath \"\"" Nov 21 14:32:24 crc kubenswrapper[4897]: I1121 14:32:24.780074 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2t86t\" (UniqueName: \"kubernetes.io/projected/0b659474-79bd-423b-9f81-1748dfa6682f-kube-api-access-2t86t\") on node \"crc\" DevicePath \"\"" Nov 21 14:32:24 crc kubenswrapper[4897]: I1121 14:32:24.780083 4897 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9-scripts\") on node \"crc\" DevicePath \"\"" Nov 21 14:32:24 crc kubenswrapper[4897]: I1121 14:32:24.780113 4897 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 21 14:32:24 crc kubenswrapper[4897]: I1121 14:32:24.780121 4897 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b659474-79bd-423b-9f81-1748dfa6682f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 21 14:32:24 crc kubenswrapper[4897]: I1121 14:32:24.780129 4897 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0b659474-79bd-423b-9f81-1748dfa6682f-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 21 14:32:24 crc kubenswrapper[4897]: I1121 14:32:24.780151 4897 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Nov 21 14:32:24 crc kubenswrapper[4897]: I1121 14:32:24.780161 4897 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0b659474-79bd-423b-9f81-1748dfa6682f-logs\") on node \"crc\" DevicePath \"\"" Nov 21 14:32:24 crc kubenswrapper[4897]: I1121 14:32:24.780172 4897 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9-logs\") on node \"crc\" DevicePath \"\"" Nov 21 14:32:24 crc kubenswrapper[4897]: I1121 14:32:24.780188 4897 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Nov 21 14:32:24 crc kubenswrapper[4897]: I1121 14:32:24.803822 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b659474-79bd-423b-9f81-1748dfa6682f-config-data" (OuterVolumeSpecName: "config-data") pod "0b659474-79bd-423b-9f81-1748dfa6682f" (UID: "0b659474-79bd-423b-9f81-1748dfa6682f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:32:24 crc kubenswrapper[4897]: I1121 14:32:24.805501 4897 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Nov 21 14:32:24 crc kubenswrapper[4897]: I1121 14:32:24.826484 4897 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Nov 21 14:32:24 crc kubenswrapper[4897]: I1121 14:32:24.882392 4897 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b659474-79bd-423b-9f81-1748dfa6682f-config-data\") on node \"crc\" DevicePath \"\"" Nov 21 14:32:24 crc kubenswrapper[4897]: I1121 14:32:24.882425 4897 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Nov 21 14:32:24 crc kubenswrapper[4897]: I1121 14:32:24.882435 4897 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Nov 21 14:32:25 crc kubenswrapper[4897]: I1121 14:32:25.151842 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-58586c84c4-nwhrn" Nov 21 14:32:25 crc kubenswrapper[4897]: I1121 14:32:25.215987 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-78cfd6c7f4-mzcx9" Nov 21 14:32:25 crc kubenswrapper[4897]: I1121 14:32:25.227132 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-58586c84c4-nwhrn" Nov 21 14:32:25 crc kubenswrapper[4897]: I1121 14:32:25.287814 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9","Type":"ContainerDied","Data":"8e790ea119a34e94e015ee943a0b0c94723c96ae1ff86aacef0fe38f7a121ee0"} Nov 21 14:32:25 crc kubenswrapper[4897]: I1121 14:32:25.287865 4897 scope.go:117] "RemoveContainer" containerID="4fa89c8d9ab0c8e911501ea40d04c77dfcf1f6332ae47d538db3cb9f8de12e45" Nov 21 14:32:25 crc kubenswrapper[4897]: I1121 14:32:25.288041 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 21 14:32:25 crc kubenswrapper[4897]: I1121 14:32:25.319168 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 21 14:32:25 crc kubenswrapper[4897]: I1121 14:32:25.319585 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0b659474-79bd-423b-9f81-1748dfa6682f","Type":"ContainerDied","Data":"9d4776eb26e498b60c2e3574a8e1596335873122df1df8b80479b99a7c9f035f"} Nov 21 14:32:25 crc kubenswrapper[4897]: I1121 14:32:25.324778 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-78cfd6c7f4-mzcx9"] Nov 21 14:32:25 crc kubenswrapper[4897]: I1121 14:32:25.324900 4897 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 21 14:32:25 crc kubenswrapper[4897]: I1121 14:32:25.325162 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-78cfd6c7f4-mzcx9" podUID="2fda0597-e34c-4e8b-a600-c919f01be4b1" containerName="barbican-api" containerID="cri-o://5f1d19bd9fe7bda7a3e07c2b27a9eb0b6bcd3c942d2b28ab2452ea7c90c9b03c" gracePeriod=30 Nov 21 14:32:25 crc kubenswrapper[4897]: I1121 14:32:25.325857 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-78cfd6c7f4-mzcx9" podUID="2fda0597-e34c-4e8b-a600-c919f01be4b1" containerName="barbican-api-log" containerID="cri-o://3205cf41928a4172b4f181fe69f55ee8dd1a3bc5673e03408997ccab345fb4fb" gracePeriod=30 Nov 21 14:32:25 crc kubenswrapper[4897]: I1121 14:32:25.333185 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-78cfd6c7f4-mzcx9" podUID="2fda0597-e34c-4e8b-a600-c919f01be4b1" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.188:9311/healthcheck\": EOF" Nov 21 14:32:25 crc kubenswrapper[4897]: I1121 14:32:25.338735 4897 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-78cfd6c7f4-mzcx9" podUID="2fda0597-e34c-4e8b-a600-c919f01be4b1" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.188:9311/healthcheck\": EOF" Nov 21 14:32:25 crc kubenswrapper[4897]: I1121 14:32:25.339082 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-78cfd6c7f4-mzcx9" podUID="2fda0597-e34c-4e8b-a600-c919f01be4b1" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.188:9311/healthcheck\": EOF" Nov 21 14:32:25 crc kubenswrapper[4897]: I1121 14:32:25.339186 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-78cfd6c7f4-mzcx9" podUID="2fda0597-e34c-4e8b-a600-c919f01be4b1" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.188:9311/healthcheck\": EOF" Nov 21 14:32:25 crc kubenswrapper[4897]: I1121 14:32:25.348735 4897 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-78cfd6c7f4-mzcx9" podUID="2fda0597-e34c-4e8b-a600-c919f01be4b1" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.188:9311/healthcheck\": EOF" Nov 21 14:32:25 crc kubenswrapper[4897]: I1121 14:32:25.348836 4897 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/barbican-api-78cfd6c7f4-mzcx9" Nov 21 14:32:25 crc kubenswrapper[4897]: I1121 14:32:25.348942 4897 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-78cfd6c7f4-mzcx9" podUID="2fda0597-e34c-4e8b-a600-c919f01be4b1" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.188:9311/healthcheck\": EOF" Nov 21 14:32:25 crc kubenswrapper[4897]: I1121 14:32:25.364423 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 21 14:32:25 crc kubenswrapper[4897]: I1121 14:32:25.378867 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 21 14:32:25 crc kubenswrapper[4897]: I1121 14:32:25.398614 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 21 14:32:25 crc kubenswrapper[4897]: E1121 14:32:25.399083 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9bc19a19-de48-43b0-bcba-12e1a44f354d" containerName="init" Nov 21 14:32:25 crc kubenswrapper[4897]: I1121 14:32:25.399098 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="9bc19a19-de48-43b0-bcba-12e1a44f354d" containerName="init" Nov 21 14:32:25 crc kubenswrapper[4897]: E1121 14:32:25.399117 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9" containerName="glance-httpd" Nov 21 14:32:25 crc kubenswrapper[4897]: I1121 14:32:25.399124 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9" containerName="glance-httpd" Nov 21 14:32:25 crc kubenswrapper[4897]: E1121 14:32:25.399156 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9" containerName="glance-log" Nov 21 14:32:25 crc kubenswrapper[4897]: I1121 14:32:25.399162 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9" containerName="glance-log" Nov 21 14:32:25 crc kubenswrapper[4897]: E1121 14:32:25.399191 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b659474-79bd-423b-9f81-1748dfa6682f" containerName="glance-httpd" Nov 21 14:32:25 crc kubenswrapper[4897]: I1121 14:32:25.399198 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b659474-79bd-423b-9f81-1748dfa6682f" containerName="glance-httpd" Nov 21 14:32:25 crc kubenswrapper[4897]: E1121 14:32:25.399214 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b659474-79bd-423b-9f81-1748dfa6682f" containerName="glance-log" Nov 21 14:32:25 crc kubenswrapper[4897]: I1121 14:32:25.399220 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b659474-79bd-423b-9f81-1748dfa6682f" containerName="glance-log" Nov 21 14:32:25 crc kubenswrapper[4897]: I1121 14:32:25.399405 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9" containerName="glance-httpd" Nov 21 14:32:25 crc kubenswrapper[4897]: I1121 14:32:25.399422 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="9bc19a19-de48-43b0-bcba-12e1a44f354d" containerName="init" Nov 21 14:32:25 crc kubenswrapper[4897]: I1121 14:32:25.399448 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b659474-79bd-423b-9f81-1748dfa6682f" containerName="glance-log" Nov 21 14:32:25 crc kubenswrapper[4897]: I1121 14:32:25.399456 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9" containerName="glance-log" Nov 21 14:32:25 crc kubenswrapper[4897]: I1121 14:32:25.399467 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b659474-79bd-423b-9f81-1748dfa6682f" containerName="glance-httpd" Nov 21 14:32:25 crc kubenswrapper[4897]: I1121 14:32:25.409697 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 21 14:32:25 crc kubenswrapper[4897]: I1121 14:32:25.412642 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 21 14:32:25 crc kubenswrapper[4897]: I1121 14:32:25.418650 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-wk7cs" Nov 21 14:32:25 crc kubenswrapper[4897]: I1121 14:32:25.418885 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Nov 21 14:32:25 crc kubenswrapper[4897]: I1121 14:32:25.419155 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 21 14:32:25 crc kubenswrapper[4897]: I1121 14:32:25.419331 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Nov 21 14:32:25 crc kubenswrapper[4897]: I1121 14:32:25.496919 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756-logs\") pod \"glance-default-internal-api-0\" (UID: \"cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756\") " pod="openstack/glance-default-internal-api-0" Nov 21 14:32:25 crc kubenswrapper[4897]: I1121 14:32:25.496965 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756-scripts\") pod \"glance-default-internal-api-0\" (UID: \"cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756\") " pod="openstack/glance-default-internal-api-0" Nov 21 14:32:25 crc kubenswrapper[4897]: I1121 14:32:25.497119 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756\") " pod="openstack/glance-default-internal-api-0" Nov 21 14:32:25 crc kubenswrapper[4897]: I1121 14:32:25.497150 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ct675\" (UniqueName: \"kubernetes.io/projected/cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756-kube-api-access-ct675\") pod \"glance-default-internal-api-0\" (UID: \"cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756\") " pod="openstack/glance-default-internal-api-0" Nov 21 14:32:25 crc kubenswrapper[4897]: I1121 14:32:25.497184 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756\") " pod="openstack/glance-default-internal-api-0" Nov 21 14:32:25 crc kubenswrapper[4897]: I1121 14:32:25.497230 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756-config-data\") pod \"glance-default-internal-api-0\" (UID: \"cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756\") " pod="openstack/glance-default-internal-api-0" Nov 21 14:32:25 crc kubenswrapper[4897]: I1121 14:32:25.497255 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756\") " pod="openstack/glance-default-internal-api-0" Nov 21 14:32:25 crc kubenswrapper[4897]: I1121 14:32:25.497273 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756\") " pod="openstack/glance-default-internal-api-0" Nov 21 14:32:25 crc kubenswrapper[4897]: I1121 14:32:25.599891 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756\") " pod="openstack/glance-default-internal-api-0" Nov 21 14:32:25 crc kubenswrapper[4897]: I1121 14:32:25.601039 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756\") " pod="openstack/glance-default-internal-api-0" Nov 21 14:32:25 crc kubenswrapper[4897]: I1121 14:32:25.602278 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ct675\" (UniqueName: \"kubernetes.io/projected/cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756-kube-api-access-ct675\") pod \"glance-default-internal-api-0\" (UID: \"cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756\") " pod="openstack/glance-default-internal-api-0" Nov 21 14:32:25 crc kubenswrapper[4897]: I1121 14:32:25.602354 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756\") " pod="openstack/glance-default-internal-api-0" Nov 21 14:32:25 crc kubenswrapper[4897]: I1121 14:32:25.602642 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756-config-data\") pod \"glance-default-internal-api-0\" (UID: \"cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756\") " pod="openstack/glance-default-internal-api-0" Nov 21 14:32:25 crc kubenswrapper[4897]: I1121 14:32:25.602688 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756\") " pod="openstack/glance-default-internal-api-0" Nov 21 14:32:25 crc kubenswrapper[4897]: I1121 14:32:25.602720 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756\") " pod="openstack/glance-default-internal-api-0" Nov 21 14:32:25 crc kubenswrapper[4897]: I1121 14:32:25.602769 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756-logs\") pod \"glance-default-internal-api-0\" (UID: \"cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756\") " pod="openstack/glance-default-internal-api-0" Nov 21 14:32:25 crc kubenswrapper[4897]: I1121 14:32:25.602803 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756-scripts\") pod \"glance-default-internal-api-0\" (UID: \"cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756\") " pod="openstack/glance-default-internal-api-0" Nov 21 14:32:25 crc kubenswrapper[4897]: I1121 14:32:25.605192 4897 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/glance-default-internal-api-0" Nov 21 14:32:25 crc kubenswrapper[4897]: I1121 14:32:25.606688 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756-logs\") pod \"glance-default-internal-api-0\" (UID: \"cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756\") " pod="openstack/glance-default-internal-api-0" Nov 21 14:32:25 crc kubenswrapper[4897]: I1121 14:32:25.619275 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756\") " pod="openstack/glance-default-internal-api-0" Nov 21 14:32:25 crc kubenswrapper[4897]: I1121 14:32:25.631787 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 21 14:32:25 crc kubenswrapper[4897]: I1121 14:32:25.632255 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756-scripts\") pod \"glance-default-internal-api-0\" (UID: \"cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756\") " pod="openstack/glance-default-internal-api-0" Nov 21 14:32:25 crc kubenswrapper[4897]: I1121 14:32:25.639843 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756\") " pod="openstack/glance-default-internal-api-0" Nov 21 14:32:25 crc kubenswrapper[4897]: I1121 14:32:25.642811 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756-config-data\") pod \"glance-default-internal-api-0\" (UID: \"cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756\") " pod="openstack/glance-default-internal-api-0" Nov 21 14:32:25 crc kubenswrapper[4897]: I1121 14:32:25.668713 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 21 14:32:25 crc kubenswrapper[4897]: I1121 14:32:25.675521 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ct675\" (UniqueName: \"kubernetes.io/projected/cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756-kube-api-access-ct675\") pod \"glance-default-internal-api-0\" (UID: \"cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756\") " pod="openstack/glance-default-internal-api-0" Nov 21 14:32:25 crc kubenswrapper[4897]: I1121 14:32:25.692686 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 21 14:32:25 crc kubenswrapper[4897]: I1121 14:32:25.700687 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 21 14:32:25 crc kubenswrapper[4897]: I1121 14:32:25.704886 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 21 14:32:25 crc kubenswrapper[4897]: I1121 14:32:25.706957 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 21 14:32:25 crc kubenswrapper[4897]: I1121 14:32:25.707107 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Nov 21 14:32:25 crc kubenswrapper[4897]: I1121 14:32:25.720724 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756\") " pod="openstack/glance-default-internal-api-0" Nov 21 14:32:25 crc kubenswrapper[4897]: I1121 14:32:25.796745 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 21 14:32:25 crc kubenswrapper[4897]: I1121 14:32:25.809296 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dpbk2\" (UniqueName: \"kubernetes.io/projected/7155191e-f228-4639-8a75-ed8ff3dce81a-kube-api-access-dpbk2\") pod \"glance-default-external-api-0\" (UID: \"7155191e-f228-4639-8a75-ed8ff3dce81a\") " pod="openstack/glance-default-external-api-0" Nov 21 14:32:25 crc kubenswrapper[4897]: I1121 14:32:25.809396 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7155191e-f228-4639-8a75-ed8ff3dce81a-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"7155191e-f228-4639-8a75-ed8ff3dce81a\") " pod="openstack/glance-default-external-api-0" Nov 21 14:32:25 crc kubenswrapper[4897]: I1121 14:32:25.809441 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7155191e-f228-4639-8a75-ed8ff3dce81a-logs\") pod \"glance-default-external-api-0\" (UID: \"7155191e-f228-4639-8a75-ed8ff3dce81a\") " pod="openstack/glance-default-external-api-0" Nov 21 14:32:25 crc kubenswrapper[4897]: I1121 14:32:25.809456 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7155191e-f228-4639-8a75-ed8ff3dce81a-scripts\") pod \"glance-default-external-api-0\" (UID: \"7155191e-f228-4639-8a75-ed8ff3dce81a\") " pod="openstack/glance-default-external-api-0" Nov 21 14:32:25 crc kubenswrapper[4897]: I1121 14:32:25.809475 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"7155191e-f228-4639-8a75-ed8ff3dce81a\") " pod="openstack/glance-default-external-api-0" Nov 21 14:32:25 crc kubenswrapper[4897]: I1121 14:32:25.809512 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7155191e-f228-4639-8a75-ed8ff3dce81a-config-data\") pod \"glance-default-external-api-0\" (UID: \"7155191e-f228-4639-8a75-ed8ff3dce81a\") " pod="openstack/glance-default-external-api-0" Nov 21 14:32:25 crc kubenswrapper[4897]: I1121 14:32:25.809555 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7155191e-f228-4639-8a75-ed8ff3dce81a-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"7155191e-f228-4639-8a75-ed8ff3dce81a\") " pod="openstack/glance-default-external-api-0" Nov 21 14:32:25 crc kubenswrapper[4897]: I1121 14:32:25.809590 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7155191e-f228-4639-8a75-ed8ff3dce81a-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"7155191e-f228-4639-8a75-ed8ff3dce81a\") " pod="openstack/glance-default-external-api-0" Nov 21 14:32:25 crc kubenswrapper[4897]: I1121 14:32:25.912177 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7155191e-f228-4639-8a75-ed8ff3dce81a-config-data\") pod \"glance-default-external-api-0\" (UID: \"7155191e-f228-4639-8a75-ed8ff3dce81a\") " pod="openstack/glance-default-external-api-0" Nov 21 14:32:25 crc kubenswrapper[4897]: I1121 14:32:25.912510 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7155191e-f228-4639-8a75-ed8ff3dce81a-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"7155191e-f228-4639-8a75-ed8ff3dce81a\") " pod="openstack/glance-default-external-api-0" Nov 21 14:32:25 crc kubenswrapper[4897]: I1121 14:32:25.912699 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7155191e-f228-4639-8a75-ed8ff3dce81a-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"7155191e-f228-4639-8a75-ed8ff3dce81a\") " pod="openstack/glance-default-external-api-0" Nov 21 14:32:25 crc kubenswrapper[4897]: I1121 14:32:25.912857 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dpbk2\" (UniqueName: \"kubernetes.io/projected/7155191e-f228-4639-8a75-ed8ff3dce81a-kube-api-access-dpbk2\") pod \"glance-default-external-api-0\" (UID: \"7155191e-f228-4639-8a75-ed8ff3dce81a\") " pod="openstack/glance-default-external-api-0" Nov 21 14:32:25 crc kubenswrapper[4897]: I1121 14:32:25.913130 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7155191e-f228-4639-8a75-ed8ff3dce81a-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"7155191e-f228-4639-8a75-ed8ff3dce81a\") " pod="openstack/glance-default-external-api-0" Nov 21 14:32:25 crc kubenswrapper[4897]: I1121 14:32:25.913306 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7155191e-f228-4639-8a75-ed8ff3dce81a-logs\") pod \"glance-default-external-api-0\" (UID: \"7155191e-f228-4639-8a75-ed8ff3dce81a\") " pod="openstack/glance-default-external-api-0" Nov 21 14:32:25 crc kubenswrapper[4897]: I1121 14:32:25.913413 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7155191e-f228-4639-8a75-ed8ff3dce81a-scripts\") pod \"glance-default-external-api-0\" (UID: \"7155191e-f228-4639-8a75-ed8ff3dce81a\") " pod="openstack/glance-default-external-api-0" Nov 21 14:32:25 crc kubenswrapper[4897]: I1121 14:32:25.913532 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"7155191e-f228-4639-8a75-ed8ff3dce81a\") " pod="openstack/glance-default-external-api-0" Nov 21 14:32:25 crc kubenswrapper[4897]: I1121 14:32:25.913748 4897 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"7155191e-f228-4639-8a75-ed8ff3dce81a\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/glance-default-external-api-0" Nov 21 14:32:25 crc kubenswrapper[4897]: I1121 14:32:25.914645 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7155191e-f228-4639-8a75-ed8ff3dce81a-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"7155191e-f228-4639-8a75-ed8ff3dce81a\") " pod="openstack/glance-default-external-api-0" Nov 21 14:32:25 crc kubenswrapper[4897]: I1121 14:32:25.914779 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7155191e-f228-4639-8a75-ed8ff3dce81a-logs\") pod \"glance-default-external-api-0\" (UID: \"7155191e-f228-4639-8a75-ed8ff3dce81a\") " pod="openstack/glance-default-external-api-0" Nov 21 14:32:25 crc kubenswrapper[4897]: I1121 14:32:25.920241 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7155191e-f228-4639-8a75-ed8ff3dce81a-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"7155191e-f228-4639-8a75-ed8ff3dce81a\") " pod="openstack/glance-default-external-api-0" Nov 21 14:32:25 crc kubenswrapper[4897]: I1121 14:32:25.926546 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7155191e-f228-4639-8a75-ed8ff3dce81a-config-data\") pod \"glance-default-external-api-0\" (UID: \"7155191e-f228-4639-8a75-ed8ff3dce81a\") " pod="openstack/glance-default-external-api-0" Nov 21 14:32:25 crc kubenswrapper[4897]: I1121 14:32:25.927380 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7155191e-f228-4639-8a75-ed8ff3dce81a-scripts\") pod \"glance-default-external-api-0\" (UID: \"7155191e-f228-4639-8a75-ed8ff3dce81a\") " pod="openstack/glance-default-external-api-0" Nov 21 14:32:25 crc kubenswrapper[4897]: I1121 14:32:25.931026 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7155191e-f228-4639-8a75-ed8ff3dce81a-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"7155191e-f228-4639-8a75-ed8ff3dce81a\") " pod="openstack/glance-default-external-api-0" Nov 21 14:32:25 crc kubenswrapper[4897]: I1121 14:32:25.936101 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dpbk2\" (UniqueName: \"kubernetes.io/projected/7155191e-f228-4639-8a75-ed8ff3dce81a-kube-api-access-dpbk2\") pod \"glance-default-external-api-0\" (UID: \"7155191e-f228-4639-8a75-ed8ff3dce81a\") " pod="openstack/glance-default-external-api-0" Nov 21 14:32:25 crc kubenswrapper[4897]: I1121 14:32:25.961838 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"7155191e-f228-4639-8a75-ed8ff3dce81a\") " pod="openstack/glance-default-external-api-0" Nov 21 14:32:26 crc kubenswrapper[4897]: I1121 14:32:26.108713 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 21 14:32:26 crc kubenswrapper[4897]: I1121 14:32:26.118954 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b659474-79bd-423b-9f81-1748dfa6682f" path="/var/lib/kubelet/pods/0b659474-79bd-423b-9f81-1748dfa6682f/volumes" Nov 21 14:32:26 crc kubenswrapper[4897]: I1121 14:32:26.120285 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9" path="/var/lib/kubelet/pods/3cc0bfd6-5357-416c-ad58-c3f08ca7d2b9/volumes" Nov 21 14:32:26 crc kubenswrapper[4897]: I1121 14:32:26.329980 4897 generic.go:334] "Generic (PLEG): container finished" podID="2fda0597-e34c-4e8b-a600-c919f01be4b1" containerID="3205cf41928a4172b4f181fe69f55ee8dd1a3bc5673e03408997ccab345fb4fb" exitCode=143 Nov 21 14:32:26 crc kubenswrapper[4897]: I1121 14:32:26.330082 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-78cfd6c7f4-mzcx9" event={"ID":"2fda0597-e34c-4e8b-a600-c919f01be4b1","Type":"ContainerDied","Data":"3205cf41928a4172b4f181fe69f55ee8dd1a3bc5673e03408997ccab345fb4fb"} Nov 21 14:32:26 crc kubenswrapper[4897]: I1121 14:32:26.663721 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6d66f584d7-hxs7n" Nov 21 14:32:26 crc kubenswrapper[4897]: I1121 14:32:26.758199 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-fcfdd6f9f-6l4ws"] Nov 21 14:32:26 crc kubenswrapper[4897]: I1121 14:32:26.758437 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-fcfdd6f9f-6l4ws" podUID="11f3e22f-051b-42ab-b70e-7813ecbbc8f5" containerName="dnsmasq-dns" containerID="cri-o://41766958aa12b4ccf753cc8c11168d6cd59f891bf4c5f073c6e21504d016ade5" gracePeriod=10 Nov 21 14:32:27 crc kubenswrapper[4897]: I1121 14:32:27.343474 4897 generic.go:334] "Generic (PLEG): container finished" podID="11f3e22f-051b-42ab-b70e-7813ecbbc8f5" containerID="41766958aa12b4ccf753cc8c11168d6cd59f891bf4c5f073c6e21504d016ade5" exitCode=0 Nov 21 14:32:27 crc kubenswrapper[4897]: I1121 14:32:27.343885 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-fcfdd6f9f-6l4ws" event={"ID":"11f3e22f-051b-42ab-b70e-7813ecbbc8f5","Type":"ContainerDied","Data":"41766958aa12b4ccf753cc8c11168d6cd59f891bf4c5f073c6e21504d016ade5"} Nov 21 14:32:27 crc kubenswrapper[4897]: I1121 14:32:27.633668 4897 scope.go:117] "RemoveContainer" containerID="6e8fbf18d13f5bc74ad04dafdafe72362e9a74b3b5600242b60f2f414a7e99ef" Nov 21 14:32:27 crc kubenswrapper[4897]: I1121 14:32:27.794903 4897 scope.go:117] "RemoveContainer" containerID="feff9dd89fce21af31b8707e4bd4156615c4e36a387f9a86b269d9ee4cd6c235" Nov 21 14:32:27 crc kubenswrapper[4897]: I1121 14:32:27.932127 4897 scope.go:117] "RemoveContainer" containerID="388dbb25a02a13435da1a25cfd5bff61ba7768252158dfca065ed5eb7fdbcfcd" Nov 21 14:32:27 crc kubenswrapper[4897]: I1121 14:32:27.963100 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-fcfdd6f9f-6l4ws" Nov 21 14:32:28 crc kubenswrapper[4897]: I1121 14:32:28.092346 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/11f3e22f-051b-42ab-b70e-7813ecbbc8f5-ovsdbserver-nb\") pod \"11f3e22f-051b-42ab-b70e-7813ecbbc8f5\" (UID: \"11f3e22f-051b-42ab-b70e-7813ecbbc8f5\") " Nov 21 14:32:28 crc kubenswrapper[4897]: I1121 14:32:28.092809 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/11f3e22f-051b-42ab-b70e-7813ecbbc8f5-dns-svc\") pod \"11f3e22f-051b-42ab-b70e-7813ecbbc8f5\" (UID: \"11f3e22f-051b-42ab-b70e-7813ecbbc8f5\") " Nov 21 14:32:28 crc kubenswrapper[4897]: I1121 14:32:28.092851 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mv89f\" (UniqueName: \"kubernetes.io/projected/11f3e22f-051b-42ab-b70e-7813ecbbc8f5-kube-api-access-mv89f\") pod \"11f3e22f-051b-42ab-b70e-7813ecbbc8f5\" (UID: \"11f3e22f-051b-42ab-b70e-7813ecbbc8f5\") " Nov 21 14:32:28 crc kubenswrapper[4897]: I1121 14:32:28.092927 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/11f3e22f-051b-42ab-b70e-7813ecbbc8f5-ovsdbserver-sb\") pod \"11f3e22f-051b-42ab-b70e-7813ecbbc8f5\" (UID: \"11f3e22f-051b-42ab-b70e-7813ecbbc8f5\") " Nov 21 14:32:28 crc kubenswrapper[4897]: I1121 14:32:28.094010 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/11f3e22f-051b-42ab-b70e-7813ecbbc8f5-dns-swift-storage-0\") pod \"11f3e22f-051b-42ab-b70e-7813ecbbc8f5\" (UID: \"11f3e22f-051b-42ab-b70e-7813ecbbc8f5\") " Nov 21 14:32:28 crc kubenswrapper[4897]: I1121 14:32:28.094075 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11f3e22f-051b-42ab-b70e-7813ecbbc8f5-config\") pod \"11f3e22f-051b-42ab-b70e-7813ecbbc8f5\" (UID: \"11f3e22f-051b-42ab-b70e-7813ecbbc8f5\") " Nov 21 14:32:28 crc kubenswrapper[4897]: I1121 14:32:28.113858 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/11f3e22f-051b-42ab-b70e-7813ecbbc8f5-kube-api-access-mv89f" (OuterVolumeSpecName: "kube-api-access-mv89f") pod "11f3e22f-051b-42ab-b70e-7813ecbbc8f5" (UID: "11f3e22f-051b-42ab-b70e-7813ecbbc8f5"). InnerVolumeSpecName "kube-api-access-mv89f". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:32:28 crc kubenswrapper[4897]: I1121 14:32:28.167859 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/11f3e22f-051b-42ab-b70e-7813ecbbc8f5-config" (OuterVolumeSpecName: "config") pod "11f3e22f-051b-42ab-b70e-7813ecbbc8f5" (UID: "11f3e22f-051b-42ab-b70e-7813ecbbc8f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:32:28 crc kubenswrapper[4897]: I1121 14:32:28.175359 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/11f3e22f-051b-42ab-b70e-7813ecbbc8f5-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "11f3e22f-051b-42ab-b70e-7813ecbbc8f5" (UID: "11f3e22f-051b-42ab-b70e-7813ecbbc8f5"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:32:28 crc kubenswrapper[4897]: I1121 14:32:28.179799 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/11f3e22f-051b-42ab-b70e-7813ecbbc8f5-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "11f3e22f-051b-42ab-b70e-7813ecbbc8f5" (UID: "11f3e22f-051b-42ab-b70e-7813ecbbc8f5"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:32:28 crc kubenswrapper[4897]: I1121 14:32:28.185525 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/11f3e22f-051b-42ab-b70e-7813ecbbc8f5-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "11f3e22f-051b-42ab-b70e-7813ecbbc8f5" (UID: "11f3e22f-051b-42ab-b70e-7813ecbbc8f5"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:32:28 crc kubenswrapper[4897]: I1121 14:32:28.189274 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/11f3e22f-051b-42ab-b70e-7813ecbbc8f5-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "11f3e22f-051b-42ab-b70e-7813ecbbc8f5" (UID: "11f3e22f-051b-42ab-b70e-7813ecbbc8f5"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:32:28 crc kubenswrapper[4897]: I1121 14:32:28.198438 4897 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/11f3e22f-051b-42ab-b70e-7813ecbbc8f5-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 21 14:32:28 crc kubenswrapper[4897]: I1121 14:32:28.198469 4897 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/11f3e22f-051b-42ab-b70e-7813ecbbc8f5-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 21 14:32:28 crc kubenswrapper[4897]: I1121 14:32:28.198478 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mv89f\" (UniqueName: \"kubernetes.io/projected/11f3e22f-051b-42ab-b70e-7813ecbbc8f5-kube-api-access-mv89f\") on node \"crc\" DevicePath \"\"" Nov 21 14:32:28 crc kubenswrapper[4897]: I1121 14:32:28.198488 4897 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/11f3e22f-051b-42ab-b70e-7813ecbbc8f5-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 21 14:32:28 crc kubenswrapper[4897]: I1121 14:32:28.198496 4897 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/11f3e22f-051b-42ab-b70e-7813ecbbc8f5-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 21 14:32:28 crc kubenswrapper[4897]: I1121 14:32:28.198508 4897 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11f3e22f-051b-42ab-b70e-7813ecbbc8f5-config\") on node \"crc\" DevicePath \"\"" Nov 21 14:32:28 crc kubenswrapper[4897]: I1121 14:32:28.312831 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 21 14:32:28 crc kubenswrapper[4897]: W1121 14:32:28.317242 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7155191e_f228_4639_8a75_ed8ff3dce81a.slice/crio-f46e3df66f856439d496964b0db493d27a14089e32fff851de3b1b81d913b3e4 WatchSource:0}: Error finding container f46e3df66f856439d496964b0db493d27a14089e32fff851de3b1b81d913b3e4: Status 404 returned error can't find the container with id f46e3df66f856439d496964b0db493d27a14089e32fff851de3b1b81d913b3e4 Nov 21 14:32:28 crc kubenswrapper[4897]: I1121 14:32:28.370225 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-fcfdd6f9f-6l4ws" event={"ID":"11f3e22f-051b-42ab-b70e-7813ecbbc8f5","Type":"ContainerDied","Data":"5ad095be28aed48a8934c0083373f12f0f4b4bea7b99c14921bb6920995de6b8"} Nov 21 14:32:28 crc kubenswrapper[4897]: I1121 14:32:28.370280 4897 scope.go:117] "RemoveContainer" containerID="41766958aa12b4ccf753cc8c11168d6cd59f891bf4c5f073c6e21504d016ade5" Nov 21 14:32:28 crc kubenswrapper[4897]: I1121 14:32:28.370496 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-fcfdd6f9f-6l4ws" Nov 21 14:32:28 crc kubenswrapper[4897]: I1121 14:32:28.376180 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-78dc44f895-5stt7" event={"ID":"1c554928-41d4-4b52-a59b-83411d77596c","Type":"ContainerStarted","Data":"4cc3fb6bd5e6b381c8a89a504a1b4c58e75d02610c0579791800a1bcb7a14d94"} Nov 21 14:32:28 crc kubenswrapper[4897]: I1121 14:32:28.384604 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-tbtcb" event={"ID":"59fd52b7-f25b-4a81-a961-41e6cb526c89","Type":"ContainerStarted","Data":"7e1b93554f3c8ad375016f453d6de46e6965cbc1cf41b512bb5c0432ea3f9e3d"} Nov 21 14:32:28 crc kubenswrapper[4897]: I1121 14:32:28.409169 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7c944bd776-fnlxn" event={"ID":"672ba474-ca76-40df-a93b-fa8d3cc91e94","Type":"ContainerStarted","Data":"d3c0bfa80f1a6be31ed0096c0d6e99e4340c44332e0210a5eb345ad95aedb5ee"} Nov 21 14:32:28 crc kubenswrapper[4897]: I1121 14:32:28.416561 4897 scope.go:117] "RemoveContainer" containerID="888737d2076501df7e2c323741a7989bc6466db72ef1a795ceedff5203a50639" Nov 21 14:32:28 crc kubenswrapper[4897]: I1121 14:32:28.446233 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-78dc44f895-5stt7" podStartSLOduration=2.249610133 podStartE2EDuration="19.446208862s" podCreationTimestamp="2025-11-21 14:32:09 +0000 UTC" firstStartedPulling="2025-11-21 14:32:10.311443928 +0000 UTC m=+1407.596037413" lastFinishedPulling="2025-11-21 14:32:27.508042677 +0000 UTC m=+1424.792636142" observedRunningTime="2025-11-21 14:32:28.399623171 +0000 UTC m=+1425.684216656" watchObservedRunningTime="2025-11-21 14:32:28.446208862 +0000 UTC m=+1425.730802337" Nov 21 14:32:28 crc kubenswrapper[4897]: I1121 14:32:28.451836 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-db-sync-tbtcb" podStartSLOduration=2.2878277799999998 podStartE2EDuration="1m26.451812342s" podCreationTimestamp="2025-11-21 14:31:02 +0000 UTC" firstStartedPulling="2025-11-21 14:31:03.353723264 +0000 UTC m=+1340.638316739" lastFinishedPulling="2025-11-21 14:32:27.517707826 +0000 UTC m=+1424.802301301" observedRunningTime="2025-11-21 14:32:28.425322831 +0000 UTC m=+1425.709916306" watchObservedRunningTime="2025-11-21 14:32:28.451812342 +0000 UTC m=+1425.736405817" Nov 21 14:32:28 crc kubenswrapper[4897]: I1121 14:32:28.457163 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"7155191e-f228-4639-8a75-ed8ff3dce81a","Type":"ContainerStarted","Data":"f46e3df66f856439d496964b0db493d27a14089e32fff851de3b1b81d913b3e4"} Nov 21 14:32:28 crc kubenswrapper[4897]: I1121 14:32:28.466569 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c3696f1d-db0b-4910-a65c-31410e2316df","Type":"ContainerStarted","Data":"5a54ebcb33a48f8815ea865396faddc9bb3678b376fb5464e8ddda342d748e77"} Nov 21 14:32:28 crc kubenswrapper[4897]: I1121 14:32:28.466713 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c3696f1d-db0b-4910-a65c-31410e2316df" containerName="ceilometer-central-agent" containerID="cri-o://4e82c69b36fc692afc0ff6d360048a4a9871804fa3ae5fb182d5567f08ec50d1" gracePeriod=30 Nov 21 14:32:28 crc kubenswrapper[4897]: I1121 14:32:28.466917 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 21 14:32:28 crc kubenswrapper[4897]: I1121 14:32:28.466980 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c3696f1d-db0b-4910-a65c-31410e2316df" containerName="proxy-httpd" containerID="cri-o://5a54ebcb33a48f8815ea865396faddc9bb3678b376fb5464e8ddda342d748e77" gracePeriod=30 Nov 21 14:32:28 crc kubenswrapper[4897]: I1121 14:32:28.467027 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c3696f1d-db0b-4910-a65c-31410e2316df" containerName="sg-core" containerID="cri-o://3c6c1e27ec72c0edf905c8a623e78bf432a92f75bf1c931a7707cf9faca2b35d" gracePeriod=30 Nov 21 14:32:28 crc kubenswrapper[4897]: I1121 14:32:28.467061 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c3696f1d-db0b-4910-a65c-31410e2316df" containerName="ceilometer-notification-agent" containerID="cri-o://b676d00cf16995ea43993b65d8f5b05023450400f02e4e944e5a3777a2e7e0a5" gracePeriod=30 Nov 21 14:32:28 crc kubenswrapper[4897]: I1121 14:32:28.479859 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-fcfdd6f9f-6l4ws"] Nov 21 14:32:28 crc kubenswrapper[4897]: I1121 14:32:28.516880 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-fcfdd6f9f-6l4ws"] Nov 21 14:32:28 crc kubenswrapper[4897]: I1121 14:32:28.528527 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-7c944bd776-fnlxn" podStartSLOduration=2.561931751 podStartE2EDuration="19.528496699s" podCreationTimestamp="2025-11-21 14:32:09 +0000 UTC" firstStartedPulling="2025-11-21 14:32:10.542434074 +0000 UTC m=+1407.827027549" lastFinishedPulling="2025-11-21 14:32:27.508999032 +0000 UTC m=+1424.793592497" observedRunningTime="2025-11-21 14:32:28.470626716 +0000 UTC m=+1425.755220201" watchObservedRunningTime="2025-11-21 14:32:28.528496699 +0000 UTC m=+1425.813090174" Nov 21 14:32:28 crc kubenswrapper[4897]: I1121 14:32:28.543577 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 21 14:32:28 crc kubenswrapper[4897]: I1121 14:32:28.552328 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.989227494 podStartE2EDuration="1m26.552305918s" podCreationTimestamp="2025-11-21 14:31:02 +0000 UTC" firstStartedPulling="2025-11-21 14:31:04.210383122 +0000 UTC m=+1341.494976587" lastFinishedPulling="2025-11-21 14:32:27.773461536 +0000 UTC m=+1425.058055011" observedRunningTime="2025-11-21 14:32:28.512932511 +0000 UTC m=+1425.797525986" watchObservedRunningTime="2025-11-21 14:32:28.552305918 +0000 UTC m=+1425.836899403" Nov 21 14:32:29 crc kubenswrapper[4897]: I1121 14:32:29.505318 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-78dc44f895-5stt7" event={"ID":"1c554928-41d4-4b52-a59b-83411d77596c","Type":"ContainerStarted","Data":"8a6077c1aab16885aa11c59b54c4b413087644e16191edac692da252536d6bfc"} Nov 21 14:32:29 crc kubenswrapper[4897]: I1121 14:32:29.512455 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7c944bd776-fnlxn" event={"ID":"672ba474-ca76-40df-a93b-fa8d3cc91e94","Type":"ContainerStarted","Data":"ceab68f8e812eb919933a5c40aa157cf1551de9c771688370bfcbe4ac7b305be"} Nov 21 14:32:29 crc kubenswrapper[4897]: I1121 14:32:29.514245 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756","Type":"ContainerStarted","Data":"fbffe9604199c2c6d9800e7f1823e0e982f185c924f967e3770744b01d341b13"} Nov 21 14:32:29 crc kubenswrapper[4897]: I1121 14:32:29.514287 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756","Type":"ContainerStarted","Data":"2d3f77279621bedaaaa2adc858cb4b8df929d9fc05eb69b73803167dbae2a405"} Nov 21 14:32:29 crc kubenswrapper[4897]: I1121 14:32:29.516002 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"7155191e-f228-4639-8a75-ed8ff3dce81a","Type":"ContainerStarted","Data":"e3f7f26e10be327c1a46809d507483a15eb0aca60ab79afe47a1fa3b5e10885f"} Nov 21 14:32:29 crc kubenswrapper[4897]: I1121 14:32:29.523140 4897 generic.go:334] "Generic (PLEG): container finished" podID="c3696f1d-db0b-4910-a65c-31410e2316df" containerID="5a54ebcb33a48f8815ea865396faddc9bb3678b376fb5464e8ddda342d748e77" exitCode=0 Nov 21 14:32:29 crc kubenswrapper[4897]: I1121 14:32:29.523182 4897 generic.go:334] "Generic (PLEG): container finished" podID="c3696f1d-db0b-4910-a65c-31410e2316df" containerID="3c6c1e27ec72c0edf905c8a623e78bf432a92f75bf1c931a7707cf9faca2b35d" exitCode=2 Nov 21 14:32:29 crc kubenswrapper[4897]: I1121 14:32:29.523193 4897 generic.go:334] "Generic (PLEG): container finished" podID="c3696f1d-db0b-4910-a65c-31410e2316df" containerID="4e82c69b36fc692afc0ff6d360048a4a9871804fa3ae5fb182d5567f08ec50d1" exitCode=0 Nov 21 14:32:29 crc kubenswrapper[4897]: I1121 14:32:29.523200 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c3696f1d-db0b-4910-a65c-31410e2316df","Type":"ContainerDied","Data":"5a54ebcb33a48f8815ea865396faddc9bb3678b376fb5464e8ddda342d748e77"} Nov 21 14:32:29 crc kubenswrapper[4897]: I1121 14:32:29.523241 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c3696f1d-db0b-4910-a65c-31410e2316df","Type":"ContainerDied","Data":"3c6c1e27ec72c0edf905c8a623e78bf432a92f75bf1c931a7707cf9faca2b35d"} Nov 21 14:32:29 crc kubenswrapper[4897]: I1121 14:32:29.523252 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c3696f1d-db0b-4910-a65c-31410e2316df","Type":"ContainerDied","Data":"4e82c69b36fc692afc0ff6d360048a4a9871804fa3ae5fb182d5567f08ec50d1"} Nov 21 14:32:30 crc kubenswrapper[4897]: I1121 14:32:30.107978 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="11f3e22f-051b-42ab-b70e-7813ecbbc8f5" path="/var/lib/kubelet/pods/11f3e22f-051b-42ab-b70e-7813ecbbc8f5/volumes" Nov 21 14:32:30 crc kubenswrapper[4897]: I1121 14:32:30.535870 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756","Type":"ContainerStarted","Data":"b40036e131ce501b96971cc68d11d31dd29c34c53797f7d0d45c993eaf89d3ec"} Nov 21 14:32:30 crc kubenswrapper[4897]: I1121 14:32:30.538835 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"7155191e-f228-4639-8a75-ed8ff3dce81a","Type":"ContainerStarted","Data":"dc5724ef2786b81a902bc7373794e6066c91ce62272f7e949534d19dee9f3fe0"} Nov 21 14:32:30 crc kubenswrapper[4897]: I1121 14:32:30.541751 4897 generic.go:334] "Generic (PLEG): container finished" podID="c3696f1d-db0b-4910-a65c-31410e2316df" containerID="b676d00cf16995ea43993b65d8f5b05023450400f02e4e944e5a3777a2e7e0a5" exitCode=0 Nov 21 14:32:30 crc kubenswrapper[4897]: I1121 14:32:30.541771 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c3696f1d-db0b-4910-a65c-31410e2316df","Type":"ContainerDied","Data":"b676d00cf16995ea43993b65d8f5b05023450400f02e4e944e5a3777a2e7e0a5"} Nov 21 14:32:30 crc kubenswrapper[4897]: I1121 14:32:30.563063 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=5.563040442 podStartE2EDuration="5.563040442s" podCreationTimestamp="2025-11-21 14:32:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:32:30.555480439 +0000 UTC m=+1427.840073914" watchObservedRunningTime="2025-11-21 14:32:30.563040442 +0000 UTC m=+1427.847633917" Nov 21 14:32:30 crc kubenswrapper[4897]: I1121 14:32:30.587156 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=5.587133068 podStartE2EDuration="5.587133068s" podCreationTimestamp="2025-11-21 14:32:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:32:30.577040268 +0000 UTC m=+1427.861633743" watchObservedRunningTime="2025-11-21 14:32:30.587133068 +0000 UTC m=+1427.871726543" Nov 21 14:32:30 crc kubenswrapper[4897]: I1121 14:32:30.963878 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-78cfd6c7f4-mzcx9" podUID="2fda0597-e34c-4e8b-a600-c919f01be4b1" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.188:9311/healthcheck\": read tcp 10.217.0.2:37004->10.217.0.188:9311: read: connection reset by peer" Nov 21 14:32:30 crc kubenswrapper[4897]: I1121 14:32:30.963932 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-78cfd6c7f4-mzcx9" podUID="2fda0597-e34c-4e8b-a600-c919f01be4b1" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.188:9311/healthcheck\": read tcp 10.217.0.2:37014->10.217.0.188:9311: read: connection reset by peer" Nov 21 14:32:31 crc kubenswrapper[4897]: I1121 14:32:31.373066 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 21 14:32:31 crc kubenswrapper[4897]: I1121 14:32:31.473396 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c3696f1d-db0b-4910-a65c-31410e2316df-sg-core-conf-yaml\") pod \"c3696f1d-db0b-4910-a65c-31410e2316df\" (UID: \"c3696f1d-db0b-4910-a65c-31410e2316df\") " Nov 21 14:32:31 crc kubenswrapper[4897]: I1121 14:32:31.473836 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c3696f1d-db0b-4910-a65c-31410e2316df-scripts\") pod \"c3696f1d-db0b-4910-a65c-31410e2316df\" (UID: \"c3696f1d-db0b-4910-a65c-31410e2316df\") " Nov 21 14:32:31 crc kubenswrapper[4897]: I1121 14:32:31.474033 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3696f1d-db0b-4910-a65c-31410e2316df-combined-ca-bundle\") pod \"c3696f1d-db0b-4910-a65c-31410e2316df\" (UID: \"c3696f1d-db0b-4910-a65c-31410e2316df\") " Nov 21 14:32:31 crc kubenswrapper[4897]: I1121 14:32:31.474067 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c3696f1d-db0b-4910-a65c-31410e2316df-log-httpd\") pod \"c3696f1d-db0b-4910-a65c-31410e2316df\" (UID: \"c3696f1d-db0b-4910-a65c-31410e2316df\") " Nov 21 14:32:31 crc kubenswrapper[4897]: I1121 14:32:31.474123 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c3696f1d-db0b-4910-a65c-31410e2316df-run-httpd\") pod \"c3696f1d-db0b-4910-a65c-31410e2316df\" (UID: \"c3696f1d-db0b-4910-a65c-31410e2316df\") " Nov 21 14:32:31 crc kubenswrapper[4897]: I1121 14:32:31.474170 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lx66x\" (UniqueName: \"kubernetes.io/projected/c3696f1d-db0b-4910-a65c-31410e2316df-kube-api-access-lx66x\") pod \"c3696f1d-db0b-4910-a65c-31410e2316df\" (UID: \"c3696f1d-db0b-4910-a65c-31410e2316df\") " Nov 21 14:32:31 crc kubenswrapper[4897]: I1121 14:32:31.474191 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3696f1d-db0b-4910-a65c-31410e2316df-config-data\") pod \"c3696f1d-db0b-4910-a65c-31410e2316df\" (UID: \"c3696f1d-db0b-4910-a65c-31410e2316df\") " Nov 21 14:32:31 crc kubenswrapper[4897]: I1121 14:32:31.475757 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c3696f1d-db0b-4910-a65c-31410e2316df-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "c3696f1d-db0b-4910-a65c-31410e2316df" (UID: "c3696f1d-db0b-4910-a65c-31410e2316df"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:32:31 crc kubenswrapper[4897]: I1121 14:32:31.475895 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c3696f1d-db0b-4910-a65c-31410e2316df-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "c3696f1d-db0b-4910-a65c-31410e2316df" (UID: "c3696f1d-db0b-4910-a65c-31410e2316df"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:32:31 crc kubenswrapper[4897]: I1121 14:32:31.481904 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c3696f1d-db0b-4910-a65c-31410e2316df-kube-api-access-lx66x" (OuterVolumeSpecName: "kube-api-access-lx66x") pod "c3696f1d-db0b-4910-a65c-31410e2316df" (UID: "c3696f1d-db0b-4910-a65c-31410e2316df"). InnerVolumeSpecName "kube-api-access-lx66x". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:32:31 crc kubenswrapper[4897]: I1121 14:32:31.481994 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c3696f1d-db0b-4910-a65c-31410e2316df-scripts" (OuterVolumeSpecName: "scripts") pod "c3696f1d-db0b-4910-a65c-31410e2316df" (UID: "c3696f1d-db0b-4910-a65c-31410e2316df"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:32:31 crc kubenswrapper[4897]: I1121 14:32:31.499574 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-78cfd6c7f4-mzcx9" Nov 21 14:32:31 crc kubenswrapper[4897]: I1121 14:32:31.518648 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c3696f1d-db0b-4910-a65c-31410e2316df-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "c3696f1d-db0b-4910-a65c-31410e2316df" (UID: "c3696f1d-db0b-4910-a65c-31410e2316df"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:32:31 crc kubenswrapper[4897]: I1121 14:32:31.560714 4897 generic.go:334] "Generic (PLEG): container finished" podID="2fda0597-e34c-4e8b-a600-c919f01be4b1" containerID="5f1d19bd9fe7bda7a3e07c2b27a9eb0b6bcd3c942d2b28ab2452ea7c90c9b03c" exitCode=0 Nov 21 14:32:31 crc kubenswrapper[4897]: I1121 14:32:31.560751 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-78cfd6c7f4-mzcx9" event={"ID":"2fda0597-e34c-4e8b-a600-c919f01be4b1","Type":"ContainerDied","Data":"5f1d19bd9fe7bda7a3e07c2b27a9eb0b6bcd3c942d2b28ab2452ea7c90c9b03c"} Nov 21 14:32:31 crc kubenswrapper[4897]: I1121 14:32:31.560796 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-78cfd6c7f4-mzcx9" event={"ID":"2fda0597-e34c-4e8b-a600-c919f01be4b1","Type":"ContainerDied","Data":"ab87fdd0c844586389f5eb55925299035e386669a41ac71ec499c5db3bfa9c78"} Nov 21 14:32:31 crc kubenswrapper[4897]: I1121 14:32:31.560815 4897 scope.go:117] "RemoveContainer" containerID="5f1d19bd9fe7bda7a3e07c2b27a9eb0b6bcd3c942d2b28ab2452ea7c90c9b03c" Nov 21 14:32:31 crc kubenswrapper[4897]: I1121 14:32:31.560810 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-78cfd6c7f4-mzcx9" Nov 21 14:32:31 crc kubenswrapper[4897]: I1121 14:32:31.567948 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c3696f1d-db0b-4910-a65c-31410e2316df","Type":"ContainerDied","Data":"f77bbf271263dc44422e9ab3ef764bb5d530727524987b2b69ac5f01e86fd312"} Nov 21 14:32:31 crc kubenswrapper[4897]: I1121 14:32:31.568044 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 21 14:32:31 crc kubenswrapper[4897]: I1121 14:32:31.575821 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fda0597-e34c-4e8b-a600-c919f01be4b1-combined-ca-bundle\") pod \"2fda0597-e34c-4e8b-a600-c919f01be4b1\" (UID: \"2fda0597-e34c-4e8b-a600-c919f01be4b1\") " Nov 21 14:32:31 crc kubenswrapper[4897]: I1121 14:32:31.576069 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2lhqk\" (UniqueName: \"kubernetes.io/projected/2fda0597-e34c-4e8b-a600-c919f01be4b1-kube-api-access-2lhqk\") pod \"2fda0597-e34c-4e8b-a600-c919f01be4b1\" (UID: \"2fda0597-e34c-4e8b-a600-c919f01be4b1\") " Nov 21 14:32:31 crc kubenswrapper[4897]: I1121 14:32:31.576118 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2fda0597-e34c-4e8b-a600-c919f01be4b1-config-data-custom\") pod \"2fda0597-e34c-4e8b-a600-c919f01be4b1\" (UID: \"2fda0597-e34c-4e8b-a600-c919f01be4b1\") " Nov 21 14:32:31 crc kubenswrapper[4897]: I1121 14:32:31.576183 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2fda0597-e34c-4e8b-a600-c919f01be4b1-logs\") pod \"2fda0597-e34c-4e8b-a600-c919f01be4b1\" (UID: \"2fda0597-e34c-4e8b-a600-c919f01be4b1\") " Nov 21 14:32:31 crc kubenswrapper[4897]: I1121 14:32:31.576286 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2fda0597-e34c-4e8b-a600-c919f01be4b1-config-data\") pod \"2fda0597-e34c-4e8b-a600-c919f01be4b1\" (UID: \"2fda0597-e34c-4e8b-a600-c919f01be4b1\") " Nov 21 14:32:31 crc kubenswrapper[4897]: I1121 14:32:31.576534 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2fda0597-e34c-4e8b-a600-c919f01be4b1-logs" (OuterVolumeSpecName: "logs") pod "2fda0597-e34c-4e8b-a600-c919f01be4b1" (UID: "2fda0597-e34c-4e8b-a600-c919f01be4b1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:32:31 crc kubenswrapper[4897]: I1121 14:32:31.576850 4897 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c3696f1d-db0b-4910-a65c-31410e2316df-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 21 14:32:31 crc kubenswrapper[4897]: I1121 14:32:31.576863 4897 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2fda0597-e34c-4e8b-a600-c919f01be4b1-logs\") on node \"crc\" DevicePath \"\"" Nov 21 14:32:31 crc kubenswrapper[4897]: I1121 14:32:31.576873 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lx66x\" (UniqueName: \"kubernetes.io/projected/c3696f1d-db0b-4910-a65c-31410e2316df-kube-api-access-lx66x\") on node \"crc\" DevicePath \"\"" Nov 21 14:32:31 crc kubenswrapper[4897]: I1121 14:32:31.576882 4897 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c3696f1d-db0b-4910-a65c-31410e2316df-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 21 14:32:31 crc kubenswrapper[4897]: I1121 14:32:31.576890 4897 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c3696f1d-db0b-4910-a65c-31410e2316df-scripts\") on node \"crc\" DevicePath \"\"" Nov 21 14:32:31 crc kubenswrapper[4897]: I1121 14:32:31.576900 4897 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c3696f1d-db0b-4910-a65c-31410e2316df-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 21 14:32:31 crc kubenswrapper[4897]: I1121 14:32:31.579360 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2fda0597-e34c-4e8b-a600-c919f01be4b1-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "2fda0597-e34c-4e8b-a600-c919f01be4b1" (UID: "2fda0597-e34c-4e8b-a600-c919f01be4b1"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:32:31 crc kubenswrapper[4897]: I1121 14:32:31.579967 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2fda0597-e34c-4e8b-a600-c919f01be4b1-kube-api-access-2lhqk" (OuterVolumeSpecName: "kube-api-access-2lhqk") pod "2fda0597-e34c-4e8b-a600-c919f01be4b1" (UID: "2fda0597-e34c-4e8b-a600-c919f01be4b1"). InnerVolumeSpecName "kube-api-access-2lhqk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:32:31 crc kubenswrapper[4897]: I1121 14:32:31.592943 4897 scope.go:117] "RemoveContainer" containerID="3205cf41928a4172b4f181fe69f55ee8dd1a3bc5673e03408997ccab345fb4fb" Nov 21 14:32:31 crc kubenswrapper[4897]: I1121 14:32:31.602648 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c3696f1d-db0b-4910-a65c-31410e2316df-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c3696f1d-db0b-4910-a65c-31410e2316df" (UID: "c3696f1d-db0b-4910-a65c-31410e2316df"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:32:31 crc kubenswrapper[4897]: I1121 14:32:31.618205 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2fda0597-e34c-4e8b-a600-c919f01be4b1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2fda0597-e34c-4e8b-a600-c919f01be4b1" (UID: "2fda0597-e34c-4e8b-a600-c919f01be4b1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:32:31 crc kubenswrapper[4897]: I1121 14:32:31.618881 4897 scope.go:117] "RemoveContainer" containerID="5f1d19bd9fe7bda7a3e07c2b27a9eb0b6bcd3c942d2b28ab2452ea7c90c9b03c" Nov 21 14:32:31 crc kubenswrapper[4897]: E1121 14:32:31.619332 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5f1d19bd9fe7bda7a3e07c2b27a9eb0b6bcd3c942d2b28ab2452ea7c90c9b03c\": container with ID starting with 5f1d19bd9fe7bda7a3e07c2b27a9eb0b6bcd3c942d2b28ab2452ea7c90c9b03c not found: ID does not exist" containerID="5f1d19bd9fe7bda7a3e07c2b27a9eb0b6bcd3c942d2b28ab2452ea7c90c9b03c" Nov 21 14:32:31 crc kubenswrapper[4897]: I1121 14:32:31.619381 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f1d19bd9fe7bda7a3e07c2b27a9eb0b6bcd3c942d2b28ab2452ea7c90c9b03c"} err="failed to get container status \"5f1d19bd9fe7bda7a3e07c2b27a9eb0b6bcd3c942d2b28ab2452ea7c90c9b03c\": rpc error: code = NotFound desc = could not find container \"5f1d19bd9fe7bda7a3e07c2b27a9eb0b6bcd3c942d2b28ab2452ea7c90c9b03c\": container with ID starting with 5f1d19bd9fe7bda7a3e07c2b27a9eb0b6bcd3c942d2b28ab2452ea7c90c9b03c not found: ID does not exist" Nov 21 14:32:31 crc kubenswrapper[4897]: I1121 14:32:31.619409 4897 scope.go:117] "RemoveContainer" containerID="3205cf41928a4172b4f181fe69f55ee8dd1a3bc5673e03408997ccab345fb4fb" Nov 21 14:32:31 crc kubenswrapper[4897]: E1121 14:32:31.619859 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3205cf41928a4172b4f181fe69f55ee8dd1a3bc5673e03408997ccab345fb4fb\": container with ID starting with 3205cf41928a4172b4f181fe69f55ee8dd1a3bc5673e03408997ccab345fb4fb not found: ID does not exist" containerID="3205cf41928a4172b4f181fe69f55ee8dd1a3bc5673e03408997ccab345fb4fb" Nov 21 14:32:31 crc kubenswrapper[4897]: I1121 14:32:31.619895 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3205cf41928a4172b4f181fe69f55ee8dd1a3bc5673e03408997ccab345fb4fb"} err="failed to get container status \"3205cf41928a4172b4f181fe69f55ee8dd1a3bc5673e03408997ccab345fb4fb\": rpc error: code = NotFound desc = could not find container \"3205cf41928a4172b4f181fe69f55ee8dd1a3bc5673e03408997ccab345fb4fb\": container with ID starting with 3205cf41928a4172b4f181fe69f55ee8dd1a3bc5673e03408997ccab345fb4fb not found: ID does not exist" Nov 21 14:32:31 crc kubenswrapper[4897]: I1121 14:32:31.619915 4897 scope.go:117] "RemoveContainer" containerID="5a54ebcb33a48f8815ea865396faddc9bb3678b376fb5464e8ddda342d748e77" Nov 21 14:32:31 crc kubenswrapper[4897]: I1121 14:32:31.620490 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c3696f1d-db0b-4910-a65c-31410e2316df-config-data" (OuterVolumeSpecName: "config-data") pod "c3696f1d-db0b-4910-a65c-31410e2316df" (UID: "c3696f1d-db0b-4910-a65c-31410e2316df"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:32:31 crc kubenswrapper[4897]: I1121 14:32:31.624663 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2fda0597-e34c-4e8b-a600-c919f01be4b1-config-data" (OuterVolumeSpecName: "config-data") pod "2fda0597-e34c-4e8b-a600-c919f01be4b1" (UID: "2fda0597-e34c-4e8b-a600-c919f01be4b1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:32:31 crc kubenswrapper[4897]: I1121 14:32:31.648811 4897 scope.go:117] "RemoveContainer" containerID="3c6c1e27ec72c0edf905c8a623e78bf432a92f75bf1c931a7707cf9faca2b35d" Nov 21 14:32:31 crc kubenswrapper[4897]: I1121 14:32:31.671722 4897 scope.go:117] "RemoveContainer" containerID="b676d00cf16995ea43993b65d8f5b05023450400f02e4e944e5a3777a2e7e0a5" Nov 21 14:32:31 crc kubenswrapper[4897]: I1121 14:32:31.679038 4897 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3696f1d-db0b-4910-a65c-31410e2316df-config-data\") on node \"crc\" DevicePath \"\"" Nov 21 14:32:31 crc kubenswrapper[4897]: I1121 14:32:31.679066 4897 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2fda0597-e34c-4e8b-a600-c919f01be4b1-config-data\") on node \"crc\" DevicePath \"\"" Nov 21 14:32:31 crc kubenswrapper[4897]: I1121 14:32:31.679077 4897 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fda0597-e34c-4e8b-a600-c919f01be4b1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 21 14:32:31 crc kubenswrapper[4897]: I1121 14:32:31.679098 4897 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3696f1d-db0b-4910-a65c-31410e2316df-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 21 14:32:31 crc kubenswrapper[4897]: I1121 14:32:31.679108 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2lhqk\" (UniqueName: \"kubernetes.io/projected/2fda0597-e34c-4e8b-a600-c919f01be4b1-kube-api-access-2lhqk\") on node \"crc\" DevicePath \"\"" Nov 21 14:32:31 crc kubenswrapper[4897]: I1121 14:32:31.679116 4897 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2fda0597-e34c-4e8b-a600-c919f01be4b1-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 21 14:32:31 crc kubenswrapper[4897]: I1121 14:32:31.695824 4897 scope.go:117] "RemoveContainer" containerID="4e82c69b36fc692afc0ff6d360048a4a9871804fa3ae5fb182d5567f08ec50d1" Nov 21 14:32:31 crc kubenswrapper[4897]: I1121 14:32:31.905840 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-78cfd6c7f4-mzcx9"] Nov 21 14:32:31 crc kubenswrapper[4897]: I1121 14:32:31.917994 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-78cfd6c7f4-mzcx9"] Nov 21 14:32:31 crc kubenswrapper[4897]: I1121 14:32:31.934541 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 21 14:32:31 crc kubenswrapper[4897]: I1121 14:32:31.943824 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 21 14:32:31 crc kubenswrapper[4897]: I1121 14:32:31.953091 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 21 14:32:31 crc kubenswrapper[4897]: E1121 14:32:31.953659 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3696f1d-db0b-4910-a65c-31410e2316df" containerName="sg-core" Nov 21 14:32:31 crc kubenswrapper[4897]: I1121 14:32:31.953681 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3696f1d-db0b-4910-a65c-31410e2316df" containerName="sg-core" Nov 21 14:32:31 crc kubenswrapper[4897]: E1121 14:32:31.953711 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11f3e22f-051b-42ab-b70e-7813ecbbc8f5" containerName="dnsmasq-dns" Nov 21 14:32:31 crc kubenswrapper[4897]: I1121 14:32:31.953719 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="11f3e22f-051b-42ab-b70e-7813ecbbc8f5" containerName="dnsmasq-dns" Nov 21 14:32:31 crc kubenswrapper[4897]: E1121 14:32:31.953738 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3696f1d-db0b-4910-a65c-31410e2316df" containerName="ceilometer-central-agent" Nov 21 14:32:31 crc kubenswrapper[4897]: I1121 14:32:31.953746 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3696f1d-db0b-4910-a65c-31410e2316df" containerName="ceilometer-central-agent" Nov 21 14:32:31 crc kubenswrapper[4897]: E1121 14:32:31.953774 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11f3e22f-051b-42ab-b70e-7813ecbbc8f5" containerName="init" Nov 21 14:32:31 crc kubenswrapper[4897]: I1121 14:32:31.953782 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="11f3e22f-051b-42ab-b70e-7813ecbbc8f5" containerName="init" Nov 21 14:32:31 crc kubenswrapper[4897]: E1121 14:32:31.953801 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fda0597-e34c-4e8b-a600-c919f01be4b1" containerName="barbican-api" Nov 21 14:32:31 crc kubenswrapper[4897]: I1121 14:32:31.953811 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fda0597-e34c-4e8b-a600-c919f01be4b1" containerName="barbican-api" Nov 21 14:32:31 crc kubenswrapper[4897]: E1121 14:32:31.953829 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3696f1d-db0b-4910-a65c-31410e2316df" containerName="proxy-httpd" Nov 21 14:32:31 crc kubenswrapper[4897]: I1121 14:32:31.953837 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3696f1d-db0b-4910-a65c-31410e2316df" containerName="proxy-httpd" Nov 21 14:32:31 crc kubenswrapper[4897]: E1121 14:32:31.953848 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fda0597-e34c-4e8b-a600-c919f01be4b1" containerName="barbican-api-log" Nov 21 14:32:31 crc kubenswrapper[4897]: I1121 14:32:31.953855 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fda0597-e34c-4e8b-a600-c919f01be4b1" containerName="barbican-api-log" Nov 21 14:32:31 crc kubenswrapper[4897]: E1121 14:32:31.953874 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3696f1d-db0b-4910-a65c-31410e2316df" containerName="ceilometer-notification-agent" Nov 21 14:32:31 crc kubenswrapper[4897]: I1121 14:32:31.953881 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3696f1d-db0b-4910-a65c-31410e2316df" containerName="ceilometer-notification-agent" Nov 21 14:32:31 crc kubenswrapper[4897]: I1121 14:32:31.954127 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="11f3e22f-051b-42ab-b70e-7813ecbbc8f5" containerName="dnsmasq-dns" Nov 21 14:32:31 crc kubenswrapper[4897]: I1121 14:32:31.954149 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="2fda0597-e34c-4e8b-a600-c919f01be4b1" containerName="barbican-api" Nov 21 14:32:31 crc kubenswrapper[4897]: I1121 14:32:31.954168 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3696f1d-db0b-4910-a65c-31410e2316df" containerName="ceilometer-notification-agent" Nov 21 14:32:31 crc kubenswrapper[4897]: I1121 14:32:31.954185 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3696f1d-db0b-4910-a65c-31410e2316df" containerName="ceilometer-central-agent" Nov 21 14:32:31 crc kubenswrapper[4897]: I1121 14:32:31.954201 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3696f1d-db0b-4910-a65c-31410e2316df" containerName="proxy-httpd" Nov 21 14:32:31 crc kubenswrapper[4897]: I1121 14:32:31.954212 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="2fda0597-e34c-4e8b-a600-c919f01be4b1" containerName="barbican-api-log" Nov 21 14:32:31 crc kubenswrapper[4897]: I1121 14:32:31.954235 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3696f1d-db0b-4910-a65c-31410e2316df" containerName="sg-core" Nov 21 14:32:31 crc kubenswrapper[4897]: I1121 14:32:31.956651 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 21 14:32:31 crc kubenswrapper[4897]: I1121 14:32:31.961479 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 21 14:32:31 crc kubenswrapper[4897]: I1121 14:32:31.963715 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 21 14:32:31 crc kubenswrapper[4897]: I1121 14:32:31.964133 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 21 14:32:32 crc kubenswrapper[4897]: I1121 14:32:32.087172 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3dc8b9a3-65bf-436b-b968-20cade83734b-config-data\") pod \"ceilometer-0\" (UID: \"3dc8b9a3-65bf-436b-b968-20cade83734b\") " pod="openstack/ceilometer-0" Nov 21 14:32:32 crc kubenswrapper[4897]: I1121 14:32:32.087232 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3dc8b9a3-65bf-436b-b968-20cade83734b-run-httpd\") pod \"ceilometer-0\" (UID: \"3dc8b9a3-65bf-436b-b968-20cade83734b\") " pod="openstack/ceilometer-0" Nov 21 14:32:32 crc kubenswrapper[4897]: I1121 14:32:32.087248 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3dc8b9a3-65bf-436b-b968-20cade83734b-scripts\") pod \"ceilometer-0\" (UID: \"3dc8b9a3-65bf-436b-b968-20cade83734b\") " pod="openstack/ceilometer-0" Nov 21 14:32:32 crc kubenswrapper[4897]: I1121 14:32:32.087271 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rfxtr\" (UniqueName: \"kubernetes.io/projected/3dc8b9a3-65bf-436b-b968-20cade83734b-kube-api-access-rfxtr\") pod \"ceilometer-0\" (UID: \"3dc8b9a3-65bf-436b-b968-20cade83734b\") " pod="openstack/ceilometer-0" Nov 21 14:32:32 crc kubenswrapper[4897]: I1121 14:32:32.087646 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3dc8b9a3-65bf-436b-b968-20cade83734b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3dc8b9a3-65bf-436b-b968-20cade83734b\") " pod="openstack/ceilometer-0" Nov 21 14:32:32 crc kubenswrapper[4897]: I1121 14:32:32.087723 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3dc8b9a3-65bf-436b-b968-20cade83734b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3dc8b9a3-65bf-436b-b968-20cade83734b\") " pod="openstack/ceilometer-0" Nov 21 14:32:32 crc kubenswrapper[4897]: I1121 14:32:32.087890 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3dc8b9a3-65bf-436b-b968-20cade83734b-log-httpd\") pod \"ceilometer-0\" (UID: \"3dc8b9a3-65bf-436b-b968-20cade83734b\") " pod="openstack/ceilometer-0" Nov 21 14:32:32 crc kubenswrapper[4897]: I1121 14:32:32.102700 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2fda0597-e34c-4e8b-a600-c919f01be4b1" path="/var/lib/kubelet/pods/2fda0597-e34c-4e8b-a600-c919f01be4b1/volumes" Nov 21 14:32:32 crc kubenswrapper[4897]: I1121 14:32:32.103539 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c3696f1d-db0b-4910-a65c-31410e2316df" path="/var/lib/kubelet/pods/c3696f1d-db0b-4910-a65c-31410e2316df/volumes" Nov 21 14:32:32 crc kubenswrapper[4897]: I1121 14:32:32.189824 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3dc8b9a3-65bf-436b-b968-20cade83734b-log-httpd\") pod \"ceilometer-0\" (UID: \"3dc8b9a3-65bf-436b-b968-20cade83734b\") " pod="openstack/ceilometer-0" Nov 21 14:32:32 crc kubenswrapper[4897]: I1121 14:32:32.189889 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3dc8b9a3-65bf-436b-b968-20cade83734b-config-data\") pod \"ceilometer-0\" (UID: \"3dc8b9a3-65bf-436b-b968-20cade83734b\") " pod="openstack/ceilometer-0" Nov 21 14:32:32 crc kubenswrapper[4897]: I1121 14:32:32.189943 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3dc8b9a3-65bf-436b-b968-20cade83734b-run-httpd\") pod \"ceilometer-0\" (UID: \"3dc8b9a3-65bf-436b-b968-20cade83734b\") " pod="openstack/ceilometer-0" Nov 21 14:32:32 crc kubenswrapper[4897]: I1121 14:32:32.189974 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3dc8b9a3-65bf-436b-b968-20cade83734b-scripts\") pod \"ceilometer-0\" (UID: \"3dc8b9a3-65bf-436b-b968-20cade83734b\") " pod="openstack/ceilometer-0" Nov 21 14:32:32 crc kubenswrapper[4897]: I1121 14:32:32.190014 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rfxtr\" (UniqueName: \"kubernetes.io/projected/3dc8b9a3-65bf-436b-b968-20cade83734b-kube-api-access-rfxtr\") pod \"ceilometer-0\" (UID: \"3dc8b9a3-65bf-436b-b968-20cade83734b\") " pod="openstack/ceilometer-0" Nov 21 14:32:32 crc kubenswrapper[4897]: I1121 14:32:32.190359 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3dc8b9a3-65bf-436b-b968-20cade83734b-log-httpd\") pod \"ceilometer-0\" (UID: \"3dc8b9a3-65bf-436b-b968-20cade83734b\") " pod="openstack/ceilometer-0" Nov 21 14:32:32 crc kubenswrapper[4897]: I1121 14:32:32.190558 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3dc8b9a3-65bf-436b-b968-20cade83734b-run-httpd\") pod \"ceilometer-0\" (UID: \"3dc8b9a3-65bf-436b-b968-20cade83734b\") " pod="openstack/ceilometer-0" Nov 21 14:32:32 crc kubenswrapper[4897]: I1121 14:32:32.190801 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3dc8b9a3-65bf-436b-b968-20cade83734b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3dc8b9a3-65bf-436b-b968-20cade83734b\") " pod="openstack/ceilometer-0" Nov 21 14:32:32 crc kubenswrapper[4897]: I1121 14:32:32.191311 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3dc8b9a3-65bf-436b-b968-20cade83734b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3dc8b9a3-65bf-436b-b968-20cade83734b\") " pod="openstack/ceilometer-0" Nov 21 14:32:32 crc kubenswrapper[4897]: I1121 14:32:32.194832 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3dc8b9a3-65bf-436b-b968-20cade83734b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3dc8b9a3-65bf-436b-b968-20cade83734b\") " pod="openstack/ceilometer-0" Nov 21 14:32:32 crc kubenswrapper[4897]: I1121 14:32:32.195414 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3dc8b9a3-65bf-436b-b968-20cade83734b-scripts\") pod \"ceilometer-0\" (UID: \"3dc8b9a3-65bf-436b-b968-20cade83734b\") " pod="openstack/ceilometer-0" Nov 21 14:32:32 crc kubenswrapper[4897]: I1121 14:32:32.195636 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3dc8b9a3-65bf-436b-b968-20cade83734b-config-data\") pod \"ceilometer-0\" (UID: \"3dc8b9a3-65bf-436b-b968-20cade83734b\") " pod="openstack/ceilometer-0" Nov 21 14:32:32 crc kubenswrapper[4897]: I1121 14:32:32.196468 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3dc8b9a3-65bf-436b-b968-20cade83734b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3dc8b9a3-65bf-436b-b968-20cade83734b\") " pod="openstack/ceilometer-0" Nov 21 14:32:32 crc kubenswrapper[4897]: I1121 14:32:32.207429 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rfxtr\" (UniqueName: \"kubernetes.io/projected/3dc8b9a3-65bf-436b-b968-20cade83734b-kube-api-access-rfxtr\") pod \"ceilometer-0\" (UID: \"3dc8b9a3-65bf-436b-b968-20cade83734b\") " pod="openstack/ceilometer-0" Nov 21 14:32:32 crc kubenswrapper[4897]: I1121 14:32:32.275148 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 21 14:32:32 crc kubenswrapper[4897]: I1121 14:32:32.580833 4897 generic.go:334] "Generic (PLEG): container finished" podID="55644d8b-b8a3-463e-8045-b7391e9c2419" containerID="e7ac0248e2b28f323c61ff8d85c2fd0aaf30199090d2984738889239f670d176" exitCode=0 Nov 21 14:32:32 crc kubenswrapper[4897]: I1121 14:32:32.580934 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-7dvt9" event={"ID":"55644d8b-b8a3-463e-8045-b7391e9c2419","Type":"ContainerDied","Data":"e7ac0248e2b28f323c61ff8d85c2fd0aaf30199090d2984738889239f670d176"} Nov 21 14:32:32 crc kubenswrapper[4897]: I1121 14:32:32.751868 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 21 14:32:32 crc kubenswrapper[4897]: W1121 14:32:32.757402 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3dc8b9a3_65bf_436b_b968_20cade83734b.slice/crio-4ec1b14863278e91b666cd671229063cc15246cd0bf6c9f0030795d2c16b256e WatchSource:0}: Error finding container 4ec1b14863278e91b666cd671229063cc15246cd0bf6c9f0030795d2c16b256e: Status 404 returned error can't find the container with id 4ec1b14863278e91b666cd671229063cc15246cd0bf6c9f0030795d2c16b256e Nov 21 14:32:33 crc kubenswrapper[4897]: I1121 14:32:33.597719 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3dc8b9a3-65bf-436b-b968-20cade83734b","Type":"ContainerStarted","Data":"4ec1b14863278e91b666cd671229063cc15246cd0bf6c9f0030795d2c16b256e"} Nov 21 14:32:33 crc kubenswrapper[4897]: I1121 14:32:33.963017 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-7dvt9" Nov 21 14:32:34 crc kubenswrapper[4897]: I1121 14:32:34.036958 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55644d8b-b8a3-463e-8045-b7391e9c2419-combined-ca-bundle\") pod \"55644d8b-b8a3-463e-8045-b7391e9c2419\" (UID: \"55644d8b-b8a3-463e-8045-b7391e9c2419\") " Nov 21 14:32:34 crc kubenswrapper[4897]: I1121 14:32:34.037110 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/55644d8b-b8a3-463e-8045-b7391e9c2419-etc-machine-id\") pod \"55644d8b-b8a3-463e-8045-b7391e9c2419\" (UID: \"55644d8b-b8a3-463e-8045-b7391e9c2419\") " Nov 21 14:32:34 crc kubenswrapper[4897]: I1121 14:32:34.037175 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/55644d8b-b8a3-463e-8045-b7391e9c2419-scripts\") pod \"55644d8b-b8a3-463e-8045-b7391e9c2419\" (UID: \"55644d8b-b8a3-463e-8045-b7391e9c2419\") " Nov 21 14:32:34 crc kubenswrapper[4897]: I1121 14:32:34.037208 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55644d8b-b8a3-463e-8045-b7391e9c2419-config-data\") pod \"55644d8b-b8a3-463e-8045-b7391e9c2419\" (UID: \"55644d8b-b8a3-463e-8045-b7391e9c2419\") " Nov 21 14:32:34 crc kubenswrapper[4897]: I1121 14:32:34.037296 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/55644d8b-b8a3-463e-8045-b7391e9c2419-db-sync-config-data\") pod \"55644d8b-b8a3-463e-8045-b7391e9c2419\" (UID: \"55644d8b-b8a3-463e-8045-b7391e9c2419\") " Nov 21 14:32:34 crc kubenswrapper[4897]: I1121 14:32:34.037292 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/55644d8b-b8a3-463e-8045-b7391e9c2419-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "55644d8b-b8a3-463e-8045-b7391e9c2419" (UID: "55644d8b-b8a3-463e-8045-b7391e9c2419"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 21 14:32:34 crc kubenswrapper[4897]: I1121 14:32:34.038739 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dzppc\" (UniqueName: \"kubernetes.io/projected/55644d8b-b8a3-463e-8045-b7391e9c2419-kube-api-access-dzppc\") pod \"55644d8b-b8a3-463e-8045-b7391e9c2419\" (UID: \"55644d8b-b8a3-463e-8045-b7391e9c2419\") " Nov 21 14:32:34 crc kubenswrapper[4897]: I1121 14:32:34.039791 4897 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/55644d8b-b8a3-463e-8045-b7391e9c2419-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 21 14:32:34 crc kubenswrapper[4897]: I1121 14:32:34.042624 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55644d8b-b8a3-463e-8045-b7391e9c2419-scripts" (OuterVolumeSpecName: "scripts") pod "55644d8b-b8a3-463e-8045-b7391e9c2419" (UID: "55644d8b-b8a3-463e-8045-b7391e9c2419"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:32:34 crc kubenswrapper[4897]: I1121 14:32:34.044219 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55644d8b-b8a3-463e-8045-b7391e9c2419-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "55644d8b-b8a3-463e-8045-b7391e9c2419" (UID: "55644d8b-b8a3-463e-8045-b7391e9c2419"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:32:34 crc kubenswrapper[4897]: I1121 14:32:34.047574 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/55644d8b-b8a3-463e-8045-b7391e9c2419-kube-api-access-dzppc" (OuterVolumeSpecName: "kube-api-access-dzppc") pod "55644d8b-b8a3-463e-8045-b7391e9c2419" (UID: "55644d8b-b8a3-463e-8045-b7391e9c2419"). InnerVolumeSpecName "kube-api-access-dzppc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:32:34 crc kubenswrapper[4897]: I1121 14:32:34.075393 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55644d8b-b8a3-463e-8045-b7391e9c2419-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "55644d8b-b8a3-463e-8045-b7391e9c2419" (UID: "55644d8b-b8a3-463e-8045-b7391e9c2419"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:32:34 crc kubenswrapper[4897]: I1121 14:32:34.098751 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55644d8b-b8a3-463e-8045-b7391e9c2419-config-data" (OuterVolumeSpecName: "config-data") pod "55644d8b-b8a3-463e-8045-b7391e9c2419" (UID: "55644d8b-b8a3-463e-8045-b7391e9c2419"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:32:34 crc kubenswrapper[4897]: I1121 14:32:34.143031 4897 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55644d8b-b8a3-463e-8045-b7391e9c2419-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 21 14:32:34 crc kubenswrapper[4897]: I1121 14:32:34.143091 4897 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/55644d8b-b8a3-463e-8045-b7391e9c2419-scripts\") on node \"crc\" DevicePath \"\"" Nov 21 14:32:34 crc kubenswrapper[4897]: I1121 14:32:34.143105 4897 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55644d8b-b8a3-463e-8045-b7391e9c2419-config-data\") on node \"crc\" DevicePath \"\"" Nov 21 14:32:34 crc kubenswrapper[4897]: I1121 14:32:34.143116 4897 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/55644d8b-b8a3-463e-8045-b7391e9c2419-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 21 14:32:34 crc kubenswrapper[4897]: I1121 14:32:34.143127 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dzppc\" (UniqueName: \"kubernetes.io/projected/55644d8b-b8a3-463e-8045-b7391e9c2419-kube-api-access-dzppc\") on node \"crc\" DevicePath \"\"" Nov 21 14:32:34 crc kubenswrapper[4897]: I1121 14:32:34.608379 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-7dvt9" event={"ID":"55644d8b-b8a3-463e-8045-b7391e9c2419","Type":"ContainerDied","Data":"33c7a02a10a9950715b35226e715a310a1019d82d1c2fdd396e29be1f520e9d4"} Nov 21 14:32:34 crc kubenswrapper[4897]: I1121 14:32:34.608609 4897 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="33c7a02a10a9950715b35226e715a310a1019d82d1c2fdd396e29be1f520e9d4" Nov 21 14:32:34 crc kubenswrapper[4897]: I1121 14:32:34.608406 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-7dvt9" Nov 21 14:32:34 crc kubenswrapper[4897]: I1121 14:32:34.611446 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3dc8b9a3-65bf-436b-b968-20cade83734b","Type":"ContainerStarted","Data":"321b7f3d268272bf59b02dea3780437f7e52fcdd88b5efebf4835bb68e5e9423"} Nov 21 14:32:34 crc kubenswrapper[4897]: I1121 14:32:34.611492 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3dc8b9a3-65bf-436b-b968-20cade83734b","Type":"ContainerStarted","Data":"2d00f6a1feaa6c3dd368484437a106b7d628e3481e0c75aed4819ed8a89197b6"} Nov 21 14:32:34 crc kubenswrapper[4897]: I1121 14:32:34.941212 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Nov 21 14:32:34 crc kubenswrapper[4897]: E1121 14:32:34.941868 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55644d8b-b8a3-463e-8045-b7391e9c2419" containerName="cinder-db-sync" Nov 21 14:32:34 crc kubenswrapper[4897]: I1121 14:32:34.941886 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="55644d8b-b8a3-463e-8045-b7391e9c2419" containerName="cinder-db-sync" Nov 21 14:32:34 crc kubenswrapper[4897]: I1121 14:32:34.942151 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="55644d8b-b8a3-463e-8045-b7391e9c2419" containerName="cinder-db-sync" Nov 21 14:32:34 crc kubenswrapper[4897]: I1121 14:32:34.943851 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 21 14:32:34 crc kubenswrapper[4897]: I1121 14:32:34.946472 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-pr8nf" Nov 21 14:32:34 crc kubenswrapper[4897]: I1121 14:32:34.946727 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Nov 21 14:32:34 crc kubenswrapper[4897]: I1121 14:32:34.946859 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Nov 21 14:32:34 crc kubenswrapper[4897]: I1121 14:32:34.946865 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Nov 21 14:32:34 crc kubenswrapper[4897]: I1121 14:32:34.984764 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 21 14:32:35 crc kubenswrapper[4897]: I1121 14:32:35.064804 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-674b76c99f-zpcvf"] Nov 21 14:32:35 crc kubenswrapper[4897]: I1121 14:32:35.073109 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-674b76c99f-zpcvf" Nov 21 14:32:35 crc kubenswrapper[4897]: I1121 14:32:35.075142 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jksv8\" (UniqueName: \"kubernetes.io/projected/79802b6b-53b5-4568-b79b-83da7dbbf123-kube-api-access-jksv8\") pod \"cinder-scheduler-0\" (UID: \"79802b6b-53b5-4568-b79b-83da7dbbf123\") " pod="openstack/cinder-scheduler-0" Nov 21 14:32:35 crc kubenswrapper[4897]: I1121 14:32:35.075226 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79802b6b-53b5-4568-b79b-83da7dbbf123-config-data\") pod \"cinder-scheduler-0\" (UID: \"79802b6b-53b5-4568-b79b-83da7dbbf123\") " pod="openstack/cinder-scheduler-0" Nov 21 14:32:35 crc kubenswrapper[4897]: I1121 14:32:35.075342 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79802b6b-53b5-4568-b79b-83da7dbbf123-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"79802b6b-53b5-4568-b79b-83da7dbbf123\") " pod="openstack/cinder-scheduler-0" Nov 21 14:32:35 crc kubenswrapper[4897]: I1121 14:32:35.075438 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/79802b6b-53b5-4568-b79b-83da7dbbf123-scripts\") pod \"cinder-scheduler-0\" (UID: \"79802b6b-53b5-4568-b79b-83da7dbbf123\") " pod="openstack/cinder-scheduler-0" Nov 21 14:32:35 crc kubenswrapper[4897]: I1121 14:32:35.075585 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/79802b6b-53b5-4568-b79b-83da7dbbf123-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"79802b6b-53b5-4568-b79b-83da7dbbf123\") " pod="openstack/cinder-scheduler-0" Nov 21 14:32:35 crc kubenswrapper[4897]: I1121 14:32:35.075639 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/79802b6b-53b5-4568-b79b-83da7dbbf123-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"79802b6b-53b5-4568-b79b-83da7dbbf123\") " pod="openstack/cinder-scheduler-0" Nov 21 14:32:35 crc kubenswrapper[4897]: I1121 14:32:35.120575 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-674b76c99f-zpcvf"] Nov 21 14:32:35 crc kubenswrapper[4897]: I1121 14:32:35.137157 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Nov 21 14:32:35 crc kubenswrapper[4897]: I1121 14:32:35.139907 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 21 14:32:35 crc kubenswrapper[4897]: I1121 14:32:35.144614 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Nov 21 14:32:35 crc kubenswrapper[4897]: I1121 14:32:35.150318 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 21 14:32:35 crc kubenswrapper[4897]: I1121 14:32:35.177628 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/881842b1-c889-41be-b92a-4690f665e361-ovsdbserver-nb\") pod \"dnsmasq-dns-674b76c99f-zpcvf\" (UID: \"881842b1-c889-41be-b92a-4690f665e361\") " pod="openstack/dnsmasq-dns-674b76c99f-zpcvf" Nov 21 14:32:35 crc kubenswrapper[4897]: I1121 14:32:35.177679 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/881842b1-c889-41be-b92a-4690f665e361-config\") pod \"dnsmasq-dns-674b76c99f-zpcvf\" (UID: \"881842b1-c889-41be-b92a-4690f665e361\") " pod="openstack/dnsmasq-dns-674b76c99f-zpcvf" Nov 21 14:32:35 crc kubenswrapper[4897]: I1121 14:32:35.177730 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/881842b1-c889-41be-b92a-4690f665e361-ovsdbserver-sb\") pod \"dnsmasq-dns-674b76c99f-zpcvf\" (UID: \"881842b1-c889-41be-b92a-4690f665e361\") " pod="openstack/dnsmasq-dns-674b76c99f-zpcvf" Nov 21 14:32:35 crc kubenswrapper[4897]: I1121 14:32:35.177762 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/79802b6b-53b5-4568-b79b-83da7dbbf123-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"79802b6b-53b5-4568-b79b-83da7dbbf123\") " pod="openstack/cinder-scheduler-0" Nov 21 14:32:35 crc kubenswrapper[4897]: I1121 14:32:35.177799 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/881842b1-c889-41be-b92a-4690f665e361-dns-svc\") pod \"dnsmasq-dns-674b76c99f-zpcvf\" (UID: \"881842b1-c889-41be-b92a-4690f665e361\") " pod="openstack/dnsmasq-dns-674b76c99f-zpcvf" Nov 21 14:32:35 crc kubenswrapper[4897]: I1121 14:32:35.177826 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/79802b6b-53b5-4568-b79b-83da7dbbf123-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"79802b6b-53b5-4568-b79b-83da7dbbf123\") " pod="openstack/cinder-scheduler-0" Nov 21 14:32:35 crc kubenswrapper[4897]: I1121 14:32:35.177850 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jksv8\" (UniqueName: \"kubernetes.io/projected/79802b6b-53b5-4568-b79b-83da7dbbf123-kube-api-access-jksv8\") pod \"cinder-scheduler-0\" (UID: \"79802b6b-53b5-4568-b79b-83da7dbbf123\") " pod="openstack/cinder-scheduler-0" Nov 21 14:32:35 crc kubenswrapper[4897]: I1121 14:32:35.177872 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79802b6b-53b5-4568-b79b-83da7dbbf123-config-data\") pod \"cinder-scheduler-0\" (UID: \"79802b6b-53b5-4568-b79b-83da7dbbf123\") " pod="openstack/cinder-scheduler-0" Nov 21 14:32:35 crc kubenswrapper[4897]: I1121 14:32:35.177937 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79802b6b-53b5-4568-b79b-83da7dbbf123-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"79802b6b-53b5-4568-b79b-83da7dbbf123\") " pod="openstack/cinder-scheduler-0" Nov 21 14:32:35 crc kubenswrapper[4897]: I1121 14:32:35.177963 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/881842b1-c889-41be-b92a-4690f665e361-dns-swift-storage-0\") pod \"dnsmasq-dns-674b76c99f-zpcvf\" (UID: \"881842b1-c889-41be-b92a-4690f665e361\") " pod="openstack/dnsmasq-dns-674b76c99f-zpcvf" Nov 21 14:32:35 crc kubenswrapper[4897]: I1121 14:32:35.177996 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vn5cf\" (UniqueName: \"kubernetes.io/projected/881842b1-c889-41be-b92a-4690f665e361-kube-api-access-vn5cf\") pod \"dnsmasq-dns-674b76c99f-zpcvf\" (UID: \"881842b1-c889-41be-b92a-4690f665e361\") " pod="openstack/dnsmasq-dns-674b76c99f-zpcvf" Nov 21 14:32:35 crc kubenswrapper[4897]: I1121 14:32:35.178056 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/79802b6b-53b5-4568-b79b-83da7dbbf123-scripts\") pod \"cinder-scheduler-0\" (UID: \"79802b6b-53b5-4568-b79b-83da7dbbf123\") " pod="openstack/cinder-scheduler-0" Nov 21 14:32:35 crc kubenswrapper[4897]: I1121 14:32:35.180593 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/79802b6b-53b5-4568-b79b-83da7dbbf123-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"79802b6b-53b5-4568-b79b-83da7dbbf123\") " pod="openstack/cinder-scheduler-0" Nov 21 14:32:35 crc kubenswrapper[4897]: I1121 14:32:35.184103 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79802b6b-53b5-4568-b79b-83da7dbbf123-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"79802b6b-53b5-4568-b79b-83da7dbbf123\") " pod="openstack/cinder-scheduler-0" Nov 21 14:32:35 crc kubenswrapper[4897]: I1121 14:32:35.186558 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79802b6b-53b5-4568-b79b-83da7dbbf123-config-data\") pod \"cinder-scheduler-0\" (UID: \"79802b6b-53b5-4568-b79b-83da7dbbf123\") " pod="openstack/cinder-scheduler-0" Nov 21 14:32:35 crc kubenswrapper[4897]: I1121 14:32:35.188292 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/79802b6b-53b5-4568-b79b-83da7dbbf123-scripts\") pod \"cinder-scheduler-0\" (UID: \"79802b6b-53b5-4568-b79b-83da7dbbf123\") " pod="openstack/cinder-scheduler-0" Nov 21 14:32:35 crc kubenswrapper[4897]: I1121 14:32:35.189669 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/79802b6b-53b5-4568-b79b-83da7dbbf123-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"79802b6b-53b5-4568-b79b-83da7dbbf123\") " pod="openstack/cinder-scheduler-0" Nov 21 14:32:35 crc kubenswrapper[4897]: I1121 14:32:35.199814 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jksv8\" (UniqueName: \"kubernetes.io/projected/79802b6b-53b5-4568-b79b-83da7dbbf123-kube-api-access-jksv8\") pod \"cinder-scheduler-0\" (UID: \"79802b6b-53b5-4568-b79b-83da7dbbf123\") " pod="openstack/cinder-scheduler-0" Nov 21 14:32:35 crc kubenswrapper[4897]: I1121 14:32:35.275981 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 21 14:32:35 crc kubenswrapper[4897]: I1121 14:32:35.279502 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cdf9e87f-440b-41db-a8d1-f3f5f8953310-logs\") pod \"cinder-api-0\" (UID: \"cdf9e87f-440b-41db-a8d1-f3f5f8953310\") " pod="openstack/cinder-api-0" Nov 21 14:32:35 crc kubenswrapper[4897]: I1121 14:32:35.279591 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cdf9e87f-440b-41db-a8d1-f3f5f8953310-config-data-custom\") pod \"cinder-api-0\" (UID: \"cdf9e87f-440b-41db-a8d1-f3f5f8953310\") " pod="openstack/cinder-api-0" Nov 21 14:32:35 crc kubenswrapper[4897]: I1121 14:32:35.279640 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/881842b1-c889-41be-b92a-4690f665e361-ovsdbserver-nb\") pod \"dnsmasq-dns-674b76c99f-zpcvf\" (UID: \"881842b1-c889-41be-b92a-4690f665e361\") " pod="openstack/dnsmasq-dns-674b76c99f-zpcvf" Nov 21 14:32:35 crc kubenswrapper[4897]: I1121 14:32:35.279678 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/881842b1-c889-41be-b92a-4690f665e361-config\") pod \"dnsmasq-dns-674b76c99f-zpcvf\" (UID: \"881842b1-c889-41be-b92a-4690f665e361\") " pod="openstack/dnsmasq-dns-674b76c99f-zpcvf" Nov 21 14:32:35 crc kubenswrapper[4897]: I1121 14:32:35.279711 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cdf9e87f-440b-41db-a8d1-f3f5f8953310-etc-machine-id\") pod \"cinder-api-0\" (UID: \"cdf9e87f-440b-41db-a8d1-f3f5f8953310\") " pod="openstack/cinder-api-0" Nov 21 14:32:35 crc kubenswrapper[4897]: I1121 14:32:35.279783 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/881842b1-c889-41be-b92a-4690f665e361-ovsdbserver-sb\") pod \"dnsmasq-dns-674b76c99f-zpcvf\" (UID: \"881842b1-c889-41be-b92a-4690f665e361\") " pod="openstack/dnsmasq-dns-674b76c99f-zpcvf" Nov 21 14:32:35 crc kubenswrapper[4897]: I1121 14:32:35.279847 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/881842b1-c889-41be-b92a-4690f665e361-dns-svc\") pod \"dnsmasq-dns-674b76c99f-zpcvf\" (UID: \"881842b1-c889-41be-b92a-4690f665e361\") " pod="openstack/dnsmasq-dns-674b76c99f-zpcvf" Nov 21 14:32:35 crc kubenswrapper[4897]: I1121 14:32:35.279874 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c4dm5\" (UniqueName: \"kubernetes.io/projected/cdf9e87f-440b-41db-a8d1-f3f5f8953310-kube-api-access-c4dm5\") pod \"cinder-api-0\" (UID: \"cdf9e87f-440b-41db-a8d1-f3f5f8953310\") " pod="openstack/cinder-api-0" Nov 21 14:32:35 crc kubenswrapper[4897]: I1121 14:32:35.279916 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cdf9e87f-440b-41db-a8d1-f3f5f8953310-scripts\") pod \"cinder-api-0\" (UID: \"cdf9e87f-440b-41db-a8d1-f3f5f8953310\") " pod="openstack/cinder-api-0" Nov 21 14:32:35 crc kubenswrapper[4897]: I1121 14:32:35.279946 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cdf9e87f-440b-41db-a8d1-f3f5f8953310-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"cdf9e87f-440b-41db-a8d1-f3f5f8953310\") " pod="openstack/cinder-api-0" Nov 21 14:32:35 crc kubenswrapper[4897]: I1121 14:32:35.280035 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cdf9e87f-440b-41db-a8d1-f3f5f8953310-config-data\") pod \"cinder-api-0\" (UID: \"cdf9e87f-440b-41db-a8d1-f3f5f8953310\") " pod="openstack/cinder-api-0" Nov 21 14:32:35 crc kubenswrapper[4897]: I1121 14:32:35.280091 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/881842b1-c889-41be-b92a-4690f665e361-dns-swift-storage-0\") pod \"dnsmasq-dns-674b76c99f-zpcvf\" (UID: \"881842b1-c889-41be-b92a-4690f665e361\") " pod="openstack/dnsmasq-dns-674b76c99f-zpcvf" Nov 21 14:32:35 crc kubenswrapper[4897]: I1121 14:32:35.280129 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vn5cf\" (UniqueName: \"kubernetes.io/projected/881842b1-c889-41be-b92a-4690f665e361-kube-api-access-vn5cf\") pod \"dnsmasq-dns-674b76c99f-zpcvf\" (UID: \"881842b1-c889-41be-b92a-4690f665e361\") " pod="openstack/dnsmasq-dns-674b76c99f-zpcvf" Nov 21 14:32:35 crc kubenswrapper[4897]: I1121 14:32:35.280896 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/881842b1-c889-41be-b92a-4690f665e361-ovsdbserver-nb\") pod \"dnsmasq-dns-674b76c99f-zpcvf\" (UID: \"881842b1-c889-41be-b92a-4690f665e361\") " pod="openstack/dnsmasq-dns-674b76c99f-zpcvf" Nov 21 14:32:35 crc kubenswrapper[4897]: I1121 14:32:35.281635 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/881842b1-c889-41be-b92a-4690f665e361-dns-svc\") pod \"dnsmasq-dns-674b76c99f-zpcvf\" (UID: \"881842b1-c889-41be-b92a-4690f665e361\") " pod="openstack/dnsmasq-dns-674b76c99f-zpcvf" Nov 21 14:32:35 crc kubenswrapper[4897]: I1121 14:32:35.281781 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/881842b1-c889-41be-b92a-4690f665e361-dns-swift-storage-0\") pod \"dnsmasq-dns-674b76c99f-zpcvf\" (UID: \"881842b1-c889-41be-b92a-4690f665e361\") " pod="openstack/dnsmasq-dns-674b76c99f-zpcvf" Nov 21 14:32:35 crc kubenswrapper[4897]: I1121 14:32:35.282244 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/881842b1-c889-41be-b92a-4690f665e361-config\") pod \"dnsmasq-dns-674b76c99f-zpcvf\" (UID: \"881842b1-c889-41be-b92a-4690f665e361\") " pod="openstack/dnsmasq-dns-674b76c99f-zpcvf" Nov 21 14:32:35 crc kubenswrapper[4897]: I1121 14:32:35.282576 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/881842b1-c889-41be-b92a-4690f665e361-ovsdbserver-sb\") pod \"dnsmasq-dns-674b76c99f-zpcvf\" (UID: \"881842b1-c889-41be-b92a-4690f665e361\") " pod="openstack/dnsmasq-dns-674b76c99f-zpcvf" Nov 21 14:32:35 crc kubenswrapper[4897]: I1121 14:32:35.315455 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vn5cf\" (UniqueName: \"kubernetes.io/projected/881842b1-c889-41be-b92a-4690f665e361-kube-api-access-vn5cf\") pod \"dnsmasq-dns-674b76c99f-zpcvf\" (UID: \"881842b1-c889-41be-b92a-4690f665e361\") " pod="openstack/dnsmasq-dns-674b76c99f-zpcvf" Nov 21 14:32:35 crc kubenswrapper[4897]: I1121 14:32:35.382363 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c4dm5\" (UniqueName: \"kubernetes.io/projected/cdf9e87f-440b-41db-a8d1-f3f5f8953310-kube-api-access-c4dm5\") pod \"cinder-api-0\" (UID: \"cdf9e87f-440b-41db-a8d1-f3f5f8953310\") " pod="openstack/cinder-api-0" Nov 21 14:32:35 crc kubenswrapper[4897]: I1121 14:32:35.382445 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cdf9e87f-440b-41db-a8d1-f3f5f8953310-scripts\") pod \"cinder-api-0\" (UID: \"cdf9e87f-440b-41db-a8d1-f3f5f8953310\") " pod="openstack/cinder-api-0" Nov 21 14:32:35 crc kubenswrapper[4897]: I1121 14:32:35.382484 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cdf9e87f-440b-41db-a8d1-f3f5f8953310-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"cdf9e87f-440b-41db-a8d1-f3f5f8953310\") " pod="openstack/cinder-api-0" Nov 21 14:32:35 crc kubenswrapper[4897]: I1121 14:32:35.382676 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cdf9e87f-440b-41db-a8d1-f3f5f8953310-config-data\") pod \"cinder-api-0\" (UID: \"cdf9e87f-440b-41db-a8d1-f3f5f8953310\") " pod="openstack/cinder-api-0" Nov 21 14:32:35 crc kubenswrapper[4897]: I1121 14:32:35.382786 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cdf9e87f-440b-41db-a8d1-f3f5f8953310-logs\") pod \"cinder-api-0\" (UID: \"cdf9e87f-440b-41db-a8d1-f3f5f8953310\") " pod="openstack/cinder-api-0" Nov 21 14:32:35 crc kubenswrapper[4897]: I1121 14:32:35.382855 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cdf9e87f-440b-41db-a8d1-f3f5f8953310-config-data-custom\") pod \"cinder-api-0\" (UID: \"cdf9e87f-440b-41db-a8d1-f3f5f8953310\") " pod="openstack/cinder-api-0" Nov 21 14:32:35 crc kubenswrapper[4897]: I1121 14:32:35.382933 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cdf9e87f-440b-41db-a8d1-f3f5f8953310-etc-machine-id\") pod \"cinder-api-0\" (UID: \"cdf9e87f-440b-41db-a8d1-f3f5f8953310\") " pod="openstack/cinder-api-0" Nov 21 14:32:35 crc kubenswrapper[4897]: I1121 14:32:35.383057 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cdf9e87f-440b-41db-a8d1-f3f5f8953310-etc-machine-id\") pod \"cinder-api-0\" (UID: \"cdf9e87f-440b-41db-a8d1-f3f5f8953310\") " pod="openstack/cinder-api-0" Nov 21 14:32:35 crc kubenswrapper[4897]: I1121 14:32:35.383275 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cdf9e87f-440b-41db-a8d1-f3f5f8953310-logs\") pod \"cinder-api-0\" (UID: \"cdf9e87f-440b-41db-a8d1-f3f5f8953310\") " pod="openstack/cinder-api-0" Nov 21 14:32:35 crc kubenswrapper[4897]: I1121 14:32:35.386781 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cdf9e87f-440b-41db-a8d1-f3f5f8953310-config-data-custom\") pod \"cinder-api-0\" (UID: \"cdf9e87f-440b-41db-a8d1-f3f5f8953310\") " pod="openstack/cinder-api-0" Nov 21 14:32:35 crc kubenswrapper[4897]: I1121 14:32:35.393287 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cdf9e87f-440b-41db-a8d1-f3f5f8953310-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"cdf9e87f-440b-41db-a8d1-f3f5f8953310\") " pod="openstack/cinder-api-0" Nov 21 14:32:35 crc kubenswrapper[4897]: I1121 14:32:35.394123 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cdf9e87f-440b-41db-a8d1-f3f5f8953310-scripts\") pod \"cinder-api-0\" (UID: \"cdf9e87f-440b-41db-a8d1-f3f5f8953310\") " pod="openstack/cinder-api-0" Nov 21 14:32:35 crc kubenswrapper[4897]: I1121 14:32:35.401697 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cdf9e87f-440b-41db-a8d1-f3f5f8953310-config-data\") pod \"cinder-api-0\" (UID: \"cdf9e87f-440b-41db-a8d1-f3f5f8953310\") " pod="openstack/cinder-api-0" Nov 21 14:32:35 crc kubenswrapper[4897]: I1121 14:32:35.401710 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-674b76c99f-zpcvf" Nov 21 14:32:35 crc kubenswrapper[4897]: I1121 14:32:35.402530 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c4dm5\" (UniqueName: \"kubernetes.io/projected/cdf9e87f-440b-41db-a8d1-f3f5f8953310-kube-api-access-c4dm5\") pod \"cinder-api-0\" (UID: \"cdf9e87f-440b-41db-a8d1-f3f5f8953310\") " pod="openstack/cinder-api-0" Nov 21 14:32:35 crc kubenswrapper[4897]: I1121 14:32:35.465242 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 21 14:32:35 crc kubenswrapper[4897]: I1121 14:32:35.632340 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3dc8b9a3-65bf-436b-b968-20cade83734b","Type":"ContainerStarted","Data":"665c0ba8bda47e588fbf7269519777bef5776b6faba0458531bf4156232646fb"} Nov 21 14:32:35 crc kubenswrapper[4897]: I1121 14:32:35.798718 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 21 14:32:35 crc kubenswrapper[4897]: I1121 14:32:35.798764 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 21 14:32:35 crc kubenswrapper[4897]: I1121 14:32:35.843212 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 21 14:32:35 crc kubenswrapper[4897]: I1121 14:32:35.858413 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 21 14:32:35 crc kubenswrapper[4897]: I1121 14:32:35.942591 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 21 14:32:36 crc kubenswrapper[4897]: I1121 14:32:36.049226 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-674b76c99f-zpcvf"] Nov 21 14:32:36 crc kubenswrapper[4897]: I1121 14:32:36.110686 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 21 14:32:36 crc kubenswrapper[4897]: I1121 14:32:36.111472 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 21 14:32:36 crc kubenswrapper[4897]: I1121 14:32:36.154478 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 21 14:32:36 crc kubenswrapper[4897]: W1121 14:32:36.158789 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcdf9e87f_440b_41db_a8d1_f3f5f8953310.slice/crio-e429315940c298bee2e17ab5392dc9e4768e1620eb12d003f3e776012cb0dcff WatchSource:0}: Error finding container e429315940c298bee2e17ab5392dc9e4768e1620eb12d003f3e776012cb0dcff: Status 404 returned error can't find the container with id e429315940c298bee2e17ab5392dc9e4768e1620eb12d003f3e776012cb0dcff Nov 21 14:32:36 crc kubenswrapper[4897]: I1121 14:32:36.159377 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 21 14:32:36 crc kubenswrapper[4897]: I1121 14:32:36.182492 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 21 14:32:36 crc kubenswrapper[4897]: I1121 14:32:36.708956 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"79802b6b-53b5-4568-b79b-83da7dbbf123","Type":"ContainerStarted","Data":"d324833442dd6e8b1ab903afa0e6d80c1cc2e4a0ea8015cdeee79bdf10a78e81"} Nov 21 14:32:36 crc kubenswrapper[4897]: I1121 14:32:36.716216 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-674b76c99f-zpcvf" event={"ID":"881842b1-c889-41be-b92a-4690f665e361","Type":"ContainerStarted","Data":"41554094680b0dd662bd112e1c1b06c186ecb0f5a61bbe9834ce77c7d1b4ff1a"} Nov 21 14:32:36 crc kubenswrapper[4897]: I1121 14:32:36.724130 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"cdf9e87f-440b-41db-a8d1-f3f5f8953310","Type":"ContainerStarted","Data":"e429315940c298bee2e17ab5392dc9e4768e1620eb12d003f3e776012cb0dcff"} Nov 21 14:32:36 crc kubenswrapper[4897]: I1121 14:32:36.724656 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 21 14:32:36 crc kubenswrapper[4897]: I1121 14:32:36.724711 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 21 14:32:36 crc kubenswrapper[4897]: I1121 14:32:36.724728 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 21 14:32:36 crc kubenswrapper[4897]: I1121 14:32:36.724745 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 21 14:32:37 crc kubenswrapper[4897]: I1121 14:32:37.755019 4897 generic.go:334] "Generic (PLEG): container finished" podID="881842b1-c889-41be-b92a-4690f665e361" containerID="d620b5b8baa9dcb6069695a000305aaac4895ad653cdba53e611946149acab93" exitCode=0 Nov 21 14:32:37 crc kubenswrapper[4897]: I1121 14:32:37.755119 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-674b76c99f-zpcvf" event={"ID":"881842b1-c889-41be-b92a-4690f665e361","Type":"ContainerDied","Data":"d620b5b8baa9dcb6069695a000305aaac4895ad653cdba53e611946149acab93"} Nov 21 14:32:37 crc kubenswrapper[4897]: I1121 14:32:37.762727 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"cdf9e87f-440b-41db-a8d1-f3f5f8953310","Type":"ContainerStarted","Data":"8f667b9a8773f8a967c628df8a268f6bc203eedd9b3a74c8e6c7b84a0f55a894"} Nov 21 14:32:37 crc kubenswrapper[4897]: I1121 14:32:37.783430 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Nov 21 14:32:38 crc kubenswrapper[4897]: I1121 14:32:38.777379 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"79802b6b-53b5-4568-b79b-83da7dbbf123","Type":"ContainerStarted","Data":"cebc20d99433055d6fe1f45416a2f0a69cbf7883d422fa7d34130d3f6586593a"} Nov 21 14:32:38 crc kubenswrapper[4897]: I1121 14:32:38.782144 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-674b76c99f-zpcvf" event={"ID":"881842b1-c889-41be-b92a-4690f665e361","Type":"ContainerStarted","Data":"ccda44133f50fc08043f235aaf8d793190f2ad7da677e86d25005b5384e7f032"} Nov 21 14:32:38 crc kubenswrapper[4897]: I1121 14:32:38.782276 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-674b76c99f-zpcvf" Nov 21 14:32:38 crc kubenswrapper[4897]: I1121 14:32:38.786174 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"cdf9e87f-440b-41db-a8d1-f3f5f8953310","Type":"ContainerStarted","Data":"7a14f7719519441504c7c2651ce49332d1db1d374b07cb9c63c4a9a12278116e"} Nov 21 14:32:38 crc kubenswrapper[4897]: I1121 14:32:38.786250 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Nov 21 14:32:38 crc kubenswrapper[4897]: I1121 14:32:38.786251 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="cdf9e87f-440b-41db-a8d1-f3f5f8953310" containerName="cinder-api-log" containerID="cri-o://8f667b9a8773f8a967c628df8a268f6bc203eedd9b3a74c8e6c7b84a0f55a894" gracePeriod=30 Nov 21 14:32:38 crc kubenswrapper[4897]: I1121 14:32:38.786253 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="cdf9e87f-440b-41db-a8d1-f3f5f8953310" containerName="cinder-api" containerID="cri-o://7a14f7719519441504c7c2651ce49332d1db1d374b07cb9c63c4a9a12278116e" gracePeriod=30 Nov 21 14:32:38 crc kubenswrapper[4897]: I1121 14:32:38.816830 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-674b76c99f-zpcvf" podStartSLOduration=4.816804575 podStartE2EDuration="4.816804575s" podCreationTimestamp="2025-11-21 14:32:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:32:38.805545203 +0000 UTC m=+1436.090138678" watchObservedRunningTime="2025-11-21 14:32:38.816804575 +0000 UTC m=+1436.101398050" Nov 21 14:32:38 crc kubenswrapper[4897]: I1121 14:32:38.833072 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.83304741 podStartE2EDuration="3.83304741s" podCreationTimestamp="2025-11-21 14:32:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:32:38.822526029 +0000 UTC m=+1436.107119504" watchObservedRunningTime="2025-11-21 14:32:38.83304741 +0000 UTC m=+1436.117640905" Nov 21 14:32:38 crc kubenswrapper[4897]: I1121 14:32:38.836824 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3dc8b9a3-65bf-436b-b968-20cade83734b","Type":"ContainerStarted","Data":"d480947193a7a84560fe60f69d8dfa6bf4beb45e84e6bc5f9af9448e0cd826ab"} Nov 21 14:32:38 crc kubenswrapper[4897]: I1121 14:32:38.837079 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 21 14:32:39 crc kubenswrapper[4897]: I1121 14:32:39.630218 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 21 14:32:39 crc kubenswrapper[4897]: I1121 14:32:39.657317 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=4.136703434 podStartE2EDuration="8.65729704s" podCreationTimestamp="2025-11-21 14:32:31 +0000 UTC" firstStartedPulling="2025-11-21 14:32:32.759965701 +0000 UTC m=+1430.044559176" lastFinishedPulling="2025-11-21 14:32:37.280559307 +0000 UTC m=+1434.565152782" observedRunningTime="2025-11-21 14:32:38.862442579 +0000 UTC m=+1436.147036054" watchObservedRunningTime="2025-11-21 14:32:39.65729704 +0000 UTC m=+1436.941890515" Nov 21 14:32:39 crc kubenswrapper[4897]: I1121 14:32:39.708481 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cdf9e87f-440b-41db-a8d1-f3f5f8953310-config-data\") pod \"cdf9e87f-440b-41db-a8d1-f3f5f8953310\" (UID: \"cdf9e87f-440b-41db-a8d1-f3f5f8953310\") " Nov 21 14:32:39 crc kubenswrapper[4897]: I1121 14:32:39.708575 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cdf9e87f-440b-41db-a8d1-f3f5f8953310-scripts\") pod \"cdf9e87f-440b-41db-a8d1-f3f5f8953310\" (UID: \"cdf9e87f-440b-41db-a8d1-f3f5f8953310\") " Nov 21 14:32:39 crc kubenswrapper[4897]: I1121 14:32:39.708630 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cdf9e87f-440b-41db-a8d1-f3f5f8953310-combined-ca-bundle\") pod \"cdf9e87f-440b-41db-a8d1-f3f5f8953310\" (UID: \"cdf9e87f-440b-41db-a8d1-f3f5f8953310\") " Nov 21 14:32:39 crc kubenswrapper[4897]: I1121 14:32:39.708721 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c4dm5\" (UniqueName: \"kubernetes.io/projected/cdf9e87f-440b-41db-a8d1-f3f5f8953310-kube-api-access-c4dm5\") pod \"cdf9e87f-440b-41db-a8d1-f3f5f8953310\" (UID: \"cdf9e87f-440b-41db-a8d1-f3f5f8953310\") " Nov 21 14:32:39 crc kubenswrapper[4897]: I1121 14:32:39.708807 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cdf9e87f-440b-41db-a8d1-f3f5f8953310-etc-machine-id\") pod \"cdf9e87f-440b-41db-a8d1-f3f5f8953310\" (UID: \"cdf9e87f-440b-41db-a8d1-f3f5f8953310\") " Nov 21 14:32:39 crc kubenswrapper[4897]: I1121 14:32:39.708826 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cdf9e87f-440b-41db-a8d1-f3f5f8953310-config-data-custom\") pod \"cdf9e87f-440b-41db-a8d1-f3f5f8953310\" (UID: \"cdf9e87f-440b-41db-a8d1-f3f5f8953310\") " Nov 21 14:32:39 crc kubenswrapper[4897]: I1121 14:32:39.708883 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cdf9e87f-440b-41db-a8d1-f3f5f8953310-logs\") pod \"cdf9e87f-440b-41db-a8d1-f3f5f8953310\" (UID: \"cdf9e87f-440b-41db-a8d1-f3f5f8953310\") " Nov 21 14:32:39 crc kubenswrapper[4897]: I1121 14:32:39.709529 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cdf9e87f-440b-41db-a8d1-f3f5f8953310-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "cdf9e87f-440b-41db-a8d1-f3f5f8953310" (UID: "cdf9e87f-440b-41db-a8d1-f3f5f8953310"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 21 14:32:39 crc kubenswrapper[4897]: I1121 14:32:39.709576 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cdf9e87f-440b-41db-a8d1-f3f5f8953310-logs" (OuterVolumeSpecName: "logs") pod "cdf9e87f-440b-41db-a8d1-f3f5f8953310" (UID: "cdf9e87f-440b-41db-a8d1-f3f5f8953310"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:32:39 crc kubenswrapper[4897]: I1121 14:32:39.720655 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cdf9e87f-440b-41db-a8d1-f3f5f8953310-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "cdf9e87f-440b-41db-a8d1-f3f5f8953310" (UID: "cdf9e87f-440b-41db-a8d1-f3f5f8953310"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:32:39 crc kubenswrapper[4897]: I1121 14:32:39.720738 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cdf9e87f-440b-41db-a8d1-f3f5f8953310-kube-api-access-c4dm5" (OuterVolumeSpecName: "kube-api-access-c4dm5") pod "cdf9e87f-440b-41db-a8d1-f3f5f8953310" (UID: "cdf9e87f-440b-41db-a8d1-f3f5f8953310"). InnerVolumeSpecName "kube-api-access-c4dm5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:32:39 crc kubenswrapper[4897]: I1121 14:32:39.728203 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cdf9e87f-440b-41db-a8d1-f3f5f8953310-scripts" (OuterVolumeSpecName: "scripts") pod "cdf9e87f-440b-41db-a8d1-f3f5f8953310" (UID: "cdf9e87f-440b-41db-a8d1-f3f5f8953310"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:32:39 crc kubenswrapper[4897]: I1121 14:32:39.784900 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cdf9e87f-440b-41db-a8d1-f3f5f8953310-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cdf9e87f-440b-41db-a8d1-f3f5f8953310" (UID: "cdf9e87f-440b-41db-a8d1-f3f5f8953310"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:32:39 crc kubenswrapper[4897]: I1121 14:32:39.786646 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cdf9e87f-440b-41db-a8d1-f3f5f8953310-config-data" (OuterVolumeSpecName: "config-data") pod "cdf9e87f-440b-41db-a8d1-f3f5f8953310" (UID: "cdf9e87f-440b-41db-a8d1-f3f5f8953310"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:32:39 crc kubenswrapper[4897]: I1121 14:32:39.811603 4897 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cdf9e87f-440b-41db-a8d1-f3f5f8953310-config-data\") on node \"crc\" DevicePath \"\"" Nov 21 14:32:39 crc kubenswrapper[4897]: I1121 14:32:39.811639 4897 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cdf9e87f-440b-41db-a8d1-f3f5f8953310-scripts\") on node \"crc\" DevicePath \"\"" Nov 21 14:32:39 crc kubenswrapper[4897]: I1121 14:32:39.811650 4897 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cdf9e87f-440b-41db-a8d1-f3f5f8953310-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 21 14:32:39 crc kubenswrapper[4897]: I1121 14:32:39.811664 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c4dm5\" (UniqueName: \"kubernetes.io/projected/cdf9e87f-440b-41db-a8d1-f3f5f8953310-kube-api-access-c4dm5\") on node \"crc\" DevicePath \"\"" Nov 21 14:32:39 crc kubenswrapper[4897]: I1121 14:32:39.811677 4897 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cdf9e87f-440b-41db-a8d1-f3f5f8953310-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 21 14:32:39 crc kubenswrapper[4897]: I1121 14:32:39.811686 4897 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cdf9e87f-440b-41db-a8d1-f3f5f8953310-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 21 14:32:39 crc kubenswrapper[4897]: I1121 14:32:39.811697 4897 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cdf9e87f-440b-41db-a8d1-f3f5f8953310-logs\") on node \"crc\" DevicePath \"\"" Nov 21 14:32:39 crc kubenswrapper[4897]: I1121 14:32:39.851177 4897 generic.go:334] "Generic (PLEG): container finished" podID="cdf9e87f-440b-41db-a8d1-f3f5f8953310" containerID="7a14f7719519441504c7c2651ce49332d1db1d374b07cb9c63c4a9a12278116e" exitCode=0 Nov 21 14:32:39 crc kubenswrapper[4897]: I1121 14:32:39.851219 4897 generic.go:334] "Generic (PLEG): container finished" podID="cdf9e87f-440b-41db-a8d1-f3f5f8953310" containerID="8f667b9a8773f8a967c628df8a268f6bc203eedd9b3a74c8e6c7b84a0f55a894" exitCode=143 Nov 21 14:32:39 crc kubenswrapper[4897]: I1121 14:32:39.851257 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"cdf9e87f-440b-41db-a8d1-f3f5f8953310","Type":"ContainerDied","Data":"7a14f7719519441504c7c2651ce49332d1db1d374b07cb9c63c4a9a12278116e"} Nov 21 14:32:39 crc kubenswrapper[4897]: I1121 14:32:39.851488 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"cdf9e87f-440b-41db-a8d1-f3f5f8953310","Type":"ContainerDied","Data":"8f667b9a8773f8a967c628df8a268f6bc203eedd9b3a74c8e6c7b84a0f55a894"} Nov 21 14:32:39 crc kubenswrapper[4897]: I1121 14:32:39.851529 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"cdf9e87f-440b-41db-a8d1-f3f5f8953310","Type":"ContainerDied","Data":"e429315940c298bee2e17ab5392dc9e4768e1620eb12d003f3e776012cb0dcff"} Nov 21 14:32:39 crc kubenswrapper[4897]: I1121 14:32:39.851551 4897 scope.go:117] "RemoveContainer" containerID="7a14f7719519441504c7c2651ce49332d1db1d374b07cb9c63c4a9a12278116e" Nov 21 14:32:39 crc kubenswrapper[4897]: I1121 14:32:39.852305 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 21 14:32:39 crc kubenswrapper[4897]: I1121 14:32:39.854021 4897 generic.go:334] "Generic (PLEG): container finished" podID="59fd52b7-f25b-4a81-a961-41e6cb526c89" containerID="7e1b93554f3c8ad375016f453d6de46e6965cbc1cf41b512bb5c0432ea3f9e3d" exitCode=0 Nov 21 14:32:39 crc kubenswrapper[4897]: I1121 14:32:39.854153 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-tbtcb" event={"ID":"59fd52b7-f25b-4a81-a961-41e6cb526c89","Type":"ContainerDied","Data":"7e1b93554f3c8ad375016f453d6de46e6965cbc1cf41b512bb5c0432ea3f9e3d"} Nov 21 14:32:39 crc kubenswrapper[4897]: I1121 14:32:39.862270 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"79802b6b-53b5-4568-b79b-83da7dbbf123","Type":"ContainerStarted","Data":"fb1780b07f00fff91feafdcf7e5944c12e5d238b232280791d8dc951cb74c7c3"} Nov 21 14:32:39 crc kubenswrapper[4897]: I1121 14:32:39.903890 4897 scope.go:117] "RemoveContainer" containerID="8f667b9a8773f8a967c628df8a268f6bc203eedd9b3a74c8e6c7b84a0f55a894" Nov 21 14:32:39 crc kubenswrapper[4897]: I1121 14:32:39.910120 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.451931588 podStartE2EDuration="5.91009765s" podCreationTimestamp="2025-11-21 14:32:34 +0000 UTC" firstStartedPulling="2025-11-21 14:32:35.946913495 +0000 UTC m=+1433.231506970" lastFinishedPulling="2025-11-21 14:32:37.405079557 +0000 UTC m=+1434.689673032" observedRunningTime="2025-11-21 14:32:39.898944331 +0000 UTC m=+1437.183537806" watchObservedRunningTime="2025-11-21 14:32:39.91009765 +0000 UTC m=+1437.194691125" Nov 21 14:32:39 crc kubenswrapper[4897]: I1121 14:32:39.938330 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Nov 21 14:32:39 crc kubenswrapper[4897]: I1121 14:32:39.948919 4897 scope.go:117] "RemoveContainer" containerID="7a14f7719519441504c7c2651ce49332d1db1d374b07cb9c63c4a9a12278116e" Nov 21 14:32:39 crc kubenswrapper[4897]: E1121 14:32:39.949940 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7a14f7719519441504c7c2651ce49332d1db1d374b07cb9c63c4a9a12278116e\": container with ID starting with 7a14f7719519441504c7c2651ce49332d1db1d374b07cb9c63c4a9a12278116e not found: ID does not exist" containerID="7a14f7719519441504c7c2651ce49332d1db1d374b07cb9c63c4a9a12278116e" Nov 21 14:32:39 crc kubenswrapper[4897]: I1121 14:32:39.949982 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a14f7719519441504c7c2651ce49332d1db1d374b07cb9c63c4a9a12278116e"} err="failed to get container status \"7a14f7719519441504c7c2651ce49332d1db1d374b07cb9c63c4a9a12278116e\": rpc error: code = NotFound desc = could not find container \"7a14f7719519441504c7c2651ce49332d1db1d374b07cb9c63c4a9a12278116e\": container with ID starting with 7a14f7719519441504c7c2651ce49332d1db1d374b07cb9c63c4a9a12278116e not found: ID does not exist" Nov 21 14:32:39 crc kubenswrapper[4897]: I1121 14:32:39.950014 4897 scope.go:117] "RemoveContainer" containerID="8f667b9a8773f8a967c628df8a268f6bc203eedd9b3a74c8e6c7b84a0f55a894" Nov 21 14:32:39 crc kubenswrapper[4897]: E1121 14:32:39.957564 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8f667b9a8773f8a967c628df8a268f6bc203eedd9b3a74c8e6c7b84a0f55a894\": container with ID starting with 8f667b9a8773f8a967c628df8a268f6bc203eedd9b3a74c8e6c7b84a0f55a894 not found: ID does not exist" containerID="8f667b9a8773f8a967c628df8a268f6bc203eedd9b3a74c8e6c7b84a0f55a894" Nov 21 14:32:39 crc kubenswrapper[4897]: I1121 14:32:39.957626 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f667b9a8773f8a967c628df8a268f6bc203eedd9b3a74c8e6c7b84a0f55a894"} err="failed to get container status \"8f667b9a8773f8a967c628df8a268f6bc203eedd9b3a74c8e6c7b84a0f55a894\": rpc error: code = NotFound desc = could not find container \"8f667b9a8773f8a967c628df8a268f6bc203eedd9b3a74c8e6c7b84a0f55a894\": container with ID starting with 8f667b9a8773f8a967c628df8a268f6bc203eedd9b3a74c8e6c7b84a0f55a894 not found: ID does not exist" Nov 21 14:32:39 crc kubenswrapper[4897]: I1121 14:32:39.957664 4897 scope.go:117] "RemoveContainer" containerID="7a14f7719519441504c7c2651ce49332d1db1d374b07cb9c63c4a9a12278116e" Nov 21 14:32:39 crc kubenswrapper[4897]: I1121 14:32:39.958780 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a14f7719519441504c7c2651ce49332d1db1d374b07cb9c63c4a9a12278116e"} err="failed to get container status \"7a14f7719519441504c7c2651ce49332d1db1d374b07cb9c63c4a9a12278116e\": rpc error: code = NotFound desc = could not find container \"7a14f7719519441504c7c2651ce49332d1db1d374b07cb9c63c4a9a12278116e\": container with ID starting with 7a14f7719519441504c7c2651ce49332d1db1d374b07cb9c63c4a9a12278116e not found: ID does not exist" Nov 21 14:32:39 crc kubenswrapper[4897]: I1121 14:32:39.958815 4897 scope.go:117] "RemoveContainer" containerID="8f667b9a8773f8a967c628df8a268f6bc203eedd9b3a74c8e6c7b84a0f55a894" Nov 21 14:32:39 crc kubenswrapper[4897]: I1121 14:32:39.962279 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f667b9a8773f8a967c628df8a268f6bc203eedd9b3a74c8e6c7b84a0f55a894"} err="failed to get container status \"8f667b9a8773f8a967c628df8a268f6bc203eedd9b3a74c8e6c7b84a0f55a894\": rpc error: code = NotFound desc = could not find container \"8f667b9a8773f8a967c628df8a268f6bc203eedd9b3a74c8e6c7b84a0f55a894\": container with ID starting with 8f667b9a8773f8a967c628df8a268f6bc203eedd9b3a74c8e6c7b84a0f55a894 not found: ID does not exist" Nov 21 14:32:39 crc kubenswrapper[4897]: I1121 14:32:39.964372 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Nov 21 14:32:39 crc kubenswrapper[4897]: I1121 14:32:39.994559 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Nov 21 14:32:39 crc kubenswrapper[4897]: E1121 14:32:39.995021 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cdf9e87f-440b-41db-a8d1-f3f5f8953310" containerName="cinder-api-log" Nov 21 14:32:39 crc kubenswrapper[4897]: I1121 14:32:39.995039 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="cdf9e87f-440b-41db-a8d1-f3f5f8953310" containerName="cinder-api-log" Nov 21 14:32:39 crc kubenswrapper[4897]: E1121 14:32:39.995096 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cdf9e87f-440b-41db-a8d1-f3f5f8953310" containerName="cinder-api" Nov 21 14:32:39 crc kubenswrapper[4897]: I1121 14:32:39.995102 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="cdf9e87f-440b-41db-a8d1-f3f5f8953310" containerName="cinder-api" Nov 21 14:32:39 crc kubenswrapper[4897]: I1121 14:32:39.995305 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="cdf9e87f-440b-41db-a8d1-f3f5f8953310" containerName="cinder-api-log" Nov 21 14:32:39 crc kubenswrapper[4897]: I1121 14:32:39.995320 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="cdf9e87f-440b-41db-a8d1-f3f5f8953310" containerName="cinder-api" Nov 21 14:32:39 crc kubenswrapper[4897]: I1121 14:32:39.996485 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 21 14:32:40 crc kubenswrapper[4897]: I1121 14:32:40.000965 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Nov 21 14:32:40 crc kubenswrapper[4897]: I1121 14:32:40.001048 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Nov 21 14:32:40 crc kubenswrapper[4897]: I1121 14:32:40.001385 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Nov 21 14:32:40 crc kubenswrapper[4897]: I1121 14:32:40.018869 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 21 14:32:40 crc kubenswrapper[4897]: I1121 14:32:40.019723 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8a67c019-b256-4890-8fb0-1dfa3a88b7b1-logs\") pod \"cinder-api-0\" (UID: \"8a67c019-b256-4890-8fb0-1dfa3a88b7b1\") " pod="openstack/cinder-api-0" Nov 21 14:32:40 crc kubenswrapper[4897]: I1121 14:32:40.019930 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b22qw\" (UniqueName: \"kubernetes.io/projected/8a67c019-b256-4890-8fb0-1dfa3a88b7b1-kube-api-access-b22qw\") pod \"cinder-api-0\" (UID: \"8a67c019-b256-4890-8fb0-1dfa3a88b7b1\") " pod="openstack/cinder-api-0" Nov 21 14:32:40 crc kubenswrapper[4897]: I1121 14:32:40.019957 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8a67c019-b256-4890-8fb0-1dfa3a88b7b1-config-data-custom\") pod \"cinder-api-0\" (UID: \"8a67c019-b256-4890-8fb0-1dfa3a88b7b1\") " pod="openstack/cinder-api-0" Nov 21 14:32:40 crc kubenswrapper[4897]: I1121 14:32:40.019992 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a67c019-b256-4890-8fb0-1dfa3a88b7b1-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"8a67c019-b256-4890-8fb0-1dfa3a88b7b1\") " pod="openstack/cinder-api-0" Nov 21 14:32:40 crc kubenswrapper[4897]: I1121 14:32:40.020032 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a67c019-b256-4890-8fb0-1dfa3a88b7b1-public-tls-certs\") pod \"cinder-api-0\" (UID: \"8a67c019-b256-4890-8fb0-1dfa3a88b7b1\") " pod="openstack/cinder-api-0" Nov 21 14:32:40 crc kubenswrapper[4897]: I1121 14:32:40.020163 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a67c019-b256-4890-8fb0-1dfa3a88b7b1-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"8a67c019-b256-4890-8fb0-1dfa3a88b7b1\") " pod="openstack/cinder-api-0" Nov 21 14:32:40 crc kubenswrapper[4897]: I1121 14:32:40.020289 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a67c019-b256-4890-8fb0-1dfa3a88b7b1-config-data\") pod \"cinder-api-0\" (UID: \"8a67c019-b256-4890-8fb0-1dfa3a88b7b1\") " pod="openstack/cinder-api-0" Nov 21 14:32:40 crc kubenswrapper[4897]: I1121 14:32:40.020414 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a67c019-b256-4890-8fb0-1dfa3a88b7b1-scripts\") pod \"cinder-api-0\" (UID: \"8a67c019-b256-4890-8fb0-1dfa3a88b7b1\") " pod="openstack/cinder-api-0" Nov 21 14:32:40 crc kubenswrapper[4897]: I1121 14:32:40.020594 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8a67c019-b256-4890-8fb0-1dfa3a88b7b1-etc-machine-id\") pod \"cinder-api-0\" (UID: \"8a67c019-b256-4890-8fb0-1dfa3a88b7b1\") " pod="openstack/cinder-api-0" Nov 21 14:32:40 crc kubenswrapper[4897]: I1121 14:32:40.107488 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cdf9e87f-440b-41db-a8d1-f3f5f8953310" path="/var/lib/kubelet/pods/cdf9e87f-440b-41db-a8d1-f3f5f8953310/volumes" Nov 21 14:32:40 crc kubenswrapper[4897]: I1121 14:32:40.123161 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a67c019-b256-4890-8fb0-1dfa3a88b7b1-config-data\") pod \"cinder-api-0\" (UID: \"8a67c019-b256-4890-8fb0-1dfa3a88b7b1\") " pod="openstack/cinder-api-0" Nov 21 14:32:40 crc kubenswrapper[4897]: I1121 14:32:40.123251 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a67c019-b256-4890-8fb0-1dfa3a88b7b1-scripts\") pod \"cinder-api-0\" (UID: \"8a67c019-b256-4890-8fb0-1dfa3a88b7b1\") " pod="openstack/cinder-api-0" Nov 21 14:32:40 crc kubenswrapper[4897]: I1121 14:32:40.123287 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8a67c019-b256-4890-8fb0-1dfa3a88b7b1-etc-machine-id\") pod \"cinder-api-0\" (UID: \"8a67c019-b256-4890-8fb0-1dfa3a88b7b1\") " pod="openstack/cinder-api-0" Nov 21 14:32:40 crc kubenswrapper[4897]: I1121 14:32:40.123320 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8a67c019-b256-4890-8fb0-1dfa3a88b7b1-logs\") pod \"cinder-api-0\" (UID: \"8a67c019-b256-4890-8fb0-1dfa3a88b7b1\") " pod="openstack/cinder-api-0" Nov 21 14:32:40 crc kubenswrapper[4897]: I1121 14:32:40.123395 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8a67c019-b256-4890-8fb0-1dfa3a88b7b1-config-data-custom\") pod \"cinder-api-0\" (UID: \"8a67c019-b256-4890-8fb0-1dfa3a88b7b1\") " pod="openstack/cinder-api-0" Nov 21 14:32:40 crc kubenswrapper[4897]: I1121 14:32:40.123411 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b22qw\" (UniqueName: \"kubernetes.io/projected/8a67c019-b256-4890-8fb0-1dfa3a88b7b1-kube-api-access-b22qw\") pod \"cinder-api-0\" (UID: \"8a67c019-b256-4890-8fb0-1dfa3a88b7b1\") " pod="openstack/cinder-api-0" Nov 21 14:32:40 crc kubenswrapper[4897]: I1121 14:32:40.123429 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a67c019-b256-4890-8fb0-1dfa3a88b7b1-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"8a67c019-b256-4890-8fb0-1dfa3a88b7b1\") " pod="openstack/cinder-api-0" Nov 21 14:32:40 crc kubenswrapper[4897]: I1121 14:32:40.123444 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a67c019-b256-4890-8fb0-1dfa3a88b7b1-public-tls-certs\") pod \"cinder-api-0\" (UID: \"8a67c019-b256-4890-8fb0-1dfa3a88b7b1\") " pod="openstack/cinder-api-0" Nov 21 14:32:40 crc kubenswrapper[4897]: I1121 14:32:40.123533 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a67c019-b256-4890-8fb0-1dfa3a88b7b1-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"8a67c019-b256-4890-8fb0-1dfa3a88b7b1\") " pod="openstack/cinder-api-0" Nov 21 14:32:40 crc kubenswrapper[4897]: I1121 14:32:40.125417 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8a67c019-b256-4890-8fb0-1dfa3a88b7b1-etc-machine-id\") pod \"cinder-api-0\" (UID: \"8a67c019-b256-4890-8fb0-1dfa3a88b7b1\") " pod="openstack/cinder-api-0" Nov 21 14:32:40 crc kubenswrapper[4897]: I1121 14:32:40.125763 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8a67c019-b256-4890-8fb0-1dfa3a88b7b1-logs\") pod \"cinder-api-0\" (UID: \"8a67c019-b256-4890-8fb0-1dfa3a88b7b1\") " pod="openstack/cinder-api-0" Nov 21 14:32:40 crc kubenswrapper[4897]: I1121 14:32:40.130540 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a67c019-b256-4890-8fb0-1dfa3a88b7b1-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"8a67c019-b256-4890-8fb0-1dfa3a88b7b1\") " pod="openstack/cinder-api-0" Nov 21 14:32:40 crc kubenswrapper[4897]: I1121 14:32:40.130617 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8a67c019-b256-4890-8fb0-1dfa3a88b7b1-config-data-custom\") pod \"cinder-api-0\" (UID: \"8a67c019-b256-4890-8fb0-1dfa3a88b7b1\") " pod="openstack/cinder-api-0" Nov 21 14:32:40 crc kubenswrapper[4897]: I1121 14:32:40.131429 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a67c019-b256-4890-8fb0-1dfa3a88b7b1-config-data\") pod \"cinder-api-0\" (UID: \"8a67c019-b256-4890-8fb0-1dfa3a88b7b1\") " pod="openstack/cinder-api-0" Nov 21 14:32:40 crc kubenswrapper[4897]: I1121 14:32:40.131806 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a67c019-b256-4890-8fb0-1dfa3a88b7b1-scripts\") pod \"cinder-api-0\" (UID: \"8a67c019-b256-4890-8fb0-1dfa3a88b7b1\") " pod="openstack/cinder-api-0" Nov 21 14:32:40 crc kubenswrapper[4897]: I1121 14:32:40.132047 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a67c019-b256-4890-8fb0-1dfa3a88b7b1-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"8a67c019-b256-4890-8fb0-1dfa3a88b7b1\") " pod="openstack/cinder-api-0" Nov 21 14:32:40 crc kubenswrapper[4897]: I1121 14:32:40.134612 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a67c019-b256-4890-8fb0-1dfa3a88b7b1-public-tls-certs\") pod \"cinder-api-0\" (UID: \"8a67c019-b256-4890-8fb0-1dfa3a88b7b1\") " pod="openstack/cinder-api-0" Nov 21 14:32:40 crc kubenswrapper[4897]: I1121 14:32:40.143183 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b22qw\" (UniqueName: \"kubernetes.io/projected/8a67c019-b256-4890-8fb0-1dfa3a88b7b1-kube-api-access-b22qw\") pod \"cinder-api-0\" (UID: \"8a67c019-b256-4890-8fb0-1dfa3a88b7b1\") " pod="openstack/cinder-api-0" Nov 21 14:32:40 crc kubenswrapper[4897]: I1121 14:32:40.276655 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Nov 21 14:32:40 crc kubenswrapper[4897]: I1121 14:32:40.325523 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 21 14:32:40 crc kubenswrapper[4897]: I1121 14:32:40.901699 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 21 14:32:40 crc kubenswrapper[4897]: W1121 14:32:40.907403 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8a67c019_b256_4890_8fb0_1dfa3a88b7b1.slice/crio-27df56e7388cce02cc3d4f28534f8adb8fca7e8602c0cc1274d82d9ce2f16cdb WatchSource:0}: Error finding container 27df56e7388cce02cc3d4f28534f8adb8fca7e8602c0cc1274d82d9ce2f16cdb: Status 404 returned error can't find the container with id 27df56e7388cce02cc3d4f28534f8adb8fca7e8602c0cc1274d82d9ce2f16cdb Nov 21 14:32:41 crc kubenswrapper[4897]: I1121 14:32:41.463753 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-tbtcb" Nov 21 14:32:41 crc kubenswrapper[4897]: I1121 14:32:41.589820 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 21 14:32:41 crc kubenswrapper[4897]: I1121 14:32:41.589956 4897 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 21 14:32:41 crc kubenswrapper[4897]: I1121 14:32:41.592155 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p4csc\" (UniqueName: \"kubernetes.io/projected/59fd52b7-f25b-4a81-a961-41e6cb526c89-kube-api-access-p4csc\") pod \"59fd52b7-f25b-4a81-a961-41e6cb526c89\" (UID: \"59fd52b7-f25b-4a81-a961-41e6cb526c89\") " Nov 21 14:32:41 crc kubenswrapper[4897]: I1121 14:32:41.592396 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59fd52b7-f25b-4a81-a961-41e6cb526c89-combined-ca-bundle\") pod \"59fd52b7-f25b-4a81-a961-41e6cb526c89\" (UID: \"59fd52b7-f25b-4a81-a961-41e6cb526c89\") " Nov 21 14:32:41 crc kubenswrapper[4897]: I1121 14:32:41.592476 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59fd52b7-f25b-4a81-a961-41e6cb526c89-config-data\") pod \"59fd52b7-f25b-4a81-a961-41e6cb526c89\" (UID: \"59fd52b7-f25b-4a81-a961-41e6cb526c89\") " Nov 21 14:32:41 crc kubenswrapper[4897]: I1121 14:32:41.598892 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 21 14:32:41 crc kubenswrapper[4897]: I1121 14:32:41.599011 4897 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 21 14:32:41 crc kubenswrapper[4897]: I1121 14:32:41.606684 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/59fd52b7-f25b-4a81-a961-41e6cb526c89-kube-api-access-p4csc" (OuterVolumeSpecName: "kube-api-access-p4csc") pod "59fd52b7-f25b-4a81-a961-41e6cb526c89" (UID: "59fd52b7-f25b-4a81-a961-41e6cb526c89"). InnerVolumeSpecName "kube-api-access-p4csc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:32:41 crc kubenswrapper[4897]: I1121 14:32:41.632415 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 21 14:32:41 crc kubenswrapper[4897]: I1121 14:32:41.633518 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 21 14:32:41 crc kubenswrapper[4897]: I1121 14:32:41.695219 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p4csc\" (UniqueName: \"kubernetes.io/projected/59fd52b7-f25b-4a81-a961-41e6cb526c89-kube-api-access-p4csc\") on node \"crc\" DevicePath \"\"" Nov 21 14:32:41 crc kubenswrapper[4897]: I1121 14:32:41.731654 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/59fd52b7-f25b-4a81-a961-41e6cb526c89-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "59fd52b7-f25b-4a81-a961-41e6cb526c89" (UID: "59fd52b7-f25b-4a81-a961-41e6cb526c89"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:32:41 crc kubenswrapper[4897]: I1121 14:32:41.776049 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/59fd52b7-f25b-4a81-a961-41e6cb526c89-config-data" (OuterVolumeSpecName: "config-data") pod "59fd52b7-f25b-4a81-a961-41e6cb526c89" (UID: "59fd52b7-f25b-4a81-a961-41e6cb526c89"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:32:41 crc kubenswrapper[4897]: I1121 14:32:41.796926 4897 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59fd52b7-f25b-4a81-a961-41e6cb526c89-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 21 14:32:41 crc kubenswrapper[4897]: I1121 14:32:41.796956 4897 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59fd52b7-f25b-4a81-a961-41e6cb526c89-config-data\") on node \"crc\" DevicePath \"\"" Nov 21 14:32:41 crc kubenswrapper[4897]: I1121 14:32:41.914717 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"8a67c019-b256-4890-8fb0-1dfa3a88b7b1","Type":"ContainerStarted","Data":"27df56e7388cce02cc3d4f28534f8adb8fca7e8602c0cc1274d82d9ce2f16cdb"} Nov 21 14:32:41 crc kubenswrapper[4897]: I1121 14:32:41.939588 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-tbtcb" Nov 21 14:32:41 crc kubenswrapper[4897]: I1121 14:32:41.942031 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-tbtcb" event={"ID":"59fd52b7-f25b-4a81-a961-41e6cb526c89","Type":"ContainerDied","Data":"cc0f36880c4fc19ca7daede88714edef43c228c9382b7a3bd03a7c5b9e7808fa"} Nov 21 14:32:41 crc kubenswrapper[4897]: I1121 14:32:41.942068 4897 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cc0f36880c4fc19ca7daede88714edef43c228c9382b7a3bd03a7c5b9e7808fa" Nov 21 14:32:42 crc kubenswrapper[4897]: I1121 14:32:42.954311 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"8a67c019-b256-4890-8fb0-1dfa3a88b7b1","Type":"ContainerStarted","Data":"1fa2bf55991a722051532e5b52139a67cda11eebc067170ddecefd409a0a63bb"} Nov 21 14:32:44 crc kubenswrapper[4897]: I1121 14:32:44.392847 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-77b5d6c768-mnz58" Nov 21 14:32:44 crc kubenswrapper[4897]: I1121 14:32:44.628089 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-8974ffc64-c68b5" Nov 21 14:32:44 crc kubenswrapper[4897]: I1121 14:32:44.663674 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-8974ffc64-c68b5" Nov 21 14:32:45 crc kubenswrapper[4897]: I1121 14:32:45.041442 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"8a67c019-b256-4890-8fb0-1dfa3a88b7b1","Type":"ContainerStarted","Data":"a2af93d39fdd11e12f6bcf2021113335f6f9e553aa057d7e5e448887d7be95c9"} Nov 21 14:32:45 crc kubenswrapper[4897]: I1121 14:32:45.041952 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Nov 21 14:32:45 crc kubenswrapper[4897]: I1121 14:32:45.103742 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=6.103720861 podStartE2EDuration="6.103720861s" podCreationTimestamp="2025-11-21 14:32:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:32:45.095999653 +0000 UTC m=+1442.380593128" watchObservedRunningTime="2025-11-21 14:32:45.103720861 +0000 UTC m=+1442.388314326" Nov 21 14:32:45 crc kubenswrapper[4897]: I1121 14:32:45.403524 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-674b76c99f-zpcvf" Nov 21 14:32:45 crc kubenswrapper[4897]: I1121 14:32:45.484753 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d66f584d7-hxs7n"] Nov 21 14:32:45 crc kubenswrapper[4897]: I1121 14:32:45.488097 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6d66f584d7-hxs7n" podUID="4b23c32f-2539-41a4-b994-eb4210356a13" containerName="dnsmasq-dns" containerID="cri-o://be5d18a893b03eba20c0f6a73a0c3f89b6a9c8d640286ef155ce37dffd3a7719" gracePeriod=10 Nov 21 14:32:45 crc kubenswrapper[4897]: I1121 14:32:45.730367 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/cinder-scheduler-0" podUID="79802b6b-53b5-4568-b79b-83da7dbbf123" containerName="cinder-scheduler" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 21 14:32:46 crc kubenswrapper[4897]: I1121 14:32:46.052747 4897 generic.go:334] "Generic (PLEG): container finished" podID="4b23c32f-2539-41a4-b994-eb4210356a13" containerID="be5d18a893b03eba20c0f6a73a0c3f89b6a9c8d640286ef155ce37dffd3a7719" exitCode=0 Nov 21 14:32:46 crc kubenswrapper[4897]: I1121 14:32:46.053388 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d66f584d7-hxs7n" event={"ID":"4b23c32f-2539-41a4-b994-eb4210356a13","Type":"ContainerDied","Data":"be5d18a893b03eba20c0f6a73a0c3f89b6a9c8d640286ef155ce37dffd3a7719"} Nov 21 14:32:46 crc kubenswrapper[4897]: I1121 14:32:46.632386 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d66f584d7-hxs7n" Nov 21 14:32:46 crc kubenswrapper[4897]: I1121 14:32:46.824043 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b23c32f-2539-41a4-b994-eb4210356a13-config\") pod \"4b23c32f-2539-41a4-b994-eb4210356a13\" (UID: \"4b23c32f-2539-41a4-b994-eb4210356a13\") " Nov 21 14:32:46 crc kubenswrapper[4897]: I1121 14:32:46.824177 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4b23c32f-2539-41a4-b994-eb4210356a13-dns-swift-storage-0\") pod \"4b23c32f-2539-41a4-b994-eb4210356a13\" (UID: \"4b23c32f-2539-41a4-b994-eb4210356a13\") " Nov 21 14:32:46 crc kubenswrapper[4897]: I1121 14:32:46.824224 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4b23c32f-2539-41a4-b994-eb4210356a13-ovsdbserver-nb\") pod \"4b23c32f-2539-41a4-b994-eb4210356a13\" (UID: \"4b23c32f-2539-41a4-b994-eb4210356a13\") " Nov 21 14:32:46 crc kubenswrapper[4897]: I1121 14:32:46.824257 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4b23c32f-2539-41a4-b994-eb4210356a13-ovsdbserver-sb\") pod \"4b23c32f-2539-41a4-b994-eb4210356a13\" (UID: \"4b23c32f-2539-41a4-b994-eb4210356a13\") " Nov 21 14:32:46 crc kubenswrapper[4897]: I1121 14:32:46.824298 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4xtl2\" (UniqueName: \"kubernetes.io/projected/4b23c32f-2539-41a4-b994-eb4210356a13-kube-api-access-4xtl2\") pod \"4b23c32f-2539-41a4-b994-eb4210356a13\" (UID: \"4b23c32f-2539-41a4-b994-eb4210356a13\") " Nov 21 14:32:46 crc kubenswrapper[4897]: I1121 14:32:46.824326 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4b23c32f-2539-41a4-b994-eb4210356a13-dns-svc\") pod \"4b23c32f-2539-41a4-b994-eb4210356a13\" (UID: \"4b23c32f-2539-41a4-b994-eb4210356a13\") " Nov 21 14:32:46 crc kubenswrapper[4897]: I1121 14:32:46.831118 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b23c32f-2539-41a4-b994-eb4210356a13-kube-api-access-4xtl2" (OuterVolumeSpecName: "kube-api-access-4xtl2") pod "4b23c32f-2539-41a4-b994-eb4210356a13" (UID: "4b23c32f-2539-41a4-b994-eb4210356a13"). InnerVolumeSpecName "kube-api-access-4xtl2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:32:46 crc kubenswrapper[4897]: I1121 14:32:46.881493 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4b23c32f-2539-41a4-b994-eb4210356a13-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "4b23c32f-2539-41a4-b994-eb4210356a13" (UID: "4b23c32f-2539-41a4-b994-eb4210356a13"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:32:46 crc kubenswrapper[4897]: I1121 14:32:46.900834 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4b23c32f-2539-41a4-b994-eb4210356a13-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "4b23c32f-2539-41a4-b994-eb4210356a13" (UID: "4b23c32f-2539-41a4-b994-eb4210356a13"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:32:46 crc kubenswrapper[4897]: I1121 14:32:46.904051 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4b23c32f-2539-41a4-b994-eb4210356a13-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "4b23c32f-2539-41a4-b994-eb4210356a13" (UID: "4b23c32f-2539-41a4-b994-eb4210356a13"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:32:46 crc kubenswrapper[4897]: I1121 14:32:46.909747 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4b23c32f-2539-41a4-b994-eb4210356a13-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4b23c32f-2539-41a4-b994-eb4210356a13" (UID: "4b23c32f-2539-41a4-b994-eb4210356a13"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:32:46 crc kubenswrapper[4897]: I1121 14:32:46.912128 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4b23c32f-2539-41a4-b994-eb4210356a13-config" (OuterVolumeSpecName: "config") pod "4b23c32f-2539-41a4-b994-eb4210356a13" (UID: "4b23c32f-2539-41a4-b994-eb4210356a13"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:32:46 crc kubenswrapper[4897]: I1121 14:32:46.927166 4897 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4b23c32f-2539-41a4-b994-eb4210356a13-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 21 14:32:46 crc kubenswrapper[4897]: I1121 14:32:46.927219 4897 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4b23c32f-2539-41a4-b994-eb4210356a13-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 21 14:32:46 crc kubenswrapper[4897]: I1121 14:32:46.927234 4897 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4b23c32f-2539-41a4-b994-eb4210356a13-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 21 14:32:46 crc kubenswrapper[4897]: I1121 14:32:46.927245 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4xtl2\" (UniqueName: \"kubernetes.io/projected/4b23c32f-2539-41a4-b994-eb4210356a13-kube-api-access-4xtl2\") on node \"crc\" DevicePath \"\"" Nov 21 14:32:46 crc kubenswrapper[4897]: I1121 14:32:46.927263 4897 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4b23c32f-2539-41a4-b994-eb4210356a13-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 21 14:32:46 crc kubenswrapper[4897]: I1121 14:32:46.927276 4897 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b23c32f-2539-41a4-b994-eb4210356a13-config\") on node \"crc\" DevicePath \"\"" Nov 21 14:32:47 crc kubenswrapper[4897]: I1121 14:32:47.066210 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d66f584d7-hxs7n" event={"ID":"4b23c32f-2539-41a4-b994-eb4210356a13","Type":"ContainerDied","Data":"6fd563e868da419d7571f71309d8dd05c6238bbd22da10b375c366fcd7059388"} Nov 21 14:32:47 crc kubenswrapper[4897]: I1121 14:32:47.066273 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d66f584d7-hxs7n" Nov 21 14:32:47 crc kubenswrapper[4897]: I1121 14:32:47.066283 4897 scope.go:117] "RemoveContainer" containerID="be5d18a893b03eba20c0f6a73a0c3f89b6a9c8d640286ef155ce37dffd3a7719" Nov 21 14:32:47 crc kubenswrapper[4897]: I1121 14:32:47.106390 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d66f584d7-hxs7n"] Nov 21 14:32:47 crc kubenswrapper[4897]: I1121 14:32:47.112583 4897 scope.go:117] "RemoveContainer" containerID="311550bf15cef4574086c7e01f62b0ad80c20ddb0d2e386febbce832e4ff6449" Nov 21 14:32:47 crc kubenswrapper[4897]: I1121 14:32:47.136724 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6d66f584d7-hxs7n"] Nov 21 14:32:48 crc kubenswrapper[4897]: I1121 14:32:48.076898 4897 generic.go:334] "Generic (PLEG): container finished" podID="4a683589-a940-405b-afe1-1838f34dac9a" containerID="b3b649aa1cd6febd3c35ff713d0112d349fcbe8d37fb13176de414dfeca6b26c" exitCode=0 Nov 21 14:32:48 crc kubenswrapper[4897]: I1121 14:32:48.076978 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-mrctr" event={"ID":"4a683589-a940-405b-afe1-1838f34dac9a","Type":"ContainerDied","Data":"b3b649aa1cd6febd3c35ff713d0112d349fcbe8d37fb13176de414dfeca6b26c"} Nov 21 14:32:48 crc kubenswrapper[4897]: I1121 14:32:48.101994 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4b23c32f-2539-41a4-b994-eb4210356a13" path="/var/lib/kubelet/pods/4b23c32f-2539-41a4-b994-eb4210356a13/volumes" Nov 21 14:32:49 crc kubenswrapper[4897]: I1121 14:32:49.381495 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Nov 21 14:32:49 crc kubenswrapper[4897]: E1121 14:32:49.382742 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b23c32f-2539-41a4-b994-eb4210356a13" containerName="dnsmasq-dns" Nov 21 14:32:49 crc kubenswrapper[4897]: I1121 14:32:49.382761 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b23c32f-2539-41a4-b994-eb4210356a13" containerName="dnsmasq-dns" Nov 21 14:32:49 crc kubenswrapper[4897]: E1121 14:32:49.382787 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b23c32f-2539-41a4-b994-eb4210356a13" containerName="init" Nov 21 14:32:49 crc kubenswrapper[4897]: I1121 14:32:49.382795 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b23c32f-2539-41a4-b994-eb4210356a13" containerName="init" Nov 21 14:32:49 crc kubenswrapper[4897]: E1121 14:32:49.382827 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59fd52b7-f25b-4a81-a961-41e6cb526c89" containerName="heat-db-sync" Nov 21 14:32:49 crc kubenswrapper[4897]: I1121 14:32:49.382836 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="59fd52b7-f25b-4a81-a961-41e6cb526c89" containerName="heat-db-sync" Nov 21 14:32:49 crc kubenswrapper[4897]: I1121 14:32:49.383107 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="59fd52b7-f25b-4a81-a961-41e6cb526c89" containerName="heat-db-sync" Nov 21 14:32:49 crc kubenswrapper[4897]: I1121 14:32:49.383141 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b23c32f-2539-41a4-b994-eb4210356a13" containerName="dnsmasq-dns" Nov 21 14:32:49 crc kubenswrapper[4897]: I1121 14:32:49.384181 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 21 14:32:49 crc kubenswrapper[4897]: I1121 14:32:49.392020 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-js4hq" Nov 21 14:32:49 crc kubenswrapper[4897]: I1121 14:32:49.392236 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Nov 21 14:32:49 crc kubenswrapper[4897]: I1121 14:32:49.392357 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Nov 21 14:32:49 crc kubenswrapper[4897]: I1121 14:32:49.398812 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 21 14:32:49 crc kubenswrapper[4897]: I1121 14:32:49.485119 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/23323788-338f-42c8-817a-e18aa170417d-openstack-config\") pod \"openstackclient\" (UID: \"23323788-338f-42c8-817a-e18aa170417d\") " pod="openstack/openstackclient" Nov 21 14:32:49 crc kubenswrapper[4897]: I1121 14:32:49.485445 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/23323788-338f-42c8-817a-e18aa170417d-openstack-config-secret\") pod \"openstackclient\" (UID: \"23323788-338f-42c8-817a-e18aa170417d\") " pod="openstack/openstackclient" Nov 21 14:32:49 crc kubenswrapper[4897]: I1121 14:32:49.485643 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23323788-338f-42c8-817a-e18aa170417d-combined-ca-bundle\") pod \"openstackclient\" (UID: \"23323788-338f-42c8-817a-e18aa170417d\") " pod="openstack/openstackclient" Nov 21 14:32:49 crc kubenswrapper[4897]: I1121 14:32:49.485668 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b2xzx\" (UniqueName: \"kubernetes.io/projected/23323788-338f-42c8-817a-e18aa170417d-kube-api-access-b2xzx\") pod \"openstackclient\" (UID: \"23323788-338f-42c8-817a-e18aa170417d\") " pod="openstack/openstackclient" Nov 21 14:32:49 crc kubenswrapper[4897]: I1121 14:32:49.586986 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b2xzx\" (UniqueName: \"kubernetes.io/projected/23323788-338f-42c8-817a-e18aa170417d-kube-api-access-b2xzx\") pod \"openstackclient\" (UID: \"23323788-338f-42c8-817a-e18aa170417d\") " pod="openstack/openstackclient" Nov 21 14:32:49 crc kubenswrapper[4897]: I1121 14:32:49.587100 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/23323788-338f-42c8-817a-e18aa170417d-openstack-config\") pod \"openstackclient\" (UID: \"23323788-338f-42c8-817a-e18aa170417d\") " pod="openstack/openstackclient" Nov 21 14:32:49 crc kubenswrapper[4897]: I1121 14:32:49.587134 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/23323788-338f-42c8-817a-e18aa170417d-openstack-config-secret\") pod \"openstackclient\" (UID: \"23323788-338f-42c8-817a-e18aa170417d\") " pod="openstack/openstackclient" Nov 21 14:32:49 crc kubenswrapper[4897]: I1121 14:32:49.587241 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23323788-338f-42c8-817a-e18aa170417d-combined-ca-bundle\") pod \"openstackclient\" (UID: \"23323788-338f-42c8-817a-e18aa170417d\") " pod="openstack/openstackclient" Nov 21 14:32:49 crc kubenswrapper[4897]: I1121 14:32:49.588577 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/23323788-338f-42c8-817a-e18aa170417d-openstack-config\") pod \"openstackclient\" (UID: \"23323788-338f-42c8-817a-e18aa170417d\") " pod="openstack/openstackclient" Nov 21 14:32:49 crc kubenswrapper[4897]: I1121 14:32:49.592042 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-mrctr" Nov 21 14:32:49 crc kubenswrapper[4897]: I1121 14:32:49.593600 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23323788-338f-42c8-817a-e18aa170417d-combined-ca-bundle\") pod \"openstackclient\" (UID: \"23323788-338f-42c8-817a-e18aa170417d\") " pod="openstack/openstackclient" Nov 21 14:32:49 crc kubenswrapper[4897]: I1121 14:32:49.594080 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/23323788-338f-42c8-817a-e18aa170417d-openstack-config-secret\") pod \"openstackclient\" (UID: \"23323788-338f-42c8-817a-e18aa170417d\") " pod="openstack/openstackclient" Nov 21 14:32:49 crc kubenswrapper[4897]: I1121 14:32:49.602623 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b2xzx\" (UniqueName: \"kubernetes.io/projected/23323788-338f-42c8-817a-e18aa170417d-kube-api-access-b2xzx\") pod \"openstackclient\" (UID: \"23323788-338f-42c8-817a-e18aa170417d\") " pod="openstack/openstackclient" Nov 21 14:32:49 crc kubenswrapper[4897]: I1121 14:32:49.690463 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a683589-a940-405b-afe1-1838f34dac9a-combined-ca-bundle\") pod \"4a683589-a940-405b-afe1-1838f34dac9a\" (UID: \"4a683589-a940-405b-afe1-1838f34dac9a\") " Nov 21 14:32:49 crc kubenswrapper[4897]: I1121 14:32:49.690533 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/4a683589-a940-405b-afe1-1838f34dac9a-config\") pod \"4a683589-a940-405b-afe1-1838f34dac9a\" (UID: \"4a683589-a940-405b-afe1-1838f34dac9a\") " Nov 21 14:32:49 crc kubenswrapper[4897]: I1121 14:32:49.690771 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t7rdp\" (UniqueName: \"kubernetes.io/projected/4a683589-a940-405b-afe1-1838f34dac9a-kube-api-access-t7rdp\") pod \"4a683589-a940-405b-afe1-1838f34dac9a\" (UID: \"4a683589-a940-405b-afe1-1838f34dac9a\") " Nov 21 14:32:49 crc kubenswrapper[4897]: I1121 14:32:49.696242 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4a683589-a940-405b-afe1-1838f34dac9a-kube-api-access-t7rdp" (OuterVolumeSpecName: "kube-api-access-t7rdp") pod "4a683589-a940-405b-afe1-1838f34dac9a" (UID: "4a683589-a940-405b-afe1-1838f34dac9a"). InnerVolumeSpecName "kube-api-access-t7rdp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:32:49 crc kubenswrapper[4897]: I1121 14:32:49.726974 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 21 14:32:49 crc kubenswrapper[4897]: I1121 14:32:49.732848 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a683589-a940-405b-afe1-1838f34dac9a-config" (OuterVolumeSpecName: "config") pod "4a683589-a940-405b-afe1-1838f34dac9a" (UID: "4a683589-a940-405b-afe1-1838f34dac9a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:32:49 crc kubenswrapper[4897]: I1121 14:32:49.736228 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a683589-a940-405b-afe1-1838f34dac9a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4a683589-a940-405b-afe1-1838f34dac9a" (UID: "4a683589-a940-405b-afe1-1838f34dac9a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:32:49 crc kubenswrapper[4897]: I1121 14:32:49.792847 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t7rdp\" (UniqueName: \"kubernetes.io/projected/4a683589-a940-405b-afe1-1838f34dac9a-kube-api-access-t7rdp\") on node \"crc\" DevicePath \"\"" Nov 21 14:32:49 crc kubenswrapper[4897]: I1121 14:32:49.792884 4897 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a683589-a940-405b-afe1-1838f34dac9a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 21 14:32:49 crc kubenswrapper[4897]: I1121 14:32:49.792894 4897 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/4a683589-a940-405b-afe1-1838f34dac9a-config\") on node \"crc\" DevicePath \"\"" Nov 21 14:32:50 crc kubenswrapper[4897]: I1121 14:32:50.104497 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-mrctr" Nov 21 14:32:50 crc kubenswrapper[4897]: I1121 14:32:50.105361 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-mrctr" event={"ID":"4a683589-a940-405b-afe1-1838f34dac9a","Type":"ContainerDied","Data":"350ee59368631067e4b119b860a3b0e5f8fb23b051990ae4884466a22966dd83"} Nov 21 14:32:50 crc kubenswrapper[4897]: I1121 14:32:50.105392 4897 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="350ee59368631067e4b119b860a3b0e5f8fb23b051990ae4884466a22966dd83" Nov 21 14:32:50 crc kubenswrapper[4897]: W1121 14:32:50.298269 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod23323788_338f_42c8_817a_e18aa170417d.slice/crio-8828cb2b28c036cbe55b9a0d794fd80ff573f8f4cfe72102f33582a5b3997b68 WatchSource:0}: Error finding container 8828cb2b28c036cbe55b9a0d794fd80ff573f8f4cfe72102f33582a5b3997b68: Status 404 returned error can't find the container with id 8828cb2b28c036cbe55b9a0d794fd80ff573f8f4cfe72102f33582a5b3997b68 Nov 21 14:32:50 crc kubenswrapper[4897]: I1121 14:32:50.305626 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 21 14:32:50 crc kubenswrapper[4897]: I1121 14:32:50.445220 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6bb4fc677f-zlcqd"] Nov 21 14:32:50 crc kubenswrapper[4897]: E1121 14:32:50.447925 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a683589-a940-405b-afe1-1838f34dac9a" containerName="neutron-db-sync" Nov 21 14:32:50 crc kubenswrapper[4897]: I1121 14:32:50.447947 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a683589-a940-405b-afe1-1838f34dac9a" containerName="neutron-db-sync" Nov 21 14:32:50 crc kubenswrapper[4897]: I1121 14:32:50.448190 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a683589-a940-405b-afe1-1838f34dac9a" containerName="neutron-db-sync" Nov 21 14:32:50 crc kubenswrapper[4897]: I1121 14:32:50.450976 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bb4fc677f-zlcqd" Nov 21 14:32:50 crc kubenswrapper[4897]: I1121 14:32:50.470118 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bb4fc677f-zlcqd"] Nov 21 14:32:50 crc kubenswrapper[4897]: I1121 14:32:50.564406 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-7446d77dd8-rg2bn"] Nov 21 14:32:50 crc kubenswrapper[4897]: I1121 14:32:50.567267 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7446d77dd8-rg2bn" Nov 21 14:32:50 crc kubenswrapper[4897]: I1121 14:32:50.572064 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Nov 21 14:32:50 crc kubenswrapper[4897]: I1121 14:32:50.572261 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Nov 21 14:32:50 crc kubenswrapper[4897]: I1121 14:32:50.572363 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Nov 21 14:32:50 crc kubenswrapper[4897]: I1121 14:32:50.573114 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-lddw4" Nov 21 14:32:50 crc kubenswrapper[4897]: I1121 14:32:50.583036 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7446d77dd8-rg2bn"] Nov 21 14:32:50 crc kubenswrapper[4897]: I1121 14:32:50.618751 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/71cdff91-5167-4b4a-9958-c45c77c2f8e1-config\") pod \"dnsmasq-dns-6bb4fc677f-zlcqd\" (UID: \"71cdff91-5167-4b4a-9958-c45c77c2f8e1\") " pod="openstack/dnsmasq-dns-6bb4fc677f-zlcqd" Nov 21 14:32:50 crc kubenswrapper[4897]: I1121 14:32:50.618814 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/71cdff91-5167-4b4a-9958-c45c77c2f8e1-dns-swift-storage-0\") pod \"dnsmasq-dns-6bb4fc677f-zlcqd\" (UID: \"71cdff91-5167-4b4a-9958-c45c77c2f8e1\") " pod="openstack/dnsmasq-dns-6bb4fc677f-zlcqd" Nov 21 14:32:50 crc kubenswrapper[4897]: I1121 14:32:50.618837 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/71cdff91-5167-4b4a-9958-c45c77c2f8e1-ovsdbserver-nb\") pod \"dnsmasq-dns-6bb4fc677f-zlcqd\" (UID: \"71cdff91-5167-4b4a-9958-c45c77c2f8e1\") " pod="openstack/dnsmasq-dns-6bb4fc677f-zlcqd" Nov 21 14:32:50 crc kubenswrapper[4897]: I1121 14:32:50.618934 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/71cdff91-5167-4b4a-9958-c45c77c2f8e1-dns-svc\") pod \"dnsmasq-dns-6bb4fc677f-zlcqd\" (UID: \"71cdff91-5167-4b4a-9958-c45c77c2f8e1\") " pod="openstack/dnsmasq-dns-6bb4fc677f-zlcqd" Nov 21 14:32:50 crc kubenswrapper[4897]: I1121 14:32:50.618973 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/71cdff91-5167-4b4a-9958-c45c77c2f8e1-ovsdbserver-sb\") pod \"dnsmasq-dns-6bb4fc677f-zlcqd\" (UID: \"71cdff91-5167-4b4a-9958-c45c77c2f8e1\") " pod="openstack/dnsmasq-dns-6bb4fc677f-zlcqd" Nov 21 14:32:50 crc kubenswrapper[4897]: I1121 14:32:50.619166 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zldpw\" (UniqueName: \"kubernetes.io/projected/71cdff91-5167-4b4a-9958-c45c77c2f8e1-kube-api-access-zldpw\") pod \"dnsmasq-dns-6bb4fc677f-zlcqd\" (UID: \"71cdff91-5167-4b4a-9958-c45c77c2f8e1\") " pod="openstack/dnsmasq-dns-6bb4fc677f-zlcqd" Nov 21 14:32:50 crc kubenswrapper[4897]: I1121 14:32:50.721727 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zldpw\" (UniqueName: \"kubernetes.io/projected/71cdff91-5167-4b4a-9958-c45c77c2f8e1-kube-api-access-zldpw\") pod \"dnsmasq-dns-6bb4fc677f-zlcqd\" (UID: \"71cdff91-5167-4b4a-9958-c45c77c2f8e1\") " pod="openstack/dnsmasq-dns-6bb4fc677f-zlcqd" Nov 21 14:32:50 crc kubenswrapper[4897]: I1121 14:32:50.721788 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/bac76d5f-def0-4fe8-babf-2c1a171f04dc-config\") pod \"neutron-7446d77dd8-rg2bn\" (UID: \"bac76d5f-def0-4fe8-babf-2c1a171f04dc\") " pod="openstack/neutron-7446d77dd8-rg2bn" Nov 21 14:32:50 crc kubenswrapper[4897]: I1121 14:32:50.721824 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/71cdff91-5167-4b4a-9958-c45c77c2f8e1-config\") pod \"dnsmasq-dns-6bb4fc677f-zlcqd\" (UID: \"71cdff91-5167-4b4a-9958-c45c77c2f8e1\") " pod="openstack/dnsmasq-dns-6bb4fc677f-zlcqd" Nov 21 14:32:50 crc kubenswrapper[4897]: I1121 14:32:50.721884 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/71cdff91-5167-4b4a-9958-c45c77c2f8e1-dns-swift-storage-0\") pod \"dnsmasq-dns-6bb4fc677f-zlcqd\" (UID: \"71cdff91-5167-4b4a-9958-c45c77c2f8e1\") " pod="openstack/dnsmasq-dns-6bb4fc677f-zlcqd" Nov 21 14:32:50 crc kubenswrapper[4897]: I1121 14:32:50.721900 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/71cdff91-5167-4b4a-9958-c45c77c2f8e1-ovsdbserver-nb\") pod \"dnsmasq-dns-6bb4fc677f-zlcqd\" (UID: \"71cdff91-5167-4b4a-9958-c45c77c2f8e1\") " pod="openstack/dnsmasq-dns-6bb4fc677f-zlcqd" Nov 21 14:32:50 crc kubenswrapper[4897]: I1121 14:32:50.721917 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/bac76d5f-def0-4fe8-babf-2c1a171f04dc-httpd-config\") pod \"neutron-7446d77dd8-rg2bn\" (UID: \"bac76d5f-def0-4fe8-babf-2c1a171f04dc\") " pod="openstack/neutron-7446d77dd8-rg2bn" Nov 21 14:32:50 crc kubenswrapper[4897]: I1121 14:32:50.721963 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/bac76d5f-def0-4fe8-babf-2c1a171f04dc-ovndb-tls-certs\") pod \"neutron-7446d77dd8-rg2bn\" (UID: \"bac76d5f-def0-4fe8-babf-2c1a171f04dc\") " pod="openstack/neutron-7446d77dd8-rg2bn" Nov 21 14:32:50 crc kubenswrapper[4897]: I1121 14:32:50.721987 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9ttr9\" (UniqueName: \"kubernetes.io/projected/bac76d5f-def0-4fe8-babf-2c1a171f04dc-kube-api-access-9ttr9\") pod \"neutron-7446d77dd8-rg2bn\" (UID: \"bac76d5f-def0-4fe8-babf-2c1a171f04dc\") " pod="openstack/neutron-7446d77dd8-rg2bn" Nov 21 14:32:50 crc kubenswrapper[4897]: I1121 14:32:50.722029 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/71cdff91-5167-4b4a-9958-c45c77c2f8e1-dns-svc\") pod \"dnsmasq-dns-6bb4fc677f-zlcqd\" (UID: \"71cdff91-5167-4b4a-9958-c45c77c2f8e1\") " pod="openstack/dnsmasq-dns-6bb4fc677f-zlcqd" Nov 21 14:32:50 crc kubenswrapper[4897]: I1121 14:32:50.722055 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/71cdff91-5167-4b4a-9958-c45c77c2f8e1-ovsdbserver-sb\") pod \"dnsmasq-dns-6bb4fc677f-zlcqd\" (UID: \"71cdff91-5167-4b4a-9958-c45c77c2f8e1\") " pod="openstack/dnsmasq-dns-6bb4fc677f-zlcqd" Nov 21 14:32:50 crc kubenswrapper[4897]: I1121 14:32:50.722108 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bac76d5f-def0-4fe8-babf-2c1a171f04dc-combined-ca-bundle\") pod \"neutron-7446d77dd8-rg2bn\" (UID: \"bac76d5f-def0-4fe8-babf-2c1a171f04dc\") " pod="openstack/neutron-7446d77dd8-rg2bn" Nov 21 14:32:50 crc kubenswrapper[4897]: I1121 14:32:50.723030 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/71cdff91-5167-4b4a-9958-c45c77c2f8e1-dns-svc\") pod \"dnsmasq-dns-6bb4fc677f-zlcqd\" (UID: \"71cdff91-5167-4b4a-9958-c45c77c2f8e1\") " pod="openstack/dnsmasq-dns-6bb4fc677f-zlcqd" Nov 21 14:32:50 crc kubenswrapper[4897]: I1121 14:32:50.723076 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/71cdff91-5167-4b4a-9958-c45c77c2f8e1-dns-swift-storage-0\") pod \"dnsmasq-dns-6bb4fc677f-zlcqd\" (UID: \"71cdff91-5167-4b4a-9958-c45c77c2f8e1\") " pod="openstack/dnsmasq-dns-6bb4fc677f-zlcqd" Nov 21 14:32:50 crc kubenswrapper[4897]: I1121 14:32:50.723089 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/71cdff91-5167-4b4a-9958-c45c77c2f8e1-config\") pod \"dnsmasq-dns-6bb4fc677f-zlcqd\" (UID: \"71cdff91-5167-4b4a-9958-c45c77c2f8e1\") " pod="openstack/dnsmasq-dns-6bb4fc677f-zlcqd" Nov 21 14:32:50 crc kubenswrapper[4897]: I1121 14:32:50.723455 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/71cdff91-5167-4b4a-9958-c45c77c2f8e1-ovsdbserver-sb\") pod \"dnsmasq-dns-6bb4fc677f-zlcqd\" (UID: \"71cdff91-5167-4b4a-9958-c45c77c2f8e1\") " pod="openstack/dnsmasq-dns-6bb4fc677f-zlcqd" Nov 21 14:32:50 crc kubenswrapper[4897]: I1121 14:32:50.723823 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/71cdff91-5167-4b4a-9958-c45c77c2f8e1-ovsdbserver-nb\") pod \"dnsmasq-dns-6bb4fc677f-zlcqd\" (UID: \"71cdff91-5167-4b4a-9958-c45c77c2f8e1\") " pod="openstack/dnsmasq-dns-6bb4fc677f-zlcqd" Nov 21 14:32:50 crc kubenswrapper[4897]: I1121 14:32:50.747950 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zldpw\" (UniqueName: \"kubernetes.io/projected/71cdff91-5167-4b4a-9958-c45c77c2f8e1-kube-api-access-zldpw\") pod \"dnsmasq-dns-6bb4fc677f-zlcqd\" (UID: \"71cdff91-5167-4b4a-9958-c45c77c2f8e1\") " pod="openstack/dnsmasq-dns-6bb4fc677f-zlcqd" Nov 21 14:32:50 crc kubenswrapper[4897]: I1121 14:32:50.823993 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/bac76d5f-def0-4fe8-babf-2c1a171f04dc-config\") pod \"neutron-7446d77dd8-rg2bn\" (UID: \"bac76d5f-def0-4fe8-babf-2c1a171f04dc\") " pod="openstack/neutron-7446d77dd8-rg2bn" Nov 21 14:32:50 crc kubenswrapper[4897]: I1121 14:32:50.824359 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/bac76d5f-def0-4fe8-babf-2c1a171f04dc-httpd-config\") pod \"neutron-7446d77dd8-rg2bn\" (UID: \"bac76d5f-def0-4fe8-babf-2c1a171f04dc\") " pod="openstack/neutron-7446d77dd8-rg2bn" Nov 21 14:32:50 crc kubenswrapper[4897]: I1121 14:32:50.824536 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/bac76d5f-def0-4fe8-babf-2c1a171f04dc-ovndb-tls-certs\") pod \"neutron-7446d77dd8-rg2bn\" (UID: \"bac76d5f-def0-4fe8-babf-2c1a171f04dc\") " pod="openstack/neutron-7446d77dd8-rg2bn" Nov 21 14:32:50 crc kubenswrapper[4897]: I1121 14:32:50.824642 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9ttr9\" (UniqueName: \"kubernetes.io/projected/bac76d5f-def0-4fe8-babf-2c1a171f04dc-kube-api-access-9ttr9\") pod \"neutron-7446d77dd8-rg2bn\" (UID: \"bac76d5f-def0-4fe8-babf-2c1a171f04dc\") " pod="openstack/neutron-7446d77dd8-rg2bn" Nov 21 14:32:50 crc kubenswrapper[4897]: I1121 14:32:50.824848 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bac76d5f-def0-4fe8-babf-2c1a171f04dc-combined-ca-bundle\") pod \"neutron-7446d77dd8-rg2bn\" (UID: \"bac76d5f-def0-4fe8-babf-2c1a171f04dc\") " pod="openstack/neutron-7446d77dd8-rg2bn" Nov 21 14:32:50 crc kubenswrapper[4897]: I1121 14:32:50.828374 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/bac76d5f-def0-4fe8-babf-2c1a171f04dc-httpd-config\") pod \"neutron-7446d77dd8-rg2bn\" (UID: \"bac76d5f-def0-4fe8-babf-2c1a171f04dc\") " pod="openstack/neutron-7446d77dd8-rg2bn" Nov 21 14:32:50 crc kubenswrapper[4897]: I1121 14:32:50.828695 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/bac76d5f-def0-4fe8-babf-2c1a171f04dc-config\") pod \"neutron-7446d77dd8-rg2bn\" (UID: \"bac76d5f-def0-4fe8-babf-2c1a171f04dc\") " pod="openstack/neutron-7446d77dd8-rg2bn" Nov 21 14:32:50 crc kubenswrapper[4897]: I1121 14:32:50.829734 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Nov 21 14:32:50 crc kubenswrapper[4897]: I1121 14:32:50.830157 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/bac76d5f-def0-4fe8-babf-2c1a171f04dc-ovndb-tls-certs\") pod \"neutron-7446d77dd8-rg2bn\" (UID: \"bac76d5f-def0-4fe8-babf-2c1a171f04dc\") " pod="openstack/neutron-7446d77dd8-rg2bn" Nov 21 14:32:50 crc kubenswrapper[4897]: I1121 14:32:50.839734 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bb4fc677f-zlcqd" Nov 21 14:32:50 crc kubenswrapper[4897]: I1121 14:32:50.841864 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9ttr9\" (UniqueName: \"kubernetes.io/projected/bac76d5f-def0-4fe8-babf-2c1a171f04dc-kube-api-access-9ttr9\") pod \"neutron-7446d77dd8-rg2bn\" (UID: \"bac76d5f-def0-4fe8-babf-2c1a171f04dc\") " pod="openstack/neutron-7446d77dd8-rg2bn" Nov 21 14:32:50 crc kubenswrapper[4897]: I1121 14:32:50.842104 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bac76d5f-def0-4fe8-babf-2c1a171f04dc-combined-ca-bundle\") pod \"neutron-7446d77dd8-rg2bn\" (UID: \"bac76d5f-def0-4fe8-babf-2c1a171f04dc\") " pod="openstack/neutron-7446d77dd8-rg2bn" Nov 21 14:32:50 crc kubenswrapper[4897]: I1121 14:32:50.890039 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7446d77dd8-rg2bn" Nov 21 14:32:50 crc kubenswrapper[4897]: I1121 14:32:50.898904 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 21 14:32:51 crc kubenswrapper[4897]: I1121 14:32:51.119893 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="79802b6b-53b5-4568-b79b-83da7dbbf123" containerName="cinder-scheduler" containerID="cri-o://cebc20d99433055d6fe1f45416a2f0a69cbf7883d422fa7d34130d3f6586593a" gracePeriod=30 Nov 21 14:32:51 crc kubenswrapper[4897]: I1121 14:32:51.120436 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"23323788-338f-42c8-817a-e18aa170417d","Type":"ContainerStarted","Data":"8828cb2b28c036cbe55b9a0d794fd80ff573f8f4cfe72102f33582a5b3997b68"} Nov 21 14:32:51 crc kubenswrapper[4897]: I1121 14:32:51.120682 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="79802b6b-53b5-4568-b79b-83da7dbbf123" containerName="probe" containerID="cri-o://fb1780b07f00fff91feafdcf7e5944c12e5d238b232280791d8dc951cb74c7c3" gracePeriod=30 Nov 21 14:32:51 crc kubenswrapper[4897]: I1121 14:32:51.438221 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bb4fc677f-zlcqd"] Nov 21 14:32:51 crc kubenswrapper[4897]: W1121 14:32:51.449582 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod71cdff91_5167_4b4a_9958_c45c77c2f8e1.slice/crio-9f3cf386c81836b9b08a9759b7c9b3cb037fcb498538cdfb9be621765190bff3 WatchSource:0}: Error finding container 9f3cf386c81836b9b08a9759b7c9b3cb037fcb498538cdfb9be621765190bff3: Status 404 returned error can't find the container with id 9f3cf386c81836b9b08a9759b7c9b3cb037fcb498538cdfb9be621765190bff3 Nov 21 14:32:51 crc kubenswrapper[4897]: I1121 14:32:51.702168 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7446d77dd8-rg2bn"] Nov 21 14:32:52 crc kubenswrapper[4897]: I1121 14:32:52.156328 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb4fc677f-zlcqd" event={"ID":"71cdff91-5167-4b4a-9958-c45c77c2f8e1","Type":"ContainerStarted","Data":"9f3cf386c81836b9b08a9759b7c9b3cb037fcb498538cdfb9be621765190bff3"} Nov 21 14:32:52 crc kubenswrapper[4897]: I1121 14:32:52.163416 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7446d77dd8-rg2bn" event={"ID":"bac76d5f-def0-4fe8-babf-2c1a171f04dc","Type":"ContainerStarted","Data":"d8b09365925618b0cb70df4372567c5ebac5e2690ccbb21641ffeefc36d8e164"} Nov 21 14:32:52 crc kubenswrapper[4897]: I1121 14:32:52.991851 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-f46d49697-qwn8n"] Nov 21 14:32:52 crc kubenswrapper[4897]: I1121 14:32:52.994118 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-f46d49697-qwn8n" Nov 21 14:32:53 crc kubenswrapper[4897]: I1121 14:32:53.006589 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Nov 21 14:32:53 crc kubenswrapper[4897]: I1121 14:32:53.007495 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Nov 21 14:32:53 crc kubenswrapper[4897]: I1121 14:32:53.025246 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-f46d49697-qwn8n"] Nov 21 14:32:53 crc kubenswrapper[4897]: I1121 14:32:53.095549 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8vgk6\" (UniqueName: \"kubernetes.io/projected/586453ed-44e4-4b53-b205-d4d325a8a702-kube-api-access-8vgk6\") pod \"neutron-f46d49697-qwn8n\" (UID: \"586453ed-44e4-4b53-b205-d4d325a8a702\") " pod="openstack/neutron-f46d49697-qwn8n" Nov 21 14:32:53 crc kubenswrapper[4897]: I1121 14:32:53.095633 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/586453ed-44e4-4b53-b205-d4d325a8a702-internal-tls-certs\") pod \"neutron-f46d49697-qwn8n\" (UID: \"586453ed-44e4-4b53-b205-d4d325a8a702\") " pod="openstack/neutron-f46d49697-qwn8n" Nov 21 14:32:53 crc kubenswrapper[4897]: I1121 14:32:53.095848 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/586453ed-44e4-4b53-b205-d4d325a8a702-combined-ca-bundle\") pod \"neutron-f46d49697-qwn8n\" (UID: \"586453ed-44e4-4b53-b205-d4d325a8a702\") " pod="openstack/neutron-f46d49697-qwn8n" Nov 21 14:32:53 crc kubenswrapper[4897]: I1121 14:32:53.095950 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/586453ed-44e4-4b53-b205-d4d325a8a702-ovndb-tls-certs\") pod \"neutron-f46d49697-qwn8n\" (UID: \"586453ed-44e4-4b53-b205-d4d325a8a702\") " pod="openstack/neutron-f46d49697-qwn8n" Nov 21 14:32:53 crc kubenswrapper[4897]: I1121 14:32:53.096104 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/586453ed-44e4-4b53-b205-d4d325a8a702-httpd-config\") pod \"neutron-f46d49697-qwn8n\" (UID: \"586453ed-44e4-4b53-b205-d4d325a8a702\") " pod="openstack/neutron-f46d49697-qwn8n" Nov 21 14:32:53 crc kubenswrapper[4897]: I1121 14:32:53.096167 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/586453ed-44e4-4b53-b205-d4d325a8a702-public-tls-certs\") pod \"neutron-f46d49697-qwn8n\" (UID: \"586453ed-44e4-4b53-b205-d4d325a8a702\") " pod="openstack/neutron-f46d49697-qwn8n" Nov 21 14:32:53 crc kubenswrapper[4897]: I1121 14:32:53.096235 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/586453ed-44e4-4b53-b205-d4d325a8a702-config\") pod \"neutron-f46d49697-qwn8n\" (UID: \"586453ed-44e4-4b53-b205-d4d325a8a702\") " pod="openstack/neutron-f46d49697-qwn8n" Nov 21 14:32:53 crc kubenswrapper[4897]: I1121 14:32:53.187859 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb4fc677f-zlcqd" event={"ID":"71cdff91-5167-4b4a-9958-c45c77c2f8e1","Type":"ContainerStarted","Data":"430d5775e2aa468d158479b0dee949ec7a55746b5d1bc8ab8b0afd85351b11d6"} Nov 21 14:32:53 crc kubenswrapper[4897]: I1121 14:32:53.189823 4897 generic.go:334] "Generic (PLEG): container finished" podID="79802b6b-53b5-4568-b79b-83da7dbbf123" containerID="fb1780b07f00fff91feafdcf7e5944c12e5d238b232280791d8dc951cb74c7c3" exitCode=0 Nov 21 14:32:53 crc kubenswrapper[4897]: I1121 14:32:53.189886 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"79802b6b-53b5-4568-b79b-83da7dbbf123","Type":"ContainerDied","Data":"fb1780b07f00fff91feafdcf7e5944c12e5d238b232280791d8dc951cb74c7c3"} Nov 21 14:32:53 crc kubenswrapper[4897]: I1121 14:32:53.191305 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7446d77dd8-rg2bn" event={"ID":"bac76d5f-def0-4fe8-babf-2c1a171f04dc","Type":"ContainerStarted","Data":"e8721fc63494ab662dbf3cfe15e933c7d2469a6932ecbc6c02e513fec6121a44"} Nov 21 14:32:53 crc kubenswrapper[4897]: I1121 14:32:53.198519 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/586453ed-44e4-4b53-b205-d4d325a8a702-ovndb-tls-certs\") pod \"neutron-f46d49697-qwn8n\" (UID: \"586453ed-44e4-4b53-b205-d4d325a8a702\") " pod="openstack/neutron-f46d49697-qwn8n" Nov 21 14:32:53 crc kubenswrapper[4897]: I1121 14:32:53.198617 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/586453ed-44e4-4b53-b205-d4d325a8a702-httpd-config\") pod \"neutron-f46d49697-qwn8n\" (UID: \"586453ed-44e4-4b53-b205-d4d325a8a702\") " pod="openstack/neutron-f46d49697-qwn8n" Nov 21 14:32:53 crc kubenswrapper[4897]: I1121 14:32:53.198642 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/586453ed-44e4-4b53-b205-d4d325a8a702-public-tls-certs\") pod \"neutron-f46d49697-qwn8n\" (UID: \"586453ed-44e4-4b53-b205-d4d325a8a702\") " pod="openstack/neutron-f46d49697-qwn8n" Nov 21 14:32:53 crc kubenswrapper[4897]: I1121 14:32:53.198668 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/586453ed-44e4-4b53-b205-d4d325a8a702-config\") pod \"neutron-f46d49697-qwn8n\" (UID: \"586453ed-44e4-4b53-b205-d4d325a8a702\") " pod="openstack/neutron-f46d49697-qwn8n" Nov 21 14:32:53 crc kubenswrapper[4897]: I1121 14:32:53.198749 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8vgk6\" (UniqueName: \"kubernetes.io/projected/586453ed-44e4-4b53-b205-d4d325a8a702-kube-api-access-8vgk6\") pod \"neutron-f46d49697-qwn8n\" (UID: \"586453ed-44e4-4b53-b205-d4d325a8a702\") " pod="openstack/neutron-f46d49697-qwn8n" Nov 21 14:32:53 crc kubenswrapper[4897]: I1121 14:32:53.198797 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/586453ed-44e4-4b53-b205-d4d325a8a702-internal-tls-certs\") pod \"neutron-f46d49697-qwn8n\" (UID: \"586453ed-44e4-4b53-b205-d4d325a8a702\") " pod="openstack/neutron-f46d49697-qwn8n" Nov 21 14:32:53 crc kubenswrapper[4897]: I1121 14:32:53.198857 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/586453ed-44e4-4b53-b205-d4d325a8a702-combined-ca-bundle\") pod \"neutron-f46d49697-qwn8n\" (UID: \"586453ed-44e4-4b53-b205-d4d325a8a702\") " pod="openstack/neutron-f46d49697-qwn8n" Nov 21 14:32:53 crc kubenswrapper[4897]: I1121 14:32:53.211001 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/586453ed-44e4-4b53-b205-d4d325a8a702-public-tls-certs\") pod \"neutron-f46d49697-qwn8n\" (UID: \"586453ed-44e4-4b53-b205-d4d325a8a702\") " pod="openstack/neutron-f46d49697-qwn8n" Nov 21 14:32:53 crc kubenswrapper[4897]: I1121 14:32:53.215129 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/586453ed-44e4-4b53-b205-d4d325a8a702-httpd-config\") pod \"neutron-f46d49697-qwn8n\" (UID: \"586453ed-44e4-4b53-b205-d4d325a8a702\") " pod="openstack/neutron-f46d49697-qwn8n" Nov 21 14:32:53 crc kubenswrapper[4897]: I1121 14:32:53.217096 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/586453ed-44e4-4b53-b205-d4d325a8a702-combined-ca-bundle\") pod \"neutron-f46d49697-qwn8n\" (UID: \"586453ed-44e4-4b53-b205-d4d325a8a702\") " pod="openstack/neutron-f46d49697-qwn8n" Nov 21 14:32:53 crc kubenswrapper[4897]: I1121 14:32:53.217218 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/586453ed-44e4-4b53-b205-d4d325a8a702-ovndb-tls-certs\") pod \"neutron-f46d49697-qwn8n\" (UID: \"586453ed-44e4-4b53-b205-d4d325a8a702\") " pod="openstack/neutron-f46d49697-qwn8n" Nov 21 14:32:53 crc kubenswrapper[4897]: I1121 14:32:53.226675 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/586453ed-44e4-4b53-b205-d4d325a8a702-config\") pod \"neutron-f46d49697-qwn8n\" (UID: \"586453ed-44e4-4b53-b205-d4d325a8a702\") " pod="openstack/neutron-f46d49697-qwn8n" Nov 21 14:32:53 crc kubenswrapper[4897]: I1121 14:32:53.227253 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/586453ed-44e4-4b53-b205-d4d325a8a702-internal-tls-certs\") pod \"neutron-f46d49697-qwn8n\" (UID: \"586453ed-44e4-4b53-b205-d4d325a8a702\") " pod="openstack/neutron-f46d49697-qwn8n" Nov 21 14:32:53 crc kubenswrapper[4897]: I1121 14:32:53.236638 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8vgk6\" (UniqueName: \"kubernetes.io/projected/586453ed-44e4-4b53-b205-d4d325a8a702-kube-api-access-8vgk6\") pod \"neutron-f46d49697-qwn8n\" (UID: \"586453ed-44e4-4b53-b205-d4d325a8a702\") " pod="openstack/neutron-f46d49697-qwn8n" Nov 21 14:32:53 crc kubenswrapper[4897]: I1121 14:32:53.333118 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-f46d49697-qwn8n" Nov 21 14:32:54 crc kubenswrapper[4897]: I1121 14:32:54.337697 4897 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/cinder-api-0" podUID="8a67c019-b256-4890-8fb0-1dfa3a88b7b1" containerName="cinder-api" probeResult="failure" output="Get \"https://10.217.0.201:8776/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 21 14:32:54 crc kubenswrapper[4897]: I1121 14:32:54.496145 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-f46d49697-qwn8n"] Nov 21 14:32:54 crc kubenswrapper[4897]: W1121 14:32:54.874904 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod586453ed_44e4_4b53_b205_d4d325a8a702.slice/crio-73d1daaae336791a3c4d4652bd936b46a55160d864dbab9a8d8dfeaa7687f842 WatchSource:0}: Error finding container 73d1daaae336791a3c4d4652bd936b46a55160d864dbab9a8d8dfeaa7687f842: Status 404 returned error can't find the container with id 73d1daaae336791a3c4d4652bd936b46a55160d864dbab9a8d8dfeaa7687f842 Nov 21 14:32:55 crc kubenswrapper[4897]: I1121 14:32:55.251166 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-f46d49697-qwn8n" event={"ID":"586453ed-44e4-4b53-b205-d4d325a8a702","Type":"ContainerStarted","Data":"73d1daaae336791a3c4d4652bd936b46a55160d864dbab9a8d8dfeaa7687f842"} Nov 21 14:32:55 crc kubenswrapper[4897]: I1121 14:32:55.257544 4897 generic.go:334] "Generic (PLEG): container finished" podID="71cdff91-5167-4b4a-9958-c45c77c2f8e1" containerID="430d5775e2aa468d158479b0dee949ec7a55746b5d1bc8ab8b0afd85351b11d6" exitCode=0 Nov 21 14:32:55 crc kubenswrapper[4897]: I1121 14:32:55.258602 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb4fc677f-zlcqd" event={"ID":"71cdff91-5167-4b4a-9958-c45c77c2f8e1","Type":"ContainerDied","Data":"430d5775e2aa468d158479b0dee949ec7a55746b5d1bc8ab8b0afd85351b11d6"} Nov 21 14:32:55 crc kubenswrapper[4897]: I1121 14:32:55.266299 4897 generic.go:334] "Generic (PLEG): container finished" podID="79802b6b-53b5-4568-b79b-83da7dbbf123" containerID="cebc20d99433055d6fe1f45416a2f0a69cbf7883d422fa7d34130d3f6586593a" exitCode=0 Nov 21 14:32:55 crc kubenswrapper[4897]: I1121 14:32:55.266352 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"79802b6b-53b5-4568-b79b-83da7dbbf123","Type":"ContainerDied","Data":"cebc20d99433055d6fe1f45416a2f0a69cbf7883d422fa7d34130d3f6586593a"} Nov 21 14:32:55 crc kubenswrapper[4897]: I1121 14:32:55.332770 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="8a67c019-b256-4890-8fb0-1dfa3a88b7b1" containerName="cinder-api" probeResult="failure" output="Get \"https://10.217.0.201:8776/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 21 14:32:55 crc kubenswrapper[4897]: I1121 14:32:55.794066 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-75bd44869-p288j"] Nov 21 14:32:55 crc kubenswrapper[4897]: I1121 14:32:55.796012 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-75bd44869-p288j" Nov 21 14:32:55 crc kubenswrapper[4897]: I1121 14:32:55.800847 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-tkmk9" Nov 21 14:32:55 crc kubenswrapper[4897]: I1121 14:32:55.801102 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-engine-config-data" Nov 21 14:32:55 crc kubenswrapper[4897]: I1121 14:32:55.801260 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Nov 21 14:32:55 crc kubenswrapper[4897]: I1121 14:32:55.812779 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-75bd44869-p288j"] Nov 21 14:32:55 crc kubenswrapper[4897]: I1121 14:32:55.898086 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-56b88658f-4ndrt"] Nov 21 14:32:55 crc kubenswrapper[4897]: I1121 14:32:55.900407 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-56b88658f-4ndrt" Nov 21 14:32:55 crc kubenswrapper[4897]: I1121 14:32:55.921089 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-api-config-data" Nov 21 14:32:55 crc kubenswrapper[4897]: I1121 14:32:55.934929 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-55749dccc7-lbx4x"] Nov 21 14:32:55 crc kubenswrapper[4897]: I1121 14:32:55.936834 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-55749dccc7-lbx4x" Nov 21 14:32:55 crc kubenswrapper[4897]: I1121 14:32:55.939098 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-cfnapi-config-data" Nov 21 14:32:55 crc kubenswrapper[4897]: I1121 14:32:55.964569 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-56b88658f-4ndrt"] Nov 21 14:32:55 crc kubenswrapper[4897]: I1121 14:32:55.978987 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b546c9e-dba8-4806-a2ed-8023271b448a-config-data\") pod \"heat-engine-75bd44869-p288j\" (UID: \"7b546c9e-dba8-4806-a2ed-8023271b448a\") " pod="openstack/heat-engine-75bd44869-p288j" Nov 21 14:32:55 crc kubenswrapper[4897]: I1121 14:32:55.979057 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7b546c9e-dba8-4806-a2ed-8023271b448a-config-data-custom\") pod \"heat-engine-75bd44869-p288j\" (UID: \"7b546c9e-dba8-4806-a2ed-8023271b448a\") " pod="openstack/heat-engine-75bd44869-p288j" Nov 21 14:32:55 crc kubenswrapper[4897]: I1121 14:32:55.979286 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b546c9e-dba8-4806-a2ed-8023271b448a-combined-ca-bundle\") pod \"heat-engine-75bd44869-p288j\" (UID: \"7b546c9e-dba8-4806-a2ed-8023271b448a\") " pod="openstack/heat-engine-75bd44869-p288j" Nov 21 14:32:55 crc kubenswrapper[4897]: I1121 14:32:55.979364 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hc55h\" (UniqueName: \"kubernetes.io/projected/7b546c9e-dba8-4806-a2ed-8023271b448a-kube-api-access-hc55h\") pod \"heat-engine-75bd44869-p288j\" (UID: \"7b546c9e-dba8-4806-a2ed-8023271b448a\") " pod="openstack/heat-engine-75bd44869-p288j" Nov 21 14:32:55 crc kubenswrapper[4897]: I1121 14:32:55.995582 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-55749dccc7-lbx4x"] Nov 21 14:32:56 crc kubenswrapper[4897]: I1121 14:32:56.060613 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bb4fc677f-zlcqd"] Nov 21 14:32:56 crc kubenswrapper[4897]: I1121 14:32:56.082959 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4w59b\" (UniqueName: \"kubernetes.io/projected/9142a1dd-8c8e-486c-b1f3-d2f1c3fd6728-kube-api-access-4w59b\") pod \"heat-api-56b88658f-4ndrt\" (UID: \"9142a1dd-8c8e-486c-b1f3-d2f1c3fd6728\") " pod="openstack/heat-api-56b88658f-4ndrt" Nov 21 14:32:56 crc kubenswrapper[4897]: I1121 14:32:56.083019 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b546c9e-dba8-4806-a2ed-8023271b448a-combined-ca-bundle\") pod \"heat-engine-75bd44869-p288j\" (UID: \"7b546c9e-dba8-4806-a2ed-8023271b448a\") " pod="openstack/heat-engine-75bd44869-p288j" Nov 21 14:32:56 crc kubenswrapper[4897]: I1121 14:32:56.083052 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e0a52a6-14c9-4529-b2a8-c26b739b7986-combined-ca-bundle\") pod \"heat-cfnapi-55749dccc7-lbx4x\" (UID: \"2e0a52a6-14c9-4529-b2a8-c26b739b7986\") " pod="openstack/heat-cfnapi-55749dccc7-lbx4x" Nov 21 14:32:56 crc kubenswrapper[4897]: I1121 14:32:56.083114 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hc55h\" (UniqueName: \"kubernetes.io/projected/7b546c9e-dba8-4806-a2ed-8023271b448a-kube-api-access-hc55h\") pod \"heat-engine-75bd44869-p288j\" (UID: \"7b546c9e-dba8-4806-a2ed-8023271b448a\") " pod="openstack/heat-engine-75bd44869-p288j" Nov 21 14:32:56 crc kubenswrapper[4897]: I1121 14:32:56.083150 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9142a1dd-8c8e-486c-b1f3-d2f1c3fd6728-config-data-custom\") pod \"heat-api-56b88658f-4ndrt\" (UID: \"9142a1dd-8c8e-486c-b1f3-d2f1c3fd6728\") " pod="openstack/heat-api-56b88658f-4ndrt" Nov 21 14:32:56 crc kubenswrapper[4897]: I1121 14:32:56.083191 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9142a1dd-8c8e-486c-b1f3-d2f1c3fd6728-config-data\") pod \"heat-api-56b88658f-4ndrt\" (UID: \"9142a1dd-8c8e-486c-b1f3-d2f1c3fd6728\") " pod="openstack/heat-api-56b88658f-4ndrt" Nov 21 14:32:56 crc kubenswrapper[4897]: I1121 14:32:56.083241 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2e0a52a6-14c9-4529-b2a8-c26b739b7986-config-data-custom\") pod \"heat-cfnapi-55749dccc7-lbx4x\" (UID: \"2e0a52a6-14c9-4529-b2a8-c26b739b7986\") " pod="openstack/heat-cfnapi-55749dccc7-lbx4x" Nov 21 14:32:56 crc kubenswrapper[4897]: I1121 14:32:56.083278 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b546c9e-dba8-4806-a2ed-8023271b448a-config-data\") pod \"heat-engine-75bd44869-p288j\" (UID: \"7b546c9e-dba8-4806-a2ed-8023271b448a\") " pod="openstack/heat-engine-75bd44869-p288j" Nov 21 14:32:56 crc kubenswrapper[4897]: I1121 14:32:56.083309 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7b546c9e-dba8-4806-a2ed-8023271b448a-config-data-custom\") pod \"heat-engine-75bd44869-p288j\" (UID: \"7b546c9e-dba8-4806-a2ed-8023271b448a\") " pod="openstack/heat-engine-75bd44869-p288j" Nov 21 14:32:56 crc kubenswrapper[4897]: I1121 14:32:56.083331 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9142a1dd-8c8e-486c-b1f3-d2f1c3fd6728-combined-ca-bundle\") pod \"heat-api-56b88658f-4ndrt\" (UID: \"9142a1dd-8c8e-486c-b1f3-d2f1c3fd6728\") " pod="openstack/heat-api-56b88658f-4ndrt" Nov 21 14:32:56 crc kubenswrapper[4897]: I1121 14:32:56.083387 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e0a52a6-14c9-4529-b2a8-c26b739b7986-config-data\") pod \"heat-cfnapi-55749dccc7-lbx4x\" (UID: \"2e0a52a6-14c9-4529-b2a8-c26b739b7986\") " pod="openstack/heat-cfnapi-55749dccc7-lbx4x" Nov 21 14:32:56 crc kubenswrapper[4897]: I1121 14:32:56.083534 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w2w4k\" (UniqueName: \"kubernetes.io/projected/2e0a52a6-14c9-4529-b2a8-c26b739b7986-kube-api-access-w2w4k\") pod \"heat-cfnapi-55749dccc7-lbx4x\" (UID: \"2e0a52a6-14c9-4529-b2a8-c26b739b7986\") " pod="openstack/heat-cfnapi-55749dccc7-lbx4x" Nov 21 14:32:56 crc kubenswrapper[4897]: I1121 14:32:56.099278 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b546c9e-dba8-4806-a2ed-8023271b448a-combined-ca-bundle\") pod \"heat-engine-75bd44869-p288j\" (UID: \"7b546c9e-dba8-4806-a2ed-8023271b448a\") " pod="openstack/heat-engine-75bd44869-p288j" Nov 21 14:32:56 crc kubenswrapper[4897]: I1121 14:32:56.120346 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7b546c9e-dba8-4806-a2ed-8023271b448a-config-data-custom\") pod \"heat-engine-75bd44869-p288j\" (UID: \"7b546c9e-dba8-4806-a2ed-8023271b448a\") " pod="openstack/heat-engine-75bd44869-p288j" Nov 21 14:32:56 crc kubenswrapper[4897]: I1121 14:32:56.123631 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b546c9e-dba8-4806-a2ed-8023271b448a-config-data\") pod \"heat-engine-75bd44869-p288j\" (UID: \"7b546c9e-dba8-4806-a2ed-8023271b448a\") " pod="openstack/heat-engine-75bd44869-p288j" Nov 21 14:32:56 crc kubenswrapper[4897]: I1121 14:32:56.140176 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7d978555f9-wjc8q"] Nov 21 14:32:56 crc kubenswrapper[4897]: I1121 14:32:56.144237 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hc55h\" (UniqueName: \"kubernetes.io/projected/7b546c9e-dba8-4806-a2ed-8023271b448a-kube-api-access-hc55h\") pod \"heat-engine-75bd44869-p288j\" (UID: \"7b546c9e-dba8-4806-a2ed-8023271b448a\") " pod="openstack/heat-engine-75bd44869-p288j" Nov 21 14:32:56 crc kubenswrapper[4897]: I1121 14:32:56.154001 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7d978555f9-wjc8q"] Nov 21 14:32:56 crc kubenswrapper[4897]: I1121 14:32:56.154105 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d978555f9-wjc8q" Nov 21 14:32:56 crc kubenswrapper[4897]: I1121 14:32:56.156620 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-75bd44869-p288j" Nov 21 14:32:56 crc kubenswrapper[4897]: I1121 14:32:56.186224 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9142a1dd-8c8e-486c-b1f3-d2f1c3fd6728-config-data\") pod \"heat-api-56b88658f-4ndrt\" (UID: \"9142a1dd-8c8e-486c-b1f3-d2f1c3fd6728\") " pod="openstack/heat-api-56b88658f-4ndrt" Nov 21 14:32:56 crc kubenswrapper[4897]: I1121 14:32:56.186281 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2e0a52a6-14c9-4529-b2a8-c26b739b7986-config-data-custom\") pod \"heat-cfnapi-55749dccc7-lbx4x\" (UID: \"2e0a52a6-14c9-4529-b2a8-c26b739b7986\") " pod="openstack/heat-cfnapi-55749dccc7-lbx4x" Nov 21 14:32:56 crc kubenswrapper[4897]: I1121 14:32:56.186317 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9142a1dd-8c8e-486c-b1f3-d2f1c3fd6728-combined-ca-bundle\") pod \"heat-api-56b88658f-4ndrt\" (UID: \"9142a1dd-8c8e-486c-b1f3-d2f1c3fd6728\") " pod="openstack/heat-api-56b88658f-4ndrt" Nov 21 14:32:56 crc kubenswrapper[4897]: I1121 14:32:56.186354 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e0a52a6-14c9-4529-b2a8-c26b739b7986-config-data\") pod \"heat-cfnapi-55749dccc7-lbx4x\" (UID: \"2e0a52a6-14c9-4529-b2a8-c26b739b7986\") " pod="openstack/heat-cfnapi-55749dccc7-lbx4x" Nov 21 14:32:56 crc kubenswrapper[4897]: I1121 14:32:56.186434 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w2w4k\" (UniqueName: \"kubernetes.io/projected/2e0a52a6-14c9-4529-b2a8-c26b739b7986-kube-api-access-w2w4k\") pod \"heat-cfnapi-55749dccc7-lbx4x\" (UID: \"2e0a52a6-14c9-4529-b2a8-c26b739b7986\") " pod="openstack/heat-cfnapi-55749dccc7-lbx4x" Nov 21 14:32:56 crc kubenswrapper[4897]: I1121 14:32:56.186463 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4w59b\" (UniqueName: \"kubernetes.io/projected/9142a1dd-8c8e-486c-b1f3-d2f1c3fd6728-kube-api-access-4w59b\") pod \"heat-api-56b88658f-4ndrt\" (UID: \"9142a1dd-8c8e-486c-b1f3-d2f1c3fd6728\") " pod="openstack/heat-api-56b88658f-4ndrt" Nov 21 14:32:56 crc kubenswrapper[4897]: I1121 14:32:56.186484 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e0a52a6-14c9-4529-b2a8-c26b739b7986-combined-ca-bundle\") pod \"heat-cfnapi-55749dccc7-lbx4x\" (UID: \"2e0a52a6-14c9-4529-b2a8-c26b739b7986\") " pod="openstack/heat-cfnapi-55749dccc7-lbx4x" Nov 21 14:32:56 crc kubenswrapper[4897]: I1121 14:32:56.186541 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9142a1dd-8c8e-486c-b1f3-d2f1c3fd6728-config-data-custom\") pod \"heat-api-56b88658f-4ndrt\" (UID: \"9142a1dd-8c8e-486c-b1f3-d2f1c3fd6728\") " pod="openstack/heat-api-56b88658f-4ndrt" Nov 21 14:32:56 crc kubenswrapper[4897]: I1121 14:32:56.196552 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e0a52a6-14c9-4529-b2a8-c26b739b7986-config-data\") pod \"heat-cfnapi-55749dccc7-lbx4x\" (UID: \"2e0a52a6-14c9-4529-b2a8-c26b739b7986\") " pod="openstack/heat-cfnapi-55749dccc7-lbx4x" Nov 21 14:32:56 crc kubenswrapper[4897]: I1121 14:32:56.207105 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9142a1dd-8c8e-486c-b1f3-d2f1c3fd6728-config-data\") pod \"heat-api-56b88658f-4ndrt\" (UID: \"9142a1dd-8c8e-486c-b1f3-d2f1c3fd6728\") " pod="openstack/heat-api-56b88658f-4ndrt" Nov 21 14:32:56 crc kubenswrapper[4897]: I1121 14:32:56.207666 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9142a1dd-8c8e-486c-b1f3-d2f1c3fd6728-config-data-custom\") pod \"heat-api-56b88658f-4ndrt\" (UID: \"9142a1dd-8c8e-486c-b1f3-d2f1c3fd6728\") " pod="openstack/heat-api-56b88658f-4ndrt" Nov 21 14:32:56 crc kubenswrapper[4897]: I1121 14:32:56.216787 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2e0a52a6-14c9-4529-b2a8-c26b739b7986-config-data-custom\") pod \"heat-cfnapi-55749dccc7-lbx4x\" (UID: \"2e0a52a6-14c9-4529-b2a8-c26b739b7986\") " pod="openstack/heat-cfnapi-55749dccc7-lbx4x" Nov 21 14:32:56 crc kubenswrapper[4897]: I1121 14:32:56.243733 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9142a1dd-8c8e-486c-b1f3-d2f1c3fd6728-combined-ca-bundle\") pod \"heat-api-56b88658f-4ndrt\" (UID: \"9142a1dd-8c8e-486c-b1f3-d2f1c3fd6728\") " pod="openstack/heat-api-56b88658f-4ndrt" Nov 21 14:32:56 crc kubenswrapper[4897]: I1121 14:32:56.245006 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w2w4k\" (UniqueName: \"kubernetes.io/projected/2e0a52a6-14c9-4529-b2a8-c26b739b7986-kube-api-access-w2w4k\") pod \"heat-cfnapi-55749dccc7-lbx4x\" (UID: \"2e0a52a6-14c9-4529-b2a8-c26b739b7986\") " pod="openstack/heat-cfnapi-55749dccc7-lbx4x" Nov 21 14:32:56 crc kubenswrapper[4897]: I1121 14:32:56.249327 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e0a52a6-14c9-4529-b2a8-c26b739b7986-combined-ca-bundle\") pod \"heat-cfnapi-55749dccc7-lbx4x\" (UID: \"2e0a52a6-14c9-4529-b2a8-c26b739b7986\") " pod="openstack/heat-cfnapi-55749dccc7-lbx4x" Nov 21 14:32:56 crc kubenswrapper[4897]: I1121 14:32:56.254135 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4w59b\" (UniqueName: \"kubernetes.io/projected/9142a1dd-8c8e-486c-b1f3-d2f1c3fd6728-kube-api-access-4w59b\") pod \"heat-api-56b88658f-4ndrt\" (UID: \"9142a1dd-8c8e-486c-b1f3-d2f1c3fd6728\") " pod="openstack/heat-api-56b88658f-4ndrt" Nov 21 14:32:56 crc kubenswrapper[4897]: I1121 14:32:56.258038 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-56b88658f-4ndrt" Nov 21 14:32:56 crc kubenswrapper[4897]: I1121 14:32:56.270628 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-55749dccc7-lbx4x" Nov 21 14:32:56 crc kubenswrapper[4897]: I1121 14:32:56.289163 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/afa69523-63c4-4a94-b94a-61c1343ef10c-config\") pod \"dnsmasq-dns-7d978555f9-wjc8q\" (UID: \"afa69523-63c4-4a94-b94a-61c1343ef10c\") " pod="openstack/dnsmasq-dns-7d978555f9-wjc8q" Nov 21 14:32:56 crc kubenswrapper[4897]: I1121 14:32:56.289287 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5lx7z\" (UniqueName: \"kubernetes.io/projected/afa69523-63c4-4a94-b94a-61c1343ef10c-kube-api-access-5lx7z\") pod \"dnsmasq-dns-7d978555f9-wjc8q\" (UID: \"afa69523-63c4-4a94-b94a-61c1343ef10c\") " pod="openstack/dnsmasq-dns-7d978555f9-wjc8q" Nov 21 14:32:56 crc kubenswrapper[4897]: I1121 14:32:56.289338 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/afa69523-63c4-4a94-b94a-61c1343ef10c-dns-svc\") pod \"dnsmasq-dns-7d978555f9-wjc8q\" (UID: \"afa69523-63c4-4a94-b94a-61c1343ef10c\") " pod="openstack/dnsmasq-dns-7d978555f9-wjc8q" Nov 21 14:32:56 crc kubenswrapper[4897]: I1121 14:32:56.289423 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/afa69523-63c4-4a94-b94a-61c1343ef10c-ovsdbserver-sb\") pod \"dnsmasq-dns-7d978555f9-wjc8q\" (UID: \"afa69523-63c4-4a94-b94a-61c1343ef10c\") " pod="openstack/dnsmasq-dns-7d978555f9-wjc8q" Nov 21 14:32:56 crc kubenswrapper[4897]: I1121 14:32:56.289485 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/afa69523-63c4-4a94-b94a-61c1343ef10c-ovsdbserver-nb\") pod \"dnsmasq-dns-7d978555f9-wjc8q\" (UID: \"afa69523-63c4-4a94-b94a-61c1343ef10c\") " pod="openstack/dnsmasq-dns-7d978555f9-wjc8q" Nov 21 14:32:56 crc kubenswrapper[4897]: I1121 14:32:56.289541 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/afa69523-63c4-4a94-b94a-61c1343ef10c-dns-swift-storage-0\") pod \"dnsmasq-dns-7d978555f9-wjc8q\" (UID: \"afa69523-63c4-4a94-b94a-61c1343ef10c\") " pod="openstack/dnsmasq-dns-7d978555f9-wjc8q" Nov 21 14:32:56 crc kubenswrapper[4897]: I1121 14:32:56.331903 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-f46d49697-qwn8n" event={"ID":"586453ed-44e4-4b53-b205-d4d325a8a702","Type":"ContainerStarted","Data":"3c5920e85646f63e6583c532bc1486d9050066d83a57efe79c6e4882f6c5ecd1"} Nov 21 14:32:56 crc kubenswrapper[4897]: I1121 14:32:56.355351 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7446d77dd8-rg2bn" event={"ID":"bac76d5f-def0-4fe8-babf-2c1a171f04dc","Type":"ContainerStarted","Data":"8d266e169a0a0fae0622b99a0bdaf334d3e2e83f1f5b17a4df91a215e16c0d8a"} Nov 21 14:32:56 crc kubenswrapper[4897]: I1121 14:32:56.393376 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/afa69523-63c4-4a94-b94a-61c1343ef10c-ovsdbserver-sb\") pod \"dnsmasq-dns-7d978555f9-wjc8q\" (UID: \"afa69523-63c4-4a94-b94a-61c1343ef10c\") " pod="openstack/dnsmasq-dns-7d978555f9-wjc8q" Nov 21 14:32:56 crc kubenswrapper[4897]: I1121 14:32:56.393531 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/afa69523-63c4-4a94-b94a-61c1343ef10c-ovsdbserver-nb\") pod \"dnsmasq-dns-7d978555f9-wjc8q\" (UID: \"afa69523-63c4-4a94-b94a-61c1343ef10c\") " pod="openstack/dnsmasq-dns-7d978555f9-wjc8q" Nov 21 14:32:56 crc kubenswrapper[4897]: I1121 14:32:56.393592 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/afa69523-63c4-4a94-b94a-61c1343ef10c-dns-swift-storage-0\") pod \"dnsmasq-dns-7d978555f9-wjc8q\" (UID: \"afa69523-63c4-4a94-b94a-61c1343ef10c\") " pod="openstack/dnsmasq-dns-7d978555f9-wjc8q" Nov 21 14:32:56 crc kubenswrapper[4897]: I1121 14:32:56.393699 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/afa69523-63c4-4a94-b94a-61c1343ef10c-config\") pod \"dnsmasq-dns-7d978555f9-wjc8q\" (UID: \"afa69523-63c4-4a94-b94a-61c1343ef10c\") " pod="openstack/dnsmasq-dns-7d978555f9-wjc8q" Nov 21 14:32:56 crc kubenswrapper[4897]: I1121 14:32:56.393834 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5lx7z\" (UniqueName: \"kubernetes.io/projected/afa69523-63c4-4a94-b94a-61c1343ef10c-kube-api-access-5lx7z\") pod \"dnsmasq-dns-7d978555f9-wjc8q\" (UID: \"afa69523-63c4-4a94-b94a-61c1343ef10c\") " pod="openstack/dnsmasq-dns-7d978555f9-wjc8q" Nov 21 14:32:56 crc kubenswrapper[4897]: I1121 14:32:56.393863 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/afa69523-63c4-4a94-b94a-61c1343ef10c-dns-svc\") pod \"dnsmasq-dns-7d978555f9-wjc8q\" (UID: \"afa69523-63c4-4a94-b94a-61c1343ef10c\") " pod="openstack/dnsmasq-dns-7d978555f9-wjc8q" Nov 21 14:32:56 crc kubenswrapper[4897]: I1121 14:32:56.395256 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/afa69523-63c4-4a94-b94a-61c1343ef10c-dns-swift-storage-0\") pod \"dnsmasq-dns-7d978555f9-wjc8q\" (UID: \"afa69523-63c4-4a94-b94a-61c1343ef10c\") " pod="openstack/dnsmasq-dns-7d978555f9-wjc8q" Nov 21 14:32:56 crc kubenswrapper[4897]: I1121 14:32:56.395785 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/afa69523-63c4-4a94-b94a-61c1343ef10c-config\") pod \"dnsmasq-dns-7d978555f9-wjc8q\" (UID: \"afa69523-63c4-4a94-b94a-61c1343ef10c\") " pod="openstack/dnsmasq-dns-7d978555f9-wjc8q" Nov 21 14:32:56 crc kubenswrapper[4897]: I1121 14:32:56.395868 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/afa69523-63c4-4a94-b94a-61c1343ef10c-ovsdbserver-sb\") pod \"dnsmasq-dns-7d978555f9-wjc8q\" (UID: \"afa69523-63c4-4a94-b94a-61c1343ef10c\") " pod="openstack/dnsmasq-dns-7d978555f9-wjc8q" Nov 21 14:32:56 crc kubenswrapper[4897]: I1121 14:32:56.396342 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/afa69523-63c4-4a94-b94a-61c1343ef10c-ovsdbserver-nb\") pod \"dnsmasq-dns-7d978555f9-wjc8q\" (UID: \"afa69523-63c4-4a94-b94a-61c1343ef10c\") " pod="openstack/dnsmasq-dns-7d978555f9-wjc8q" Nov 21 14:32:56 crc kubenswrapper[4897]: I1121 14:32:56.399281 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/afa69523-63c4-4a94-b94a-61c1343ef10c-dns-svc\") pod \"dnsmasq-dns-7d978555f9-wjc8q\" (UID: \"afa69523-63c4-4a94-b94a-61c1343ef10c\") " pod="openstack/dnsmasq-dns-7d978555f9-wjc8q" Nov 21 14:32:56 crc kubenswrapper[4897]: I1121 14:32:56.422324 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5lx7z\" (UniqueName: \"kubernetes.io/projected/afa69523-63c4-4a94-b94a-61c1343ef10c-kube-api-access-5lx7z\") pod \"dnsmasq-dns-7d978555f9-wjc8q\" (UID: \"afa69523-63c4-4a94-b94a-61c1343ef10c\") " pod="openstack/dnsmasq-dns-7d978555f9-wjc8q" Nov 21 14:32:56 crc kubenswrapper[4897]: I1121 14:32:56.662056 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d978555f9-wjc8q" Nov 21 14:32:56 crc kubenswrapper[4897]: I1121 14:32:56.791294 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 21 14:32:56 crc kubenswrapper[4897]: I1121 14:32:56.901876 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-75bd44869-p288j"] Nov 21 14:32:56 crc kubenswrapper[4897]: I1121 14:32:56.910228 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79802b6b-53b5-4568-b79b-83da7dbbf123-config-data\") pod \"79802b6b-53b5-4568-b79b-83da7dbbf123\" (UID: \"79802b6b-53b5-4568-b79b-83da7dbbf123\") " Nov 21 14:32:56 crc kubenswrapper[4897]: I1121 14:32:56.910328 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/79802b6b-53b5-4568-b79b-83da7dbbf123-etc-machine-id\") pod \"79802b6b-53b5-4568-b79b-83da7dbbf123\" (UID: \"79802b6b-53b5-4568-b79b-83da7dbbf123\") " Nov 21 14:32:56 crc kubenswrapper[4897]: I1121 14:32:56.910371 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jksv8\" (UniqueName: \"kubernetes.io/projected/79802b6b-53b5-4568-b79b-83da7dbbf123-kube-api-access-jksv8\") pod \"79802b6b-53b5-4568-b79b-83da7dbbf123\" (UID: \"79802b6b-53b5-4568-b79b-83da7dbbf123\") " Nov 21 14:32:56 crc kubenswrapper[4897]: I1121 14:32:56.910398 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/79802b6b-53b5-4568-b79b-83da7dbbf123-scripts\") pod \"79802b6b-53b5-4568-b79b-83da7dbbf123\" (UID: \"79802b6b-53b5-4568-b79b-83da7dbbf123\") " Nov 21 14:32:56 crc kubenswrapper[4897]: I1121 14:32:56.910446 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79802b6b-53b5-4568-b79b-83da7dbbf123-combined-ca-bundle\") pod \"79802b6b-53b5-4568-b79b-83da7dbbf123\" (UID: \"79802b6b-53b5-4568-b79b-83da7dbbf123\") " Nov 21 14:32:56 crc kubenswrapper[4897]: I1121 14:32:56.910575 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/79802b6b-53b5-4568-b79b-83da7dbbf123-config-data-custom\") pod \"79802b6b-53b5-4568-b79b-83da7dbbf123\" (UID: \"79802b6b-53b5-4568-b79b-83da7dbbf123\") " Nov 21 14:32:56 crc kubenswrapper[4897]: I1121 14:32:56.920242 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79802b6b-53b5-4568-b79b-83da7dbbf123-kube-api-access-jksv8" (OuterVolumeSpecName: "kube-api-access-jksv8") pod "79802b6b-53b5-4568-b79b-83da7dbbf123" (UID: "79802b6b-53b5-4568-b79b-83da7dbbf123"). InnerVolumeSpecName "kube-api-access-jksv8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:32:56 crc kubenswrapper[4897]: I1121 14:32:56.920336 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/79802b6b-53b5-4568-b79b-83da7dbbf123-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "79802b6b-53b5-4568-b79b-83da7dbbf123" (UID: "79802b6b-53b5-4568-b79b-83da7dbbf123"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 21 14:32:56 crc kubenswrapper[4897]: I1121 14:32:56.922961 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79802b6b-53b5-4568-b79b-83da7dbbf123-scripts" (OuterVolumeSpecName: "scripts") pod "79802b6b-53b5-4568-b79b-83da7dbbf123" (UID: "79802b6b-53b5-4568-b79b-83da7dbbf123"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:32:56 crc kubenswrapper[4897]: I1121 14:32:56.949404 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79802b6b-53b5-4568-b79b-83da7dbbf123-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "79802b6b-53b5-4568-b79b-83da7dbbf123" (UID: "79802b6b-53b5-4568-b79b-83da7dbbf123"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:32:57 crc kubenswrapper[4897]: I1121 14:32:57.013415 4897 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/79802b6b-53b5-4568-b79b-83da7dbbf123-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 21 14:32:57 crc kubenswrapper[4897]: I1121 14:32:57.014333 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jksv8\" (UniqueName: \"kubernetes.io/projected/79802b6b-53b5-4568-b79b-83da7dbbf123-kube-api-access-jksv8\") on node \"crc\" DevicePath \"\"" Nov 21 14:32:57 crc kubenswrapper[4897]: I1121 14:32:57.014406 4897 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/79802b6b-53b5-4568-b79b-83da7dbbf123-scripts\") on node \"crc\" DevicePath \"\"" Nov 21 14:32:57 crc kubenswrapper[4897]: I1121 14:32:57.014487 4897 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/79802b6b-53b5-4568-b79b-83da7dbbf123-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 21 14:32:57 crc kubenswrapper[4897]: I1121 14:32:57.046878 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79802b6b-53b5-4568-b79b-83da7dbbf123-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "79802b6b-53b5-4568-b79b-83da7dbbf123" (UID: "79802b6b-53b5-4568-b79b-83da7dbbf123"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:32:57 crc kubenswrapper[4897]: I1121 14:32:57.124679 4897 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79802b6b-53b5-4568-b79b-83da7dbbf123-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 21 14:32:57 crc kubenswrapper[4897]: I1121 14:32:57.127795 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79802b6b-53b5-4568-b79b-83da7dbbf123-config-data" (OuterVolumeSpecName: "config-data") pod "79802b6b-53b5-4568-b79b-83da7dbbf123" (UID: "79802b6b-53b5-4568-b79b-83da7dbbf123"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:32:57 crc kubenswrapper[4897]: I1121 14:32:57.226489 4897 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79802b6b-53b5-4568-b79b-83da7dbbf123-config-data\") on node \"crc\" DevicePath \"\"" Nov 21 14:32:57 crc kubenswrapper[4897]: I1121 14:32:57.282348 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-56b88658f-4ndrt"] Nov 21 14:32:57 crc kubenswrapper[4897]: I1121 14:32:57.395893 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-75bd44869-p288j" event={"ID":"7b546c9e-dba8-4806-a2ed-8023271b448a","Type":"ContainerStarted","Data":"541dc73ecc166fc2449316454225cc83f3c390225f4a3426e6e554d0f34d67d0"} Nov 21 14:32:57 crc kubenswrapper[4897]: I1121 14:32:57.426066 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"79802b6b-53b5-4568-b79b-83da7dbbf123","Type":"ContainerDied","Data":"d324833442dd6e8b1ab903afa0e6d80c1cc2e4a0ea8015cdeee79bdf10a78e81"} Nov 21 14:32:57 crc kubenswrapper[4897]: I1121 14:32:57.426305 4897 scope.go:117] "RemoveContainer" containerID="fb1780b07f00fff91feafdcf7e5944c12e5d238b232280791d8dc951cb74c7c3" Nov 21 14:32:57 crc kubenswrapper[4897]: I1121 14:32:57.426554 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 21 14:32:57 crc kubenswrapper[4897]: I1121 14:32:57.468735 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-56b88658f-4ndrt" event={"ID":"9142a1dd-8c8e-486c-b1f3-d2f1c3fd6728","Type":"ContainerStarted","Data":"7b690b5df4c8f5c06b4be7e86651a188f861add64001bf4bdc893f9d70cda30f"} Nov 21 14:32:57 crc kubenswrapper[4897]: I1121 14:32:57.470098 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-7446d77dd8-rg2bn" Nov 21 14:32:57 crc kubenswrapper[4897]: I1121 14:32:57.483568 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7d978555f9-wjc8q"] Nov 21 14:32:57 crc kubenswrapper[4897]: I1121 14:32:57.576341 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-7446d77dd8-rg2bn" podStartSLOduration=7.576322856 podStartE2EDuration="7.576322856s" podCreationTimestamp="2025-11-21 14:32:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:32:57.532783259 +0000 UTC m=+1454.817376734" watchObservedRunningTime="2025-11-21 14:32:57.576322856 +0000 UTC m=+1454.860916331" Nov 21 14:32:57 crc kubenswrapper[4897]: I1121 14:32:57.578783 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 21 14:32:57 crc kubenswrapper[4897]: I1121 14:32:57.596592 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 21 14:32:57 crc kubenswrapper[4897]: I1121 14:32:57.649719 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Nov 21 14:32:57 crc kubenswrapper[4897]: E1121 14:32:57.650512 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79802b6b-53b5-4568-b79b-83da7dbbf123" containerName="cinder-scheduler" Nov 21 14:32:57 crc kubenswrapper[4897]: I1121 14:32:57.650532 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="79802b6b-53b5-4568-b79b-83da7dbbf123" containerName="cinder-scheduler" Nov 21 14:32:57 crc kubenswrapper[4897]: E1121 14:32:57.650555 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79802b6b-53b5-4568-b79b-83da7dbbf123" containerName="probe" Nov 21 14:32:57 crc kubenswrapper[4897]: I1121 14:32:57.650561 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="79802b6b-53b5-4568-b79b-83da7dbbf123" containerName="probe" Nov 21 14:32:57 crc kubenswrapper[4897]: I1121 14:32:57.650796 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="79802b6b-53b5-4568-b79b-83da7dbbf123" containerName="probe" Nov 21 14:32:57 crc kubenswrapper[4897]: I1121 14:32:57.650806 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="79802b6b-53b5-4568-b79b-83da7dbbf123" containerName="cinder-scheduler" Nov 21 14:32:57 crc kubenswrapper[4897]: I1121 14:32:57.651966 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 21 14:32:57 crc kubenswrapper[4897]: I1121 14:32:57.654420 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Nov 21 14:32:57 crc kubenswrapper[4897]: I1121 14:32:57.661773 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 21 14:32:57 crc kubenswrapper[4897]: I1121 14:32:57.677419 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-55749dccc7-lbx4x"] Nov 21 14:32:57 crc kubenswrapper[4897]: I1121 14:32:57.760769 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/367ce30d-1813-40f4-b9f4-abeb1ce23d89-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"367ce30d-1813-40f4-b9f4-abeb1ce23d89\") " pod="openstack/cinder-scheduler-0" Nov 21 14:32:57 crc kubenswrapper[4897]: I1121 14:32:57.760871 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/367ce30d-1813-40f4-b9f4-abeb1ce23d89-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"367ce30d-1813-40f4-b9f4-abeb1ce23d89\") " pod="openstack/cinder-scheduler-0" Nov 21 14:32:57 crc kubenswrapper[4897]: I1121 14:32:57.760896 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gkf65\" (UniqueName: \"kubernetes.io/projected/367ce30d-1813-40f4-b9f4-abeb1ce23d89-kube-api-access-gkf65\") pod \"cinder-scheduler-0\" (UID: \"367ce30d-1813-40f4-b9f4-abeb1ce23d89\") " pod="openstack/cinder-scheduler-0" Nov 21 14:32:57 crc kubenswrapper[4897]: I1121 14:32:57.760941 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/367ce30d-1813-40f4-b9f4-abeb1ce23d89-config-data\") pod \"cinder-scheduler-0\" (UID: \"367ce30d-1813-40f4-b9f4-abeb1ce23d89\") " pod="openstack/cinder-scheduler-0" Nov 21 14:32:57 crc kubenswrapper[4897]: I1121 14:32:57.760974 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/367ce30d-1813-40f4-b9f4-abeb1ce23d89-scripts\") pod \"cinder-scheduler-0\" (UID: \"367ce30d-1813-40f4-b9f4-abeb1ce23d89\") " pod="openstack/cinder-scheduler-0" Nov 21 14:32:57 crc kubenswrapper[4897]: I1121 14:32:57.761059 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/367ce30d-1813-40f4-b9f4-abeb1ce23d89-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"367ce30d-1813-40f4-b9f4-abeb1ce23d89\") " pod="openstack/cinder-scheduler-0" Nov 21 14:32:57 crc kubenswrapper[4897]: I1121 14:32:57.863043 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/367ce30d-1813-40f4-b9f4-abeb1ce23d89-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"367ce30d-1813-40f4-b9f4-abeb1ce23d89\") " pod="openstack/cinder-scheduler-0" Nov 21 14:32:57 crc kubenswrapper[4897]: I1121 14:32:57.863167 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/367ce30d-1813-40f4-b9f4-abeb1ce23d89-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"367ce30d-1813-40f4-b9f4-abeb1ce23d89\") " pod="openstack/cinder-scheduler-0" Nov 21 14:32:57 crc kubenswrapper[4897]: I1121 14:32:57.863244 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/367ce30d-1813-40f4-b9f4-abeb1ce23d89-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"367ce30d-1813-40f4-b9f4-abeb1ce23d89\") " pod="openstack/cinder-scheduler-0" Nov 21 14:32:57 crc kubenswrapper[4897]: I1121 14:32:57.863266 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gkf65\" (UniqueName: \"kubernetes.io/projected/367ce30d-1813-40f4-b9f4-abeb1ce23d89-kube-api-access-gkf65\") pod \"cinder-scheduler-0\" (UID: \"367ce30d-1813-40f4-b9f4-abeb1ce23d89\") " pod="openstack/cinder-scheduler-0" Nov 21 14:32:57 crc kubenswrapper[4897]: I1121 14:32:57.863286 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/367ce30d-1813-40f4-b9f4-abeb1ce23d89-config-data\") pod \"cinder-scheduler-0\" (UID: \"367ce30d-1813-40f4-b9f4-abeb1ce23d89\") " pod="openstack/cinder-scheduler-0" Nov 21 14:32:57 crc kubenswrapper[4897]: I1121 14:32:57.863339 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/367ce30d-1813-40f4-b9f4-abeb1ce23d89-scripts\") pod \"cinder-scheduler-0\" (UID: \"367ce30d-1813-40f4-b9f4-abeb1ce23d89\") " pod="openstack/cinder-scheduler-0" Nov 21 14:32:57 crc kubenswrapper[4897]: I1121 14:32:57.869085 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/367ce30d-1813-40f4-b9f4-abeb1ce23d89-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"367ce30d-1813-40f4-b9f4-abeb1ce23d89\") " pod="openstack/cinder-scheduler-0" Nov 21 14:32:57 crc kubenswrapper[4897]: I1121 14:32:57.870788 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/367ce30d-1813-40f4-b9f4-abeb1ce23d89-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"367ce30d-1813-40f4-b9f4-abeb1ce23d89\") " pod="openstack/cinder-scheduler-0" Nov 21 14:32:57 crc kubenswrapper[4897]: I1121 14:32:57.871097 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/367ce30d-1813-40f4-b9f4-abeb1ce23d89-scripts\") pod \"cinder-scheduler-0\" (UID: \"367ce30d-1813-40f4-b9f4-abeb1ce23d89\") " pod="openstack/cinder-scheduler-0" Nov 21 14:32:57 crc kubenswrapper[4897]: I1121 14:32:57.873487 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/367ce30d-1813-40f4-b9f4-abeb1ce23d89-config-data\") pod \"cinder-scheduler-0\" (UID: \"367ce30d-1813-40f4-b9f4-abeb1ce23d89\") " pod="openstack/cinder-scheduler-0" Nov 21 14:32:57 crc kubenswrapper[4897]: I1121 14:32:57.876268 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/367ce30d-1813-40f4-b9f4-abeb1ce23d89-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"367ce30d-1813-40f4-b9f4-abeb1ce23d89\") " pod="openstack/cinder-scheduler-0" Nov 21 14:32:57 crc kubenswrapper[4897]: I1121 14:32:57.883199 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gkf65\" (UniqueName: \"kubernetes.io/projected/367ce30d-1813-40f4-b9f4-abeb1ce23d89-kube-api-access-gkf65\") pod \"cinder-scheduler-0\" (UID: \"367ce30d-1813-40f4-b9f4-abeb1ce23d89\") " pod="openstack/cinder-scheduler-0" Nov 21 14:32:57 crc kubenswrapper[4897]: I1121 14:32:57.966400 4897 scope.go:117] "RemoveContainer" containerID="cebc20d99433055d6fe1f45416a2f0a69cbf7883d422fa7d34130d3f6586593a" Nov 21 14:32:57 crc kubenswrapper[4897]: I1121 14:32:57.981042 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 21 14:32:58 crc kubenswrapper[4897]: I1121 14:32:58.107080 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="79802b6b-53b5-4568-b79b-83da7dbbf123" path="/var/lib/kubelet/pods/79802b6b-53b5-4568-b79b-83da7dbbf123/volumes" Nov 21 14:32:58 crc kubenswrapper[4897]: I1121 14:32:58.488101 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb4fc677f-zlcqd" event={"ID":"71cdff91-5167-4b4a-9958-c45c77c2f8e1","Type":"ContainerStarted","Data":"4d12d8ee4151d2e16345893cb49edfc6add6c954da730fb07daf61b9fe664b06"} Nov 21 14:32:58 crc kubenswrapper[4897]: I1121 14:32:58.489827 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-55749dccc7-lbx4x" event={"ID":"2e0a52a6-14c9-4529-b2a8-c26b739b7986","Type":"ContainerStarted","Data":"d3434a310a2ffe5692eefe2d9d9c5ed874cd1e06a83f89380323be1bb6181df3"} Nov 21 14:32:58 crc kubenswrapper[4897]: I1121 14:32:58.499927 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d978555f9-wjc8q" event={"ID":"afa69523-63c4-4a94-b94a-61c1343ef10c","Type":"ContainerStarted","Data":"ce254d218e4e45e561a94f1873bfd2ea63536085260967c8f8df6325db3bcfec"} Nov 21 14:32:58 crc kubenswrapper[4897]: I1121 14:32:58.598353 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 21 14:32:59 crc kubenswrapper[4897]: I1121 14:32:59.342638 4897 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/cinder-api-0" podUID="8a67c019-b256-4890-8fb0-1dfa3a88b7b1" containerName="cinder-api" probeResult="failure" output="Get \"https://10.217.0.201:8776/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 21 14:32:59 crc kubenswrapper[4897]: I1121 14:32:59.516741 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-75bd44869-p288j" event={"ID":"7b546c9e-dba8-4806-a2ed-8023271b448a","Type":"ContainerStarted","Data":"8e41603de28669b082f8f5511f675dd2ebd4d7fd9fc98678968408620410fbb2"} Nov 21 14:32:59 crc kubenswrapper[4897]: I1121 14:32:59.517796 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"367ce30d-1813-40f4-b9f4-abeb1ce23d89","Type":"ContainerStarted","Data":"66c2a9bdc6790084eca1f0a486779fb9fede395171fab5991d270a4206576f67"} Nov 21 14:32:59 crc kubenswrapper[4897]: I1121 14:32:59.519762 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d978555f9-wjc8q" event={"ID":"afa69523-63c4-4a94-b94a-61c1343ef10c","Type":"ContainerStarted","Data":"1597597d5344b5251bb3061739b1b61fdd9a42c65f23f537f22dac42ac7b2194"} Nov 21 14:32:59 crc kubenswrapper[4897]: I1121 14:32:59.521852 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6bb4fc677f-zlcqd" podUID="71cdff91-5167-4b4a-9958-c45c77c2f8e1" containerName="dnsmasq-dns" containerID="cri-o://4d12d8ee4151d2e16345893cb49edfc6add6c954da730fb07daf61b9fe664b06" gracePeriod=10 Nov 21 14:32:59 crc kubenswrapper[4897]: I1121 14:32:59.522118 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-f46d49697-qwn8n" event={"ID":"586453ed-44e4-4b53-b205-d4d325a8a702","Type":"ContainerStarted","Data":"333acb73a94b9e6ae08c7f571f8fc7733bb995f7fe687f1034549e2f9e99589e"} Nov 21 14:32:59 crc kubenswrapper[4897]: I1121 14:32:59.522156 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6bb4fc677f-zlcqd" Nov 21 14:32:59 crc kubenswrapper[4897]: I1121 14:32:59.548095 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6bb4fc677f-zlcqd" podStartSLOduration=9.548077245 podStartE2EDuration="9.548077245s" podCreationTimestamp="2025-11-21 14:32:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:32:59.537202953 +0000 UTC m=+1456.821796438" watchObservedRunningTime="2025-11-21 14:32:59.548077245 +0000 UTC m=+1456.832670720" Nov 21 14:32:59 crc kubenswrapper[4897]: I1121 14:32:59.653307 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-844b5bbdcf-vkljz"] Nov 21 14:32:59 crc kubenswrapper[4897]: I1121 14:32:59.655202 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-844b5bbdcf-vkljz" Nov 21 14:32:59 crc kubenswrapper[4897]: I1121 14:32:59.657982 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Nov 21 14:32:59 crc kubenswrapper[4897]: I1121 14:32:59.658051 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Nov 21 14:32:59 crc kubenswrapper[4897]: I1121 14:32:59.671594 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Nov 21 14:32:59 crc kubenswrapper[4897]: I1121 14:32:59.693914 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-844b5bbdcf-vkljz"] Nov 21 14:32:59 crc kubenswrapper[4897]: I1121 14:32:59.808512 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/71d7b310-03e0-42fa-aaed-0e938d161755-public-tls-certs\") pod \"swift-proxy-844b5bbdcf-vkljz\" (UID: \"71d7b310-03e0-42fa-aaed-0e938d161755\") " pod="openstack/swift-proxy-844b5bbdcf-vkljz" Nov 21 14:32:59 crc kubenswrapper[4897]: I1121 14:32:59.808567 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/71d7b310-03e0-42fa-aaed-0e938d161755-internal-tls-certs\") pod \"swift-proxy-844b5bbdcf-vkljz\" (UID: \"71d7b310-03e0-42fa-aaed-0e938d161755\") " pod="openstack/swift-proxy-844b5bbdcf-vkljz" Nov 21 14:32:59 crc kubenswrapper[4897]: I1121 14:32:59.808600 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71d7b310-03e0-42fa-aaed-0e938d161755-config-data\") pod \"swift-proxy-844b5bbdcf-vkljz\" (UID: \"71d7b310-03e0-42fa-aaed-0e938d161755\") " pod="openstack/swift-proxy-844b5bbdcf-vkljz" Nov 21 14:32:59 crc kubenswrapper[4897]: I1121 14:32:59.808626 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/71d7b310-03e0-42fa-aaed-0e938d161755-run-httpd\") pod \"swift-proxy-844b5bbdcf-vkljz\" (UID: \"71d7b310-03e0-42fa-aaed-0e938d161755\") " pod="openstack/swift-proxy-844b5bbdcf-vkljz" Nov 21 14:32:59 crc kubenswrapper[4897]: I1121 14:32:59.808650 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/71d7b310-03e0-42fa-aaed-0e938d161755-etc-swift\") pod \"swift-proxy-844b5bbdcf-vkljz\" (UID: \"71d7b310-03e0-42fa-aaed-0e938d161755\") " pod="openstack/swift-proxy-844b5bbdcf-vkljz" Nov 21 14:32:59 crc kubenswrapper[4897]: I1121 14:32:59.808682 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71d7b310-03e0-42fa-aaed-0e938d161755-combined-ca-bundle\") pod \"swift-proxy-844b5bbdcf-vkljz\" (UID: \"71d7b310-03e0-42fa-aaed-0e938d161755\") " pod="openstack/swift-proxy-844b5bbdcf-vkljz" Nov 21 14:32:59 crc kubenswrapper[4897]: I1121 14:32:59.808732 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/71d7b310-03e0-42fa-aaed-0e938d161755-log-httpd\") pod \"swift-proxy-844b5bbdcf-vkljz\" (UID: \"71d7b310-03e0-42fa-aaed-0e938d161755\") " pod="openstack/swift-proxy-844b5bbdcf-vkljz" Nov 21 14:32:59 crc kubenswrapper[4897]: I1121 14:32:59.808794 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-66xhx\" (UniqueName: \"kubernetes.io/projected/71d7b310-03e0-42fa-aaed-0e938d161755-kube-api-access-66xhx\") pod \"swift-proxy-844b5bbdcf-vkljz\" (UID: \"71d7b310-03e0-42fa-aaed-0e938d161755\") " pod="openstack/swift-proxy-844b5bbdcf-vkljz" Nov 21 14:32:59 crc kubenswrapper[4897]: I1121 14:32:59.910270 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71d7b310-03e0-42fa-aaed-0e938d161755-config-data\") pod \"swift-proxy-844b5bbdcf-vkljz\" (UID: \"71d7b310-03e0-42fa-aaed-0e938d161755\") " pod="openstack/swift-proxy-844b5bbdcf-vkljz" Nov 21 14:32:59 crc kubenswrapper[4897]: I1121 14:32:59.910619 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/71d7b310-03e0-42fa-aaed-0e938d161755-run-httpd\") pod \"swift-proxy-844b5bbdcf-vkljz\" (UID: \"71d7b310-03e0-42fa-aaed-0e938d161755\") " pod="openstack/swift-proxy-844b5bbdcf-vkljz" Nov 21 14:32:59 crc kubenswrapper[4897]: I1121 14:32:59.910647 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/71d7b310-03e0-42fa-aaed-0e938d161755-etc-swift\") pod \"swift-proxy-844b5bbdcf-vkljz\" (UID: \"71d7b310-03e0-42fa-aaed-0e938d161755\") " pod="openstack/swift-proxy-844b5bbdcf-vkljz" Nov 21 14:32:59 crc kubenswrapper[4897]: I1121 14:32:59.910681 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71d7b310-03e0-42fa-aaed-0e938d161755-combined-ca-bundle\") pod \"swift-proxy-844b5bbdcf-vkljz\" (UID: \"71d7b310-03e0-42fa-aaed-0e938d161755\") " pod="openstack/swift-proxy-844b5bbdcf-vkljz" Nov 21 14:32:59 crc kubenswrapper[4897]: I1121 14:32:59.910739 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/71d7b310-03e0-42fa-aaed-0e938d161755-log-httpd\") pod \"swift-proxy-844b5bbdcf-vkljz\" (UID: \"71d7b310-03e0-42fa-aaed-0e938d161755\") " pod="openstack/swift-proxy-844b5bbdcf-vkljz" Nov 21 14:32:59 crc kubenswrapper[4897]: I1121 14:32:59.910809 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-66xhx\" (UniqueName: \"kubernetes.io/projected/71d7b310-03e0-42fa-aaed-0e938d161755-kube-api-access-66xhx\") pod \"swift-proxy-844b5bbdcf-vkljz\" (UID: \"71d7b310-03e0-42fa-aaed-0e938d161755\") " pod="openstack/swift-proxy-844b5bbdcf-vkljz" Nov 21 14:32:59 crc kubenswrapper[4897]: I1121 14:32:59.910890 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/71d7b310-03e0-42fa-aaed-0e938d161755-public-tls-certs\") pod \"swift-proxy-844b5bbdcf-vkljz\" (UID: \"71d7b310-03e0-42fa-aaed-0e938d161755\") " pod="openstack/swift-proxy-844b5bbdcf-vkljz" Nov 21 14:32:59 crc kubenswrapper[4897]: I1121 14:32:59.910921 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/71d7b310-03e0-42fa-aaed-0e938d161755-internal-tls-certs\") pod \"swift-proxy-844b5bbdcf-vkljz\" (UID: \"71d7b310-03e0-42fa-aaed-0e938d161755\") " pod="openstack/swift-proxy-844b5bbdcf-vkljz" Nov 21 14:32:59 crc kubenswrapper[4897]: I1121 14:32:59.911029 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/71d7b310-03e0-42fa-aaed-0e938d161755-run-httpd\") pod \"swift-proxy-844b5bbdcf-vkljz\" (UID: \"71d7b310-03e0-42fa-aaed-0e938d161755\") " pod="openstack/swift-proxy-844b5bbdcf-vkljz" Nov 21 14:32:59 crc kubenswrapper[4897]: I1121 14:32:59.911270 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/71d7b310-03e0-42fa-aaed-0e938d161755-log-httpd\") pod \"swift-proxy-844b5bbdcf-vkljz\" (UID: \"71d7b310-03e0-42fa-aaed-0e938d161755\") " pod="openstack/swift-proxy-844b5bbdcf-vkljz" Nov 21 14:32:59 crc kubenswrapper[4897]: I1121 14:32:59.924622 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/71d7b310-03e0-42fa-aaed-0e938d161755-public-tls-certs\") pod \"swift-proxy-844b5bbdcf-vkljz\" (UID: \"71d7b310-03e0-42fa-aaed-0e938d161755\") " pod="openstack/swift-proxy-844b5bbdcf-vkljz" Nov 21 14:32:59 crc kubenswrapper[4897]: I1121 14:32:59.925018 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71d7b310-03e0-42fa-aaed-0e938d161755-combined-ca-bundle\") pod \"swift-proxy-844b5bbdcf-vkljz\" (UID: \"71d7b310-03e0-42fa-aaed-0e938d161755\") " pod="openstack/swift-proxy-844b5bbdcf-vkljz" Nov 21 14:32:59 crc kubenswrapper[4897]: I1121 14:32:59.934332 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71d7b310-03e0-42fa-aaed-0e938d161755-config-data\") pod \"swift-proxy-844b5bbdcf-vkljz\" (UID: \"71d7b310-03e0-42fa-aaed-0e938d161755\") " pod="openstack/swift-proxy-844b5bbdcf-vkljz" Nov 21 14:32:59 crc kubenswrapper[4897]: I1121 14:32:59.935911 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/71d7b310-03e0-42fa-aaed-0e938d161755-etc-swift\") pod \"swift-proxy-844b5bbdcf-vkljz\" (UID: \"71d7b310-03e0-42fa-aaed-0e938d161755\") " pod="openstack/swift-proxy-844b5bbdcf-vkljz" Nov 21 14:32:59 crc kubenswrapper[4897]: I1121 14:32:59.944168 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/71d7b310-03e0-42fa-aaed-0e938d161755-internal-tls-certs\") pod \"swift-proxy-844b5bbdcf-vkljz\" (UID: \"71d7b310-03e0-42fa-aaed-0e938d161755\") " pod="openstack/swift-proxy-844b5bbdcf-vkljz" Nov 21 14:32:59 crc kubenswrapper[4897]: I1121 14:32:59.954247 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-66xhx\" (UniqueName: \"kubernetes.io/projected/71d7b310-03e0-42fa-aaed-0e938d161755-kube-api-access-66xhx\") pod \"swift-proxy-844b5bbdcf-vkljz\" (UID: \"71d7b310-03e0-42fa-aaed-0e938d161755\") " pod="openstack/swift-proxy-844b5bbdcf-vkljz" Nov 21 14:33:00 crc kubenswrapper[4897]: I1121 14:33:00.024187 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-844b5bbdcf-vkljz" Nov 21 14:33:00 crc kubenswrapper[4897]: I1121 14:33:00.148413 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 21 14:33:00 crc kubenswrapper[4897]: I1121 14:33:00.148966 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3dc8b9a3-65bf-436b-b968-20cade83734b" containerName="ceilometer-central-agent" containerID="cri-o://2d00f6a1feaa6c3dd368484437a106b7d628e3481e0c75aed4819ed8a89197b6" gracePeriod=30 Nov 21 14:33:00 crc kubenswrapper[4897]: I1121 14:33:00.150031 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3dc8b9a3-65bf-436b-b968-20cade83734b" containerName="proxy-httpd" containerID="cri-o://d480947193a7a84560fe60f69d8dfa6bf4beb45e84e6bc5f9af9448e0cd826ab" gracePeriod=30 Nov 21 14:33:00 crc kubenswrapper[4897]: I1121 14:33:00.150067 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3dc8b9a3-65bf-436b-b968-20cade83734b" containerName="ceilometer-notification-agent" containerID="cri-o://321b7f3d268272bf59b02dea3780437f7e52fcdd88b5efebf4835bb68e5e9423" gracePeriod=30 Nov 21 14:33:00 crc kubenswrapper[4897]: I1121 14:33:00.150145 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3dc8b9a3-65bf-436b-b968-20cade83734b" containerName="sg-core" containerID="cri-o://665c0ba8bda47e588fbf7269519777bef5776b6faba0458531bf4156232646fb" gracePeriod=30 Nov 21 14:33:00 crc kubenswrapper[4897]: I1121 14:33:00.203674 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="3dc8b9a3-65bf-436b-b968-20cade83734b" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.197:3000/\": EOF" Nov 21 14:33:00 crc kubenswrapper[4897]: I1121 14:33:00.345781 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="8a67c019-b256-4890-8fb0-1dfa3a88b7b1" containerName="cinder-api" probeResult="failure" output="Get \"https://10.217.0.201:8776/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 21 14:33:00 crc kubenswrapper[4897]: I1121 14:33:00.551258 4897 generic.go:334] "Generic (PLEG): container finished" podID="71cdff91-5167-4b4a-9958-c45c77c2f8e1" containerID="4d12d8ee4151d2e16345893cb49edfc6add6c954da730fb07daf61b9fe664b06" exitCode=0 Nov 21 14:33:00 crc kubenswrapper[4897]: I1121 14:33:00.551362 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb4fc677f-zlcqd" event={"ID":"71cdff91-5167-4b4a-9958-c45c77c2f8e1","Type":"ContainerDied","Data":"4d12d8ee4151d2e16345893cb49edfc6add6c954da730fb07daf61b9fe664b06"} Nov 21 14:33:00 crc kubenswrapper[4897]: I1121 14:33:00.556731 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"367ce30d-1813-40f4-b9f4-abeb1ce23d89","Type":"ContainerStarted","Data":"d049a51284939259c3e269b26655bfc11e9528d5f878ae9f05984d8e28adcbc4"} Nov 21 14:33:00 crc kubenswrapper[4897]: I1121 14:33:00.562411 4897 generic.go:334] "Generic (PLEG): container finished" podID="3dc8b9a3-65bf-436b-b968-20cade83734b" containerID="665c0ba8bda47e588fbf7269519777bef5776b6faba0458531bf4156232646fb" exitCode=2 Nov 21 14:33:00 crc kubenswrapper[4897]: I1121 14:33:00.562483 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3dc8b9a3-65bf-436b-b968-20cade83734b","Type":"ContainerDied","Data":"665c0ba8bda47e588fbf7269519777bef5776b6faba0458531bf4156232646fb"} Nov 21 14:33:00 crc kubenswrapper[4897]: I1121 14:33:00.568876 4897 generic.go:334] "Generic (PLEG): container finished" podID="afa69523-63c4-4a94-b94a-61c1343ef10c" containerID="1597597d5344b5251bb3061739b1b61fdd9a42c65f23f537f22dac42ac7b2194" exitCode=0 Nov 21 14:33:00 crc kubenswrapper[4897]: I1121 14:33:00.569412 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d978555f9-wjc8q" event={"ID":"afa69523-63c4-4a94-b94a-61c1343ef10c","Type":"ContainerDied","Data":"1597597d5344b5251bb3061739b1b61fdd9a42c65f23f537f22dac42ac7b2194"} Nov 21 14:33:00 crc kubenswrapper[4897]: I1121 14:33:00.570386 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-f46d49697-qwn8n" Nov 21 14:33:00 crc kubenswrapper[4897]: I1121 14:33:00.669460 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-75bd44869-p288j" podStartSLOduration=5.669439954 podStartE2EDuration="5.669439954s" podCreationTimestamp="2025-11-21 14:32:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:33:00.591473462 +0000 UTC m=+1457.876066937" watchObservedRunningTime="2025-11-21 14:33:00.669439954 +0000 UTC m=+1457.954033429" Nov 21 14:33:00 crc kubenswrapper[4897]: I1121 14:33:00.750402 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-f46d49697-qwn8n" podStartSLOduration=8.750089797 podStartE2EDuration="8.750089797s" podCreationTimestamp="2025-11-21 14:32:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:33:00.683791118 +0000 UTC m=+1457.968384603" watchObservedRunningTime="2025-11-21 14:33:00.750089797 +0000 UTC m=+1458.034683272" Nov 21 14:33:00 crc kubenswrapper[4897]: I1121 14:33:00.852937 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-844b5bbdcf-vkljz"] Nov 21 14:33:01 crc kubenswrapper[4897]: I1121 14:33:01.592876 4897 generic.go:334] "Generic (PLEG): container finished" podID="3dc8b9a3-65bf-436b-b968-20cade83734b" containerID="d480947193a7a84560fe60f69d8dfa6bf4beb45e84e6bc5f9af9448e0cd826ab" exitCode=0 Nov 21 14:33:01 crc kubenswrapper[4897]: I1121 14:33:01.593526 4897 generic.go:334] "Generic (PLEG): container finished" podID="3dc8b9a3-65bf-436b-b968-20cade83734b" containerID="2d00f6a1feaa6c3dd368484437a106b7d628e3481e0c75aed4819ed8a89197b6" exitCode=0 Nov 21 14:33:01 crc kubenswrapper[4897]: I1121 14:33:01.593168 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3dc8b9a3-65bf-436b-b968-20cade83734b","Type":"ContainerDied","Data":"d480947193a7a84560fe60f69d8dfa6bf4beb45e84e6bc5f9af9448e0cd826ab"} Nov 21 14:33:01 crc kubenswrapper[4897]: I1121 14:33:01.593615 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3dc8b9a3-65bf-436b-b968-20cade83734b","Type":"ContainerDied","Data":"2d00f6a1feaa6c3dd368484437a106b7d628e3481e0c75aed4819ed8a89197b6"} Nov 21 14:33:01 crc kubenswrapper[4897]: I1121 14:33:01.597141 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-844b5bbdcf-vkljz" event={"ID":"71d7b310-03e0-42fa-aaed-0e938d161755","Type":"ContainerStarted","Data":"13df9390a2bf9509775a79e9e2ac65b59584bf20ae3cb8414f5f6585c4962237"} Nov 21 14:33:01 crc kubenswrapper[4897]: I1121 14:33:01.708340 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bb4fc677f-zlcqd" Nov 21 14:33:01 crc kubenswrapper[4897]: I1121 14:33:01.786376 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/71cdff91-5167-4b4a-9958-c45c77c2f8e1-dns-swift-storage-0\") pod \"71cdff91-5167-4b4a-9958-c45c77c2f8e1\" (UID: \"71cdff91-5167-4b4a-9958-c45c77c2f8e1\") " Nov 21 14:33:01 crc kubenswrapper[4897]: I1121 14:33:01.786526 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/71cdff91-5167-4b4a-9958-c45c77c2f8e1-ovsdbserver-nb\") pod \"71cdff91-5167-4b4a-9958-c45c77c2f8e1\" (UID: \"71cdff91-5167-4b4a-9958-c45c77c2f8e1\") " Nov 21 14:33:01 crc kubenswrapper[4897]: I1121 14:33:01.786627 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/71cdff91-5167-4b4a-9958-c45c77c2f8e1-ovsdbserver-sb\") pod \"71cdff91-5167-4b4a-9958-c45c77c2f8e1\" (UID: \"71cdff91-5167-4b4a-9958-c45c77c2f8e1\") " Nov 21 14:33:01 crc kubenswrapper[4897]: I1121 14:33:01.786651 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/71cdff91-5167-4b4a-9958-c45c77c2f8e1-dns-svc\") pod \"71cdff91-5167-4b4a-9958-c45c77c2f8e1\" (UID: \"71cdff91-5167-4b4a-9958-c45c77c2f8e1\") " Nov 21 14:33:01 crc kubenswrapper[4897]: I1121 14:33:01.786674 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/71cdff91-5167-4b4a-9958-c45c77c2f8e1-config\") pod \"71cdff91-5167-4b4a-9958-c45c77c2f8e1\" (UID: \"71cdff91-5167-4b4a-9958-c45c77c2f8e1\") " Nov 21 14:33:01 crc kubenswrapper[4897]: I1121 14:33:01.786798 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zldpw\" (UniqueName: \"kubernetes.io/projected/71cdff91-5167-4b4a-9958-c45c77c2f8e1-kube-api-access-zldpw\") pod \"71cdff91-5167-4b4a-9958-c45c77c2f8e1\" (UID: \"71cdff91-5167-4b4a-9958-c45c77c2f8e1\") " Nov 21 14:33:01 crc kubenswrapper[4897]: I1121 14:33:01.795689 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/71cdff91-5167-4b4a-9958-c45c77c2f8e1-kube-api-access-zldpw" (OuterVolumeSpecName: "kube-api-access-zldpw") pod "71cdff91-5167-4b4a-9958-c45c77c2f8e1" (UID: "71cdff91-5167-4b4a-9958-c45c77c2f8e1"). InnerVolumeSpecName "kube-api-access-zldpw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:33:01 crc kubenswrapper[4897]: I1121 14:33:01.881923 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/71cdff91-5167-4b4a-9958-c45c77c2f8e1-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "71cdff91-5167-4b4a-9958-c45c77c2f8e1" (UID: "71cdff91-5167-4b4a-9958-c45c77c2f8e1"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:33:01 crc kubenswrapper[4897]: I1121 14:33:01.886932 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/71cdff91-5167-4b4a-9958-c45c77c2f8e1-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "71cdff91-5167-4b4a-9958-c45c77c2f8e1" (UID: "71cdff91-5167-4b4a-9958-c45c77c2f8e1"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:33:01 crc kubenswrapper[4897]: I1121 14:33:01.889001 4897 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/71cdff91-5167-4b4a-9958-c45c77c2f8e1-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 21 14:33:01 crc kubenswrapper[4897]: I1121 14:33:01.889033 4897 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/71cdff91-5167-4b4a-9958-c45c77c2f8e1-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 21 14:33:01 crc kubenswrapper[4897]: I1121 14:33:01.889044 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zldpw\" (UniqueName: \"kubernetes.io/projected/71cdff91-5167-4b4a-9958-c45c77c2f8e1-kube-api-access-zldpw\") on node \"crc\" DevicePath \"\"" Nov 21 14:33:01 crc kubenswrapper[4897]: I1121 14:33:01.923632 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/71cdff91-5167-4b4a-9958-c45c77c2f8e1-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "71cdff91-5167-4b4a-9958-c45c77c2f8e1" (UID: "71cdff91-5167-4b4a-9958-c45c77c2f8e1"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:33:01 crc kubenswrapper[4897]: I1121 14:33:01.940163 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/71cdff91-5167-4b4a-9958-c45c77c2f8e1-config" (OuterVolumeSpecName: "config") pod "71cdff91-5167-4b4a-9958-c45c77c2f8e1" (UID: "71cdff91-5167-4b4a-9958-c45c77c2f8e1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:33:01 crc kubenswrapper[4897]: I1121 14:33:01.960970 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/71cdff91-5167-4b4a-9958-c45c77c2f8e1-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "71cdff91-5167-4b4a-9958-c45c77c2f8e1" (UID: "71cdff91-5167-4b4a-9958-c45c77c2f8e1"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:33:01 crc kubenswrapper[4897]: I1121 14:33:01.994791 4897 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/71cdff91-5167-4b4a-9958-c45c77c2f8e1-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 21 14:33:01 crc kubenswrapper[4897]: I1121 14:33:01.994825 4897 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/71cdff91-5167-4b4a-9958-c45c77c2f8e1-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 21 14:33:01 crc kubenswrapper[4897]: I1121 14:33:01.994837 4897 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/71cdff91-5167-4b4a-9958-c45c77c2f8e1-config\") on node \"crc\" DevicePath \"\"" Nov 21 14:33:02 crc kubenswrapper[4897]: I1121 14:33:02.276228 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="3dc8b9a3-65bf-436b-b968-20cade83734b" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.197:3000/\": dial tcp 10.217.0.197:3000: connect: connection refused" Nov 21 14:33:02 crc kubenswrapper[4897]: I1121 14:33:02.386919 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-fv2n4"] Nov 21 14:33:02 crc kubenswrapper[4897]: E1121 14:33:02.387483 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71cdff91-5167-4b4a-9958-c45c77c2f8e1" containerName="dnsmasq-dns" Nov 21 14:33:02 crc kubenswrapper[4897]: I1121 14:33:02.387521 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="71cdff91-5167-4b4a-9958-c45c77c2f8e1" containerName="dnsmasq-dns" Nov 21 14:33:02 crc kubenswrapper[4897]: E1121 14:33:02.387538 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71cdff91-5167-4b4a-9958-c45c77c2f8e1" containerName="init" Nov 21 14:33:02 crc kubenswrapper[4897]: I1121 14:33:02.387547 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="71cdff91-5167-4b4a-9958-c45c77c2f8e1" containerName="init" Nov 21 14:33:02 crc kubenswrapper[4897]: I1121 14:33:02.387876 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="71cdff91-5167-4b4a-9958-c45c77c2f8e1" containerName="dnsmasq-dns" Nov 21 14:33:02 crc kubenswrapper[4897]: I1121 14:33:02.399435 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fv2n4" Nov 21 14:33:02 crc kubenswrapper[4897]: I1121 14:33:02.407859 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fv2n4"] Nov 21 14:33:02 crc kubenswrapper[4897]: I1121 14:33:02.505825 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7b274ff-8198-43ca-a3bb-b551c97762dc-catalog-content\") pod \"certified-operators-fv2n4\" (UID: \"e7b274ff-8198-43ca-a3bb-b551c97762dc\") " pod="openshift-marketplace/certified-operators-fv2n4" Nov 21 14:33:02 crc kubenswrapper[4897]: I1121 14:33:02.505894 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-phct8\" (UniqueName: \"kubernetes.io/projected/e7b274ff-8198-43ca-a3bb-b551c97762dc-kube-api-access-phct8\") pod \"certified-operators-fv2n4\" (UID: \"e7b274ff-8198-43ca-a3bb-b551c97762dc\") " pod="openshift-marketplace/certified-operators-fv2n4" Nov 21 14:33:02 crc kubenswrapper[4897]: I1121 14:33:02.505990 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7b274ff-8198-43ca-a3bb-b551c97762dc-utilities\") pod \"certified-operators-fv2n4\" (UID: \"e7b274ff-8198-43ca-a3bb-b551c97762dc\") " pod="openshift-marketplace/certified-operators-fv2n4" Nov 21 14:33:02 crc kubenswrapper[4897]: I1121 14:33:02.607907 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7b274ff-8198-43ca-a3bb-b551c97762dc-catalog-content\") pod \"certified-operators-fv2n4\" (UID: \"e7b274ff-8198-43ca-a3bb-b551c97762dc\") " pod="openshift-marketplace/certified-operators-fv2n4" Nov 21 14:33:02 crc kubenswrapper[4897]: I1121 14:33:02.608229 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-phct8\" (UniqueName: \"kubernetes.io/projected/e7b274ff-8198-43ca-a3bb-b551c97762dc-kube-api-access-phct8\") pod \"certified-operators-fv2n4\" (UID: \"e7b274ff-8198-43ca-a3bb-b551c97762dc\") " pod="openshift-marketplace/certified-operators-fv2n4" Nov 21 14:33:02 crc kubenswrapper[4897]: I1121 14:33:02.608294 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7b274ff-8198-43ca-a3bb-b551c97762dc-utilities\") pod \"certified-operators-fv2n4\" (UID: \"e7b274ff-8198-43ca-a3bb-b551c97762dc\") " pod="openshift-marketplace/certified-operators-fv2n4" Nov 21 14:33:02 crc kubenswrapper[4897]: I1121 14:33:02.608652 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7b274ff-8198-43ca-a3bb-b551c97762dc-catalog-content\") pod \"certified-operators-fv2n4\" (UID: \"e7b274ff-8198-43ca-a3bb-b551c97762dc\") " pod="openshift-marketplace/certified-operators-fv2n4" Nov 21 14:33:02 crc kubenswrapper[4897]: I1121 14:33:02.608744 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7b274ff-8198-43ca-a3bb-b551c97762dc-utilities\") pod \"certified-operators-fv2n4\" (UID: \"e7b274ff-8198-43ca-a3bb-b551c97762dc\") " pod="openshift-marketplace/certified-operators-fv2n4" Nov 21 14:33:02 crc kubenswrapper[4897]: I1121 14:33:02.647457 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d978555f9-wjc8q" event={"ID":"afa69523-63c4-4a94-b94a-61c1343ef10c","Type":"ContainerStarted","Data":"6cb3731d3191a0096a5051532c43fd96973f7663ef9d3571618a165b3dd1afca"} Nov 21 14:33:02 crc kubenswrapper[4897]: I1121 14:33:02.648721 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-phct8\" (UniqueName: \"kubernetes.io/projected/e7b274ff-8198-43ca-a3bb-b551c97762dc-kube-api-access-phct8\") pod \"certified-operators-fv2n4\" (UID: \"e7b274ff-8198-43ca-a3bb-b551c97762dc\") " pod="openshift-marketplace/certified-operators-fv2n4" Nov 21 14:33:02 crc kubenswrapper[4897]: I1121 14:33:02.648835 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7d978555f9-wjc8q" Nov 21 14:33:02 crc kubenswrapper[4897]: I1121 14:33:02.672854 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb4fc677f-zlcqd" event={"ID":"71cdff91-5167-4b4a-9958-c45c77c2f8e1","Type":"ContainerDied","Data":"9f3cf386c81836b9b08a9759b7c9b3cb037fcb498538cdfb9be621765190bff3"} Nov 21 14:33:02 crc kubenswrapper[4897]: I1121 14:33:02.672914 4897 scope.go:117] "RemoveContainer" containerID="4d12d8ee4151d2e16345893cb49edfc6add6c954da730fb07daf61b9fe664b06" Nov 21 14:33:02 crc kubenswrapper[4897]: I1121 14:33:02.673088 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bb4fc677f-zlcqd" Nov 21 14:33:02 crc kubenswrapper[4897]: I1121 14:33:02.702163 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"367ce30d-1813-40f4-b9f4-abeb1ce23d89","Type":"ContainerStarted","Data":"f5f44f5d8844f4ffbf6013415f159ac819618606a9c36d5b89f51738a1ccf753"} Nov 21 14:33:02 crc kubenswrapper[4897]: I1121 14:33:02.709673 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7d978555f9-wjc8q" podStartSLOduration=7.709648698 podStartE2EDuration="7.709648698s" podCreationTimestamp="2025-11-21 14:32:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:33:02.678291578 +0000 UTC m=+1459.962885053" watchObservedRunningTime="2025-11-21 14:33:02.709648698 +0000 UTC m=+1459.994242173" Nov 21 14:33:02 crc kubenswrapper[4897]: I1121 14:33:02.719830 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-844b5bbdcf-vkljz" event={"ID":"71d7b310-03e0-42fa-aaed-0e938d161755","Type":"ContainerStarted","Data":"44e6fab9d88d5b6a55dfb0aa8c712986701518eee5c79e9dd8ba123d2d943760"} Nov 21 14:33:02 crc kubenswrapper[4897]: I1121 14:33:02.786537 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bb4fc677f-zlcqd"] Nov 21 14:33:02 crc kubenswrapper[4897]: I1121 14:33:02.801088 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fv2n4" Nov 21 14:33:02 crc kubenswrapper[4897]: I1121 14:33:02.814570 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6bb4fc677f-zlcqd"] Nov 21 14:33:02 crc kubenswrapper[4897]: I1121 14:33:02.836310 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=5.836284416 podStartE2EDuration="5.836284416s" podCreationTimestamp="2025-11-21 14:32:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:33:02.791279248 +0000 UTC m=+1460.075872723" watchObservedRunningTime="2025-11-21 14:33:02.836284416 +0000 UTC m=+1460.120877891" Nov 21 14:33:02 crc kubenswrapper[4897]: I1121 14:33:02.981680 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Nov 21 14:33:03 crc kubenswrapper[4897]: I1121 14:33:03.288120 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-fbfff68bc-mgc44"] Nov 21 14:33:03 crc kubenswrapper[4897]: I1121 14:33:03.300337 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-fbfff68bc-mgc44" Nov 21 14:33:03 crc kubenswrapper[4897]: I1121 14:33:03.313328 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-fbfff68bc-mgc44"] Nov 21 14:33:03 crc kubenswrapper[4897]: I1121 14:33:03.334336 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-66d9b5b984-fjrkk"] Nov 21 14:33:03 crc kubenswrapper[4897]: I1121 14:33:03.337804 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-66d9b5b984-fjrkk" Nov 21 14:33:03 crc kubenswrapper[4897]: I1121 14:33:03.384932 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-5bc97dfdf4-wdbht"] Nov 21 14:33:03 crc kubenswrapper[4897]: I1121 14:33:03.386410 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-5bc97dfdf4-wdbht" Nov 21 14:33:03 crc kubenswrapper[4897]: I1121 14:33:03.462608 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-5bc97dfdf4-wdbht"] Nov 21 14:33:03 crc kubenswrapper[4897]: I1121 14:33:03.466053 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lwb8g\" (UniqueName: \"kubernetes.io/projected/2ad7667f-2a3d-445b-a582-ad2bbecc23b7-kube-api-access-lwb8g\") pod \"heat-api-5bc97dfdf4-wdbht\" (UID: \"2ad7667f-2a3d-445b-a582-ad2bbecc23b7\") " pod="openstack/heat-api-5bc97dfdf4-wdbht" Nov 21 14:33:03 crc kubenswrapper[4897]: I1121 14:33:03.466150 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ad7667f-2a3d-445b-a582-ad2bbecc23b7-config-data\") pod \"heat-api-5bc97dfdf4-wdbht\" (UID: \"2ad7667f-2a3d-445b-a582-ad2bbecc23b7\") " pod="openstack/heat-api-5bc97dfdf4-wdbht" Nov 21 14:33:03 crc kubenswrapper[4897]: I1121 14:33:03.466314 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a34d9b31-8c74-47b4-bedc-0cf717bc8f3f-config-data-custom\") pod \"heat-cfnapi-66d9b5b984-fjrkk\" (UID: \"a34d9b31-8c74-47b4-bedc-0cf717bc8f3f\") " pod="openstack/heat-cfnapi-66d9b5b984-fjrkk" Nov 21 14:33:03 crc kubenswrapper[4897]: I1121 14:33:03.466390 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a34d9b31-8c74-47b4-bedc-0cf717bc8f3f-combined-ca-bundle\") pod \"heat-cfnapi-66d9b5b984-fjrkk\" (UID: \"a34d9b31-8c74-47b4-bedc-0cf717bc8f3f\") " pod="openstack/heat-cfnapi-66d9b5b984-fjrkk" Nov 21 14:33:03 crc kubenswrapper[4897]: I1121 14:33:03.466437 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a31a2ff8-2330-4e11-90cb-83d8b22244b4-config-data-custom\") pod \"heat-engine-fbfff68bc-mgc44\" (UID: \"a31a2ff8-2330-4e11-90cb-83d8b22244b4\") " pod="openstack/heat-engine-fbfff68bc-mgc44" Nov 21 14:33:03 crc kubenswrapper[4897]: I1121 14:33:03.466530 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a31a2ff8-2330-4e11-90cb-83d8b22244b4-config-data\") pod \"heat-engine-fbfff68bc-mgc44\" (UID: \"a31a2ff8-2330-4e11-90cb-83d8b22244b4\") " pod="openstack/heat-engine-fbfff68bc-mgc44" Nov 21 14:33:03 crc kubenswrapper[4897]: I1121 14:33:03.474207 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wf2fn\" (UniqueName: \"kubernetes.io/projected/a34d9b31-8c74-47b4-bedc-0cf717bc8f3f-kube-api-access-wf2fn\") pod \"heat-cfnapi-66d9b5b984-fjrkk\" (UID: \"a34d9b31-8c74-47b4-bedc-0cf717bc8f3f\") " pod="openstack/heat-cfnapi-66d9b5b984-fjrkk" Nov 21 14:33:03 crc kubenswrapper[4897]: I1121 14:33:03.474280 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2ad7667f-2a3d-445b-a582-ad2bbecc23b7-config-data-custom\") pod \"heat-api-5bc97dfdf4-wdbht\" (UID: \"2ad7667f-2a3d-445b-a582-ad2bbecc23b7\") " pod="openstack/heat-api-5bc97dfdf4-wdbht" Nov 21 14:33:03 crc kubenswrapper[4897]: I1121 14:33:03.474330 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s7l7z\" (UniqueName: \"kubernetes.io/projected/a31a2ff8-2330-4e11-90cb-83d8b22244b4-kube-api-access-s7l7z\") pod \"heat-engine-fbfff68bc-mgc44\" (UID: \"a31a2ff8-2330-4e11-90cb-83d8b22244b4\") " pod="openstack/heat-engine-fbfff68bc-mgc44" Nov 21 14:33:03 crc kubenswrapper[4897]: I1121 14:33:03.474389 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a31a2ff8-2330-4e11-90cb-83d8b22244b4-combined-ca-bundle\") pod \"heat-engine-fbfff68bc-mgc44\" (UID: \"a31a2ff8-2330-4e11-90cb-83d8b22244b4\") " pod="openstack/heat-engine-fbfff68bc-mgc44" Nov 21 14:33:03 crc kubenswrapper[4897]: I1121 14:33:03.474413 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ad7667f-2a3d-445b-a582-ad2bbecc23b7-combined-ca-bundle\") pod \"heat-api-5bc97dfdf4-wdbht\" (UID: \"2ad7667f-2a3d-445b-a582-ad2bbecc23b7\") " pod="openstack/heat-api-5bc97dfdf4-wdbht" Nov 21 14:33:03 crc kubenswrapper[4897]: I1121 14:33:03.474440 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a34d9b31-8c74-47b4-bedc-0cf717bc8f3f-config-data\") pod \"heat-cfnapi-66d9b5b984-fjrkk\" (UID: \"a34d9b31-8c74-47b4-bedc-0cf717bc8f3f\") " pod="openstack/heat-cfnapi-66d9b5b984-fjrkk" Nov 21 14:33:03 crc kubenswrapper[4897]: I1121 14:33:03.482548 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-66d9b5b984-fjrkk"] Nov 21 14:33:03 crc kubenswrapper[4897]: I1121 14:33:03.579497 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a31a2ff8-2330-4e11-90cb-83d8b22244b4-config-data\") pod \"heat-engine-fbfff68bc-mgc44\" (UID: \"a31a2ff8-2330-4e11-90cb-83d8b22244b4\") " pod="openstack/heat-engine-fbfff68bc-mgc44" Nov 21 14:33:03 crc kubenswrapper[4897]: I1121 14:33:03.589014 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wf2fn\" (UniqueName: \"kubernetes.io/projected/a34d9b31-8c74-47b4-bedc-0cf717bc8f3f-kube-api-access-wf2fn\") pod \"heat-cfnapi-66d9b5b984-fjrkk\" (UID: \"a34d9b31-8c74-47b4-bedc-0cf717bc8f3f\") " pod="openstack/heat-cfnapi-66d9b5b984-fjrkk" Nov 21 14:33:03 crc kubenswrapper[4897]: I1121 14:33:03.589121 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2ad7667f-2a3d-445b-a582-ad2bbecc23b7-config-data-custom\") pod \"heat-api-5bc97dfdf4-wdbht\" (UID: \"2ad7667f-2a3d-445b-a582-ad2bbecc23b7\") " pod="openstack/heat-api-5bc97dfdf4-wdbht" Nov 21 14:33:03 crc kubenswrapper[4897]: I1121 14:33:03.589189 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s7l7z\" (UniqueName: \"kubernetes.io/projected/a31a2ff8-2330-4e11-90cb-83d8b22244b4-kube-api-access-s7l7z\") pod \"heat-engine-fbfff68bc-mgc44\" (UID: \"a31a2ff8-2330-4e11-90cb-83d8b22244b4\") " pod="openstack/heat-engine-fbfff68bc-mgc44" Nov 21 14:33:03 crc kubenswrapper[4897]: I1121 14:33:03.589295 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a31a2ff8-2330-4e11-90cb-83d8b22244b4-combined-ca-bundle\") pod \"heat-engine-fbfff68bc-mgc44\" (UID: \"a31a2ff8-2330-4e11-90cb-83d8b22244b4\") " pod="openstack/heat-engine-fbfff68bc-mgc44" Nov 21 14:33:03 crc kubenswrapper[4897]: I1121 14:33:03.589330 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ad7667f-2a3d-445b-a582-ad2bbecc23b7-combined-ca-bundle\") pod \"heat-api-5bc97dfdf4-wdbht\" (UID: \"2ad7667f-2a3d-445b-a582-ad2bbecc23b7\") " pod="openstack/heat-api-5bc97dfdf4-wdbht" Nov 21 14:33:03 crc kubenswrapper[4897]: I1121 14:33:03.589377 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a34d9b31-8c74-47b4-bedc-0cf717bc8f3f-config-data\") pod \"heat-cfnapi-66d9b5b984-fjrkk\" (UID: \"a34d9b31-8c74-47b4-bedc-0cf717bc8f3f\") " pod="openstack/heat-cfnapi-66d9b5b984-fjrkk" Nov 21 14:33:03 crc kubenswrapper[4897]: I1121 14:33:03.589655 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lwb8g\" (UniqueName: \"kubernetes.io/projected/2ad7667f-2a3d-445b-a582-ad2bbecc23b7-kube-api-access-lwb8g\") pod \"heat-api-5bc97dfdf4-wdbht\" (UID: \"2ad7667f-2a3d-445b-a582-ad2bbecc23b7\") " pod="openstack/heat-api-5bc97dfdf4-wdbht" Nov 21 14:33:03 crc kubenswrapper[4897]: I1121 14:33:03.589738 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ad7667f-2a3d-445b-a582-ad2bbecc23b7-config-data\") pod \"heat-api-5bc97dfdf4-wdbht\" (UID: \"2ad7667f-2a3d-445b-a582-ad2bbecc23b7\") " pod="openstack/heat-api-5bc97dfdf4-wdbht" Nov 21 14:33:03 crc kubenswrapper[4897]: I1121 14:33:03.589859 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a34d9b31-8c74-47b4-bedc-0cf717bc8f3f-config-data-custom\") pod \"heat-cfnapi-66d9b5b984-fjrkk\" (UID: \"a34d9b31-8c74-47b4-bedc-0cf717bc8f3f\") " pod="openstack/heat-cfnapi-66d9b5b984-fjrkk" Nov 21 14:33:03 crc kubenswrapper[4897]: I1121 14:33:03.589931 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a34d9b31-8c74-47b4-bedc-0cf717bc8f3f-combined-ca-bundle\") pod \"heat-cfnapi-66d9b5b984-fjrkk\" (UID: \"a34d9b31-8c74-47b4-bedc-0cf717bc8f3f\") " pod="openstack/heat-cfnapi-66d9b5b984-fjrkk" Nov 21 14:33:03 crc kubenswrapper[4897]: I1121 14:33:03.589989 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a31a2ff8-2330-4e11-90cb-83d8b22244b4-config-data-custom\") pod \"heat-engine-fbfff68bc-mgc44\" (UID: \"a31a2ff8-2330-4e11-90cb-83d8b22244b4\") " pod="openstack/heat-engine-fbfff68bc-mgc44" Nov 21 14:33:03 crc kubenswrapper[4897]: I1121 14:33:03.612661 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a34d9b31-8c74-47b4-bedc-0cf717bc8f3f-config-data\") pod \"heat-cfnapi-66d9b5b984-fjrkk\" (UID: \"a34d9b31-8c74-47b4-bedc-0cf717bc8f3f\") " pod="openstack/heat-cfnapi-66d9b5b984-fjrkk" Nov 21 14:33:03 crc kubenswrapper[4897]: I1121 14:33:03.613668 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a31a2ff8-2330-4e11-90cb-83d8b22244b4-config-data\") pod \"heat-engine-fbfff68bc-mgc44\" (UID: \"a31a2ff8-2330-4e11-90cb-83d8b22244b4\") " pod="openstack/heat-engine-fbfff68bc-mgc44" Nov 21 14:33:03 crc kubenswrapper[4897]: I1121 14:33:03.614274 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lwb8g\" (UniqueName: \"kubernetes.io/projected/2ad7667f-2a3d-445b-a582-ad2bbecc23b7-kube-api-access-lwb8g\") pod \"heat-api-5bc97dfdf4-wdbht\" (UID: \"2ad7667f-2a3d-445b-a582-ad2bbecc23b7\") " pod="openstack/heat-api-5bc97dfdf4-wdbht" Nov 21 14:33:03 crc kubenswrapper[4897]: I1121 14:33:03.614300 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ad7667f-2a3d-445b-a582-ad2bbecc23b7-combined-ca-bundle\") pod \"heat-api-5bc97dfdf4-wdbht\" (UID: \"2ad7667f-2a3d-445b-a582-ad2bbecc23b7\") " pod="openstack/heat-api-5bc97dfdf4-wdbht" Nov 21 14:33:03 crc kubenswrapper[4897]: I1121 14:33:03.615566 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a34d9b31-8c74-47b4-bedc-0cf717bc8f3f-config-data-custom\") pod \"heat-cfnapi-66d9b5b984-fjrkk\" (UID: \"a34d9b31-8c74-47b4-bedc-0cf717bc8f3f\") " pod="openstack/heat-cfnapi-66d9b5b984-fjrkk" Nov 21 14:33:03 crc kubenswrapper[4897]: I1121 14:33:03.615741 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a31a2ff8-2330-4e11-90cb-83d8b22244b4-combined-ca-bundle\") pod \"heat-engine-fbfff68bc-mgc44\" (UID: \"a31a2ff8-2330-4e11-90cb-83d8b22244b4\") " pod="openstack/heat-engine-fbfff68bc-mgc44" Nov 21 14:33:03 crc kubenswrapper[4897]: I1121 14:33:03.616337 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ad7667f-2a3d-445b-a582-ad2bbecc23b7-config-data\") pod \"heat-api-5bc97dfdf4-wdbht\" (UID: \"2ad7667f-2a3d-445b-a582-ad2bbecc23b7\") " pod="openstack/heat-api-5bc97dfdf4-wdbht" Nov 21 14:33:03 crc kubenswrapper[4897]: I1121 14:33:03.619197 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2ad7667f-2a3d-445b-a582-ad2bbecc23b7-config-data-custom\") pod \"heat-api-5bc97dfdf4-wdbht\" (UID: \"2ad7667f-2a3d-445b-a582-ad2bbecc23b7\") " pod="openstack/heat-api-5bc97dfdf4-wdbht" Nov 21 14:33:03 crc kubenswrapper[4897]: I1121 14:33:03.627222 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s7l7z\" (UniqueName: \"kubernetes.io/projected/a31a2ff8-2330-4e11-90cb-83d8b22244b4-kube-api-access-s7l7z\") pod \"heat-engine-fbfff68bc-mgc44\" (UID: \"a31a2ff8-2330-4e11-90cb-83d8b22244b4\") " pod="openstack/heat-engine-fbfff68bc-mgc44" Nov 21 14:33:03 crc kubenswrapper[4897]: I1121 14:33:03.643275 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wf2fn\" (UniqueName: \"kubernetes.io/projected/a34d9b31-8c74-47b4-bedc-0cf717bc8f3f-kube-api-access-wf2fn\") pod \"heat-cfnapi-66d9b5b984-fjrkk\" (UID: \"a34d9b31-8c74-47b4-bedc-0cf717bc8f3f\") " pod="openstack/heat-cfnapi-66d9b5b984-fjrkk" Nov 21 14:33:03 crc kubenswrapper[4897]: I1121 14:33:03.654626 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a34d9b31-8c74-47b4-bedc-0cf717bc8f3f-combined-ca-bundle\") pod \"heat-cfnapi-66d9b5b984-fjrkk\" (UID: \"a34d9b31-8c74-47b4-bedc-0cf717bc8f3f\") " pod="openstack/heat-cfnapi-66d9b5b984-fjrkk" Nov 21 14:33:03 crc kubenswrapper[4897]: I1121 14:33:03.654687 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a31a2ff8-2330-4e11-90cb-83d8b22244b4-config-data-custom\") pod \"heat-engine-fbfff68bc-mgc44\" (UID: \"a31a2ff8-2330-4e11-90cb-83d8b22244b4\") " pod="openstack/heat-engine-fbfff68bc-mgc44" Nov 21 14:33:03 crc kubenswrapper[4897]: I1121 14:33:03.717283 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-66d9b5b984-fjrkk" Nov 21 14:33:03 crc kubenswrapper[4897]: I1121 14:33:03.760045 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-5bc97dfdf4-wdbht" Nov 21 14:33:03 crc kubenswrapper[4897]: I1121 14:33:03.780847 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-844b5bbdcf-vkljz" event={"ID":"71d7b310-03e0-42fa-aaed-0e938d161755","Type":"ContainerStarted","Data":"b6ab56a2eccf3cc97584cbb5122eea6e5b20dcdca1cad46148287df5519bdf2a"} Nov 21 14:33:03 crc kubenswrapper[4897]: I1121 14:33:03.780887 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-844b5bbdcf-vkljz" Nov 21 14:33:03 crc kubenswrapper[4897]: I1121 14:33:03.780898 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-844b5bbdcf-vkljz" Nov 21 14:33:03 crc kubenswrapper[4897]: I1121 14:33:03.821029 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-844b5bbdcf-vkljz" podStartSLOduration=4.821010849 podStartE2EDuration="4.821010849s" podCreationTimestamp="2025-11-21 14:32:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:33:03.814213106 +0000 UTC m=+1461.098806581" watchObservedRunningTime="2025-11-21 14:33:03.821010849 +0000 UTC m=+1461.105604314" Nov 21 14:33:03 crc kubenswrapper[4897]: I1121 14:33:03.954927 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-fbfff68bc-mgc44" Nov 21 14:33:04 crc kubenswrapper[4897]: I1121 14:33:04.059803 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Nov 21 14:33:04 crc kubenswrapper[4897]: I1121 14:33:04.160525 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="71cdff91-5167-4b4a-9958-c45c77c2f8e1" path="/var/lib/kubelet/pods/71cdff91-5167-4b4a-9958-c45c77c2f8e1/volumes" Nov 21 14:33:04 crc kubenswrapper[4897]: E1121 14:33:04.473158 4897 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3dc8b9a3_65bf_436b_b968_20cade83734b.slice/crio-321b7f3d268272bf59b02dea3780437f7e52fcdd88b5efebf4835bb68e5e9423.scope\": RecentStats: unable to find data in memory cache]" Nov 21 14:33:04 crc kubenswrapper[4897]: I1121 14:33:04.760224 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-dtwql"] Nov 21 14:33:04 crc kubenswrapper[4897]: I1121 14:33:04.774323 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dtwql" Nov 21 14:33:04 crc kubenswrapper[4897]: I1121 14:33:04.788211 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dtwql"] Nov 21 14:33:04 crc kubenswrapper[4897]: I1121 14:33:04.851995 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d3cad20-fb17-4726-b8fa-9fd5303b5eea-catalog-content\") pod \"redhat-operators-dtwql\" (UID: \"6d3cad20-fb17-4726-b8fa-9fd5303b5eea\") " pod="openshift-marketplace/redhat-operators-dtwql" Nov 21 14:33:04 crc kubenswrapper[4897]: I1121 14:33:04.852142 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7l7sr\" (UniqueName: \"kubernetes.io/projected/6d3cad20-fb17-4726-b8fa-9fd5303b5eea-kube-api-access-7l7sr\") pod \"redhat-operators-dtwql\" (UID: \"6d3cad20-fb17-4726-b8fa-9fd5303b5eea\") " pod="openshift-marketplace/redhat-operators-dtwql" Nov 21 14:33:04 crc kubenswrapper[4897]: I1121 14:33:04.852412 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d3cad20-fb17-4726-b8fa-9fd5303b5eea-utilities\") pod \"redhat-operators-dtwql\" (UID: \"6d3cad20-fb17-4726-b8fa-9fd5303b5eea\") " pod="openshift-marketplace/redhat-operators-dtwql" Nov 21 14:33:04 crc kubenswrapper[4897]: I1121 14:33:04.958865 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d3cad20-fb17-4726-b8fa-9fd5303b5eea-utilities\") pod \"redhat-operators-dtwql\" (UID: \"6d3cad20-fb17-4726-b8fa-9fd5303b5eea\") " pod="openshift-marketplace/redhat-operators-dtwql" Nov 21 14:33:04 crc kubenswrapper[4897]: I1121 14:33:04.959125 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d3cad20-fb17-4726-b8fa-9fd5303b5eea-catalog-content\") pod \"redhat-operators-dtwql\" (UID: \"6d3cad20-fb17-4726-b8fa-9fd5303b5eea\") " pod="openshift-marketplace/redhat-operators-dtwql" Nov 21 14:33:04 crc kubenswrapper[4897]: I1121 14:33:04.959184 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7l7sr\" (UniqueName: \"kubernetes.io/projected/6d3cad20-fb17-4726-b8fa-9fd5303b5eea-kube-api-access-7l7sr\") pod \"redhat-operators-dtwql\" (UID: \"6d3cad20-fb17-4726-b8fa-9fd5303b5eea\") " pod="openshift-marketplace/redhat-operators-dtwql" Nov 21 14:33:04 crc kubenswrapper[4897]: I1121 14:33:04.960029 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d3cad20-fb17-4726-b8fa-9fd5303b5eea-utilities\") pod \"redhat-operators-dtwql\" (UID: \"6d3cad20-fb17-4726-b8fa-9fd5303b5eea\") " pod="openshift-marketplace/redhat-operators-dtwql" Nov 21 14:33:04 crc kubenswrapper[4897]: I1121 14:33:04.960091 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d3cad20-fb17-4726-b8fa-9fd5303b5eea-catalog-content\") pod \"redhat-operators-dtwql\" (UID: \"6d3cad20-fb17-4726-b8fa-9fd5303b5eea\") " pod="openshift-marketplace/redhat-operators-dtwql" Nov 21 14:33:04 crc kubenswrapper[4897]: I1121 14:33:04.982845 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7l7sr\" (UniqueName: \"kubernetes.io/projected/6d3cad20-fb17-4726-b8fa-9fd5303b5eea-kube-api-access-7l7sr\") pod \"redhat-operators-dtwql\" (UID: \"6d3cad20-fb17-4726-b8fa-9fd5303b5eea\") " pod="openshift-marketplace/redhat-operators-dtwql" Nov 21 14:33:05 crc kubenswrapper[4897]: I1121 14:33:05.109652 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dtwql" Nov 21 14:33:06 crc kubenswrapper[4897]: I1121 14:33:06.158152 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-75bd44869-p288j" Nov 21 14:33:08 crc kubenswrapper[4897]: I1121 14:33:08.862956 4897 generic.go:334] "Generic (PLEG): container finished" podID="3dc8b9a3-65bf-436b-b968-20cade83734b" containerID="321b7f3d268272bf59b02dea3780437f7e52fcdd88b5efebf4835bb68e5e9423" exitCode=0 Nov 21 14:33:08 crc kubenswrapper[4897]: I1121 14:33:08.863042 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3dc8b9a3-65bf-436b-b968-20cade83734b","Type":"ContainerDied","Data":"321b7f3d268272bf59b02dea3780437f7e52fcdd88b5efebf4835bb68e5e9423"} Nov 21 14:33:08 crc kubenswrapper[4897]: I1121 14:33:08.976519 4897 scope.go:117] "RemoveContainer" containerID="430d5775e2aa468d158479b0dee949ec7a55746b5d1bc8ab8b0afd85351b11d6" Nov 21 14:33:09 crc kubenswrapper[4897]: I1121 14:33:09.040847 4897 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/swift-proxy-844b5bbdcf-vkljz" podUID="71d7b310-03e0-42fa-aaed-0e938d161755" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Nov 21 14:33:09 crc kubenswrapper[4897]: I1121 14:33:09.048659 4897 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/swift-proxy-844b5bbdcf-vkljz" podUID="71d7b310-03e0-42fa-aaed-0e938d161755" containerName="proxy-server" probeResult="failure" output="HTTP probe failed with statuscode: 503" Nov 21 14:33:09 crc kubenswrapper[4897]: I1121 14:33:09.592672 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fv2n4"] Nov 21 14:33:09 crc kubenswrapper[4897]: I1121 14:33:09.668338 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Nov 21 14:33:09 crc kubenswrapper[4897]: I1121 14:33:09.696790 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dtwql"] Nov 21 14:33:09 crc kubenswrapper[4897]: I1121 14:33:09.732364 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-66d9b5b984-fjrkk"] Nov 21 14:33:09 crc kubenswrapper[4897]: I1121 14:33:09.859126 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-5bc97dfdf4-wdbht"] Nov 21 14:33:10 crc kubenswrapper[4897]: I1121 14:33:10.005704 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-55749dccc7-lbx4x"] Nov 21 14:33:10 crc kubenswrapper[4897]: I1121 14:33:10.032571 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-56b88658f-4ndrt"] Nov 21 14:33:10 crc kubenswrapper[4897]: I1121 14:33:10.063074 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-8459c6d79f-p9s4l"] Nov 21 14:33:10 crc kubenswrapper[4897]: I1121 14:33:10.064725 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-8459c6d79f-p9s4l" Nov 21 14:33:10 crc kubenswrapper[4897]: I1121 14:33:10.072952 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-cfnapi-internal-svc" Nov 21 14:33:10 crc kubenswrapper[4897]: I1121 14:33:10.082248 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-cfnapi-public-svc" Nov 21 14:33:10 crc kubenswrapper[4897]: I1121 14:33:10.086552 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-5d8b84cf9f-4lz8g"] Nov 21 14:33:10 crc kubenswrapper[4897]: I1121 14:33:10.088167 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-5d8b84cf9f-4lz8g" Nov 21 14:33:10 crc kubenswrapper[4897]: I1121 14:33:10.106918 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-api-public-svc" Nov 21 14:33:10 crc kubenswrapper[4897]: I1121 14:33:10.107110 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-api-internal-svc" Nov 21 14:33:10 crc kubenswrapper[4897]: I1121 14:33:10.147735 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-8459c6d79f-p9s4l"] Nov 21 14:33:10 crc kubenswrapper[4897]: I1121 14:33:10.170559 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-5d8b84cf9f-4lz8g"] Nov 21 14:33:10 crc kubenswrapper[4897]: I1121 14:33:10.187581 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-fbfff68bc-mgc44"] Nov 21 14:33:10 crc kubenswrapper[4897]: I1121 14:33:10.202247 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7cfe6c1-f519-438c-b6f1-fea0fcc70cc3-config-data\") pod \"heat-api-5d8b84cf9f-4lz8g\" (UID: \"d7cfe6c1-f519-438c-b6f1-fea0fcc70cc3\") " pod="openstack/heat-api-5d8b84cf9f-4lz8g" Nov 21 14:33:10 crc kubenswrapper[4897]: I1121 14:33:10.202337 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7cfe6c1-f519-438c-b6f1-fea0fcc70cc3-internal-tls-certs\") pod \"heat-api-5d8b84cf9f-4lz8g\" (UID: \"d7cfe6c1-f519-438c-b6f1-fea0fcc70cc3\") " pod="openstack/heat-api-5d8b84cf9f-4lz8g" Nov 21 14:33:10 crc kubenswrapper[4897]: I1121 14:33:10.202403 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d7cfe6c1-f519-438c-b6f1-fea0fcc70cc3-config-data-custom\") pod \"heat-api-5d8b84cf9f-4lz8g\" (UID: \"d7cfe6c1-f519-438c-b6f1-fea0fcc70cc3\") " pod="openstack/heat-api-5d8b84cf9f-4lz8g" Nov 21 14:33:10 crc kubenswrapper[4897]: I1121 14:33:10.202427 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0bf363b7-2fe8-4676-aa3c-4ae74c2c7d05-internal-tls-certs\") pod \"heat-cfnapi-8459c6d79f-p9s4l\" (UID: \"0bf363b7-2fe8-4676-aa3c-4ae74c2c7d05\") " pod="openstack/heat-cfnapi-8459c6d79f-p9s4l" Nov 21 14:33:10 crc kubenswrapper[4897]: I1121 14:33:10.202538 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7cfe6c1-f519-438c-b6f1-fea0fcc70cc3-combined-ca-bundle\") pod \"heat-api-5d8b84cf9f-4lz8g\" (UID: \"d7cfe6c1-f519-438c-b6f1-fea0fcc70cc3\") " pod="openstack/heat-api-5d8b84cf9f-4lz8g" Nov 21 14:33:10 crc kubenswrapper[4897]: I1121 14:33:10.202636 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7cfe6c1-f519-438c-b6f1-fea0fcc70cc3-public-tls-certs\") pod \"heat-api-5d8b84cf9f-4lz8g\" (UID: \"d7cfe6c1-f519-438c-b6f1-fea0fcc70cc3\") " pod="openstack/heat-api-5d8b84cf9f-4lz8g" Nov 21 14:33:10 crc kubenswrapper[4897]: I1121 14:33:10.202795 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5g4k9\" (UniqueName: \"kubernetes.io/projected/d7cfe6c1-f519-438c-b6f1-fea0fcc70cc3-kube-api-access-5g4k9\") pod \"heat-api-5d8b84cf9f-4lz8g\" (UID: \"d7cfe6c1-f519-438c-b6f1-fea0fcc70cc3\") " pod="openstack/heat-api-5d8b84cf9f-4lz8g" Nov 21 14:33:10 crc kubenswrapper[4897]: I1121 14:33:10.202855 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0bf363b7-2fe8-4676-aa3c-4ae74c2c7d05-public-tls-certs\") pod \"heat-cfnapi-8459c6d79f-p9s4l\" (UID: \"0bf363b7-2fe8-4676-aa3c-4ae74c2c7d05\") " pod="openstack/heat-cfnapi-8459c6d79f-p9s4l" Nov 21 14:33:10 crc kubenswrapper[4897]: I1121 14:33:10.202893 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0bf363b7-2fe8-4676-aa3c-4ae74c2c7d05-combined-ca-bundle\") pod \"heat-cfnapi-8459c6d79f-p9s4l\" (UID: \"0bf363b7-2fe8-4676-aa3c-4ae74c2c7d05\") " pod="openstack/heat-cfnapi-8459c6d79f-p9s4l" Nov 21 14:33:10 crc kubenswrapper[4897]: I1121 14:33:10.202925 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0bf363b7-2fe8-4676-aa3c-4ae74c2c7d05-config-data\") pod \"heat-cfnapi-8459c6d79f-p9s4l\" (UID: \"0bf363b7-2fe8-4676-aa3c-4ae74c2c7d05\") " pod="openstack/heat-cfnapi-8459c6d79f-p9s4l" Nov 21 14:33:10 crc kubenswrapper[4897]: I1121 14:33:10.202957 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4mgtw\" (UniqueName: \"kubernetes.io/projected/0bf363b7-2fe8-4676-aa3c-4ae74c2c7d05-kube-api-access-4mgtw\") pod \"heat-cfnapi-8459c6d79f-p9s4l\" (UID: \"0bf363b7-2fe8-4676-aa3c-4ae74c2c7d05\") " pod="openstack/heat-cfnapi-8459c6d79f-p9s4l" Nov 21 14:33:10 crc kubenswrapper[4897]: I1121 14:33:10.203000 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0bf363b7-2fe8-4676-aa3c-4ae74c2c7d05-config-data-custom\") pod \"heat-cfnapi-8459c6d79f-p9s4l\" (UID: \"0bf363b7-2fe8-4676-aa3c-4ae74c2c7d05\") " pod="openstack/heat-cfnapi-8459c6d79f-p9s4l" Nov 21 14:33:10 crc kubenswrapper[4897]: I1121 14:33:10.306698 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7cfe6c1-f519-438c-b6f1-fea0fcc70cc3-public-tls-certs\") pod \"heat-api-5d8b84cf9f-4lz8g\" (UID: \"d7cfe6c1-f519-438c-b6f1-fea0fcc70cc3\") " pod="openstack/heat-api-5d8b84cf9f-4lz8g" Nov 21 14:33:10 crc kubenswrapper[4897]: I1121 14:33:10.306807 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5g4k9\" (UniqueName: \"kubernetes.io/projected/d7cfe6c1-f519-438c-b6f1-fea0fcc70cc3-kube-api-access-5g4k9\") pod \"heat-api-5d8b84cf9f-4lz8g\" (UID: \"d7cfe6c1-f519-438c-b6f1-fea0fcc70cc3\") " pod="openstack/heat-api-5d8b84cf9f-4lz8g" Nov 21 14:33:10 crc kubenswrapper[4897]: I1121 14:33:10.306854 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0bf363b7-2fe8-4676-aa3c-4ae74c2c7d05-public-tls-certs\") pod \"heat-cfnapi-8459c6d79f-p9s4l\" (UID: \"0bf363b7-2fe8-4676-aa3c-4ae74c2c7d05\") " pod="openstack/heat-cfnapi-8459c6d79f-p9s4l" Nov 21 14:33:10 crc kubenswrapper[4897]: I1121 14:33:10.306887 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0bf363b7-2fe8-4676-aa3c-4ae74c2c7d05-combined-ca-bundle\") pod \"heat-cfnapi-8459c6d79f-p9s4l\" (UID: \"0bf363b7-2fe8-4676-aa3c-4ae74c2c7d05\") " pod="openstack/heat-cfnapi-8459c6d79f-p9s4l" Nov 21 14:33:10 crc kubenswrapper[4897]: I1121 14:33:10.306919 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0bf363b7-2fe8-4676-aa3c-4ae74c2c7d05-config-data\") pod \"heat-cfnapi-8459c6d79f-p9s4l\" (UID: \"0bf363b7-2fe8-4676-aa3c-4ae74c2c7d05\") " pod="openstack/heat-cfnapi-8459c6d79f-p9s4l" Nov 21 14:33:10 crc kubenswrapper[4897]: I1121 14:33:10.306953 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4mgtw\" (UniqueName: \"kubernetes.io/projected/0bf363b7-2fe8-4676-aa3c-4ae74c2c7d05-kube-api-access-4mgtw\") pod \"heat-cfnapi-8459c6d79f-p9s4l\" (UID: \"0bf363b7-2fe8-4676-aa3c-4ae74c2c7d05\") " pod="openstack/heat-cfnapi-8459c6d79f-p9s4l" Nov 21 14:33:10 crc kubenswrapper[4897]: I1121 14:33:10.306983 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0bf363b7-2fe8-4676-aa3c-4ae74c2c7d05-config-data-custom\") pod \"heat-cfnapi-8459c6d79f-p9s4l\" (UID: \"0bf363b7-2fe8-4676-aa3c-4ae74c2c7d05\") " pod="openstack/heat-cfnapi-8459c6d79f-p9s4l" Nov 21 14:33:10 crc kubenswrapper[4897]: I1121 14:33:10.307054 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7cfe6c1-f519-438c-b6f1-fea0fcc70cc3-config-data\") pod \"heat-api-5d8b84cf9f-4lz8g\" (UID: \"d7cfe6c1-f519-438c-b6f1-fea0fcc70cc3\") " pod="openstack/heat-api-5d8b84cf9f-4lz8g" Nov 21 14:33:10 crc kubenswrapper[4897]: I1121 14:33:10.307091 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7cfe6c1-f519-438c-b6f1-fea0fcc70cc3-internal-tls-certs\") pod \"heat-api-5d8b84cf9f-4lz8g\" (UID: \"d7cfe6c1-f519-438c-b6f1-fea0fcc70cc3\") " pod="openstack/heat-api-5d8b84cf9f-4lz8g" Nov 21 14:33:10 crc kubenswrapper[4897]: I1121 14:33:10.307146 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d7cfe6c1-f519-438c-b6f1-fea0fcc70cc3-config-data-custom\") pod \"heat-api-5d8b84cf9f-4lz8g\" (UID: \"d7cfe6c1-f519-438c-b6f1-fea0fcc70cc3\") " pod="openstack/heat-api-5d8b84cf9f-4lz8g" Nov 21 14:33:10 crc kubenswrapper[4897]: I1121 14:33:10.307171 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0bf363b7-2fe8-4676-aa3c-4ae74c2c7d05-internal-tls-certs\") pod \"heat-cfnapi-8459c6d79f-p9s4l\" (UID: \"0bf363b7-2fe8-4676-aa3c-4ae74c2c7d05\") " pod="openstack/heat-cfnapi-8459c6d79f-p9s4l" Nov 21 14:33:10 crc kubenswrapper[4897]: I1121 14:33:10.307256 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7cfe6c1-f519-438c-b6f1-fea0fcc70cc3-combined-ca-bundle\") pod \"heat-api-5d8b84cf9f-4lz8g\" (UID: \"d7cfe6c1-f519-438c-b6f1-fea0fcc70cc3\") " pod="openstack/heat-api-5d8b84cf9f-4lz8g" Nov 21 14:33:10 crc kubenswrapper[4897]: I1121 14:33:10.318208 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7cfe6c1-f519-438c-b6f1-fea0fcc70cc3-internal-tls-certs\") pod \"heat-api-5d8b84cf9f-4lz8g\" (UID: \"d7cfe6c1-f519-438c-b6f1-fea0fcc70cc3\") " pod="openstack/heat-api-5d8b84cf9f-4lz8g" Nov 21 14:33:10 crc kubenswrapper[4897]: I1121 14:33:10.326578 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0bf363b7-2fe8-4676-aa3c-4ae74c2c7d05-config-data\") pod \"heat-cfnapi-8459c6d79f-p9s4l\" (UID: \"0bf363b7-2fe8-4676-aa3c-4ae74c2c7d05\") " pod="openstack/heat-cfnapi-8459c6d79f-p9s4l" Nov 21 14:33:10 crc kubenswrapper[4897]: I1121 14:33:10.329652 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0bf363b7-2fe8-4676-aa3c-4ae74c2c7d05-internal-tls-certs\") pod \"heat-cfnapi-8459c6d79f-p9s4l\" (UID: \"0bf363b7-2fe8-4676-aa3c-4ae74c2c7d05\") " pod="openstack/heat-cfnapi-8459c6d79f-p9s4l" Nov 21 14:33:10 crc kubenswrapper[4897]: I1121 14:33:10.336563 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7cfe6c1-f519-438c-b6f1-fea0fcc70cc3-config-data\") pod \"heat-api-5d8b84cf9f-4lz8g\" (UID: \"d7cfe6c1-f519-438c-b6f1-fea0fcc70cc3\") " pod="openstack/heat-api-5d8b84cf9f-4lz8g" Nov 21 14:33:10 crc kubenswrapper[4897]: I1121 14:33:10.338442 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d7cfe6c1-f519-438c-b6f1-fea0fcc70cc3-config-data-custom\") pod \"heat-api-5d8b84cf9f-4lz8g\" (UID: \"d7cfe6c1-f519-438c-b6f1-fea0fcc70cc3\") " pod="openstack/heat-api-5d8b84cf9f-4lz8g" Nov 21 14:33:10 crc kubenswrapper[4897]: I1121 14:33:10.338526 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7cfe6c1-f519-438c-b6f1-fea0fcc70cc3-combined-ca-bundle\") pod \"heat-api-5d8b84cf9f-4lz8g\" (UID: \"d7cfe6c1-f519-438c-b6f1-fea0fcc70cc3\") " pod="openstack/heat-api-5d8b84cf9f-4lz8g" Nov 21 14:33:10 crc kubenswrapper[4897]: I1121 14:33:10.346959 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0bf363b7-2fe8-4676-aa3c-4ae74c2c7d05-public-tls-certs\") pod \"heat-cfnapi-8459c6d79f-p9s4l\" (UID: \"0bf363b7-2fe8-4676-aa3c-4ae74c2c7d05\") " pod="openstack/heat-cfnapi-8459c6d79f-p9s4l" Nov 21 14:33:10 crc kubenswrapper[4897]: I1121 14:33:10.349078 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7cfe6c1-f519-438c-b6f1-fea0fcc70cc3-public-tls-certs\") pod \"heat-api-5d8b84cf9f-4lz8g\" (UID: \"d7cfe6c1-f519-438c-b6f1-fea0fcc70cc3\") " pod="openstack/heat-api-5d8b84cf9f-4lz8g" Nov 21 14:33:10 crc kubenswrapper[4897]: I1121 14:33:10.359756 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4mgtw\" (UniqueName: \"kubernetes.io/projected/0bf363b7-2fe8-4676-aa3c-4ae74c2c7d05-kube-api-access-4mgtw\") pod \"heat-cfnapi-8459c6d79f-p9s4l\" (UID: \"0bf363b7-2fe8-4676-aa3c-4ae74c2c7d05\") " pod="openstack/heat-cfnapi-8459c6d79f-p9s4l" Nov 21 14:33:10 crc kubenswrapper[4897]: I1121 14:33:10.361306 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0bf363b7-2fe8-4676-aa3c-4ae74c2c7d05-config-data-custom\") pod \"heat-cfnapi-8459c6d79f-p9s4l\" (UID: \"0bf363b7-2fe8-4676-aa3c-4ae74c2c7d05\") " pod="openstack/heat-cfnapi-8459c6d79f-p9s4l" Nov 21 14:33:10 crc kubenswrapper[4897]: I1121 14:33:10.397477 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5g4k9\" (UniqueName: \"kubernetes.io/projected/d7cfe6c1-f519-438c-b6f1-fea0fcc70cc3-kube-api-access-5g4k9\") pod \"heat-api-5d8b84cf9f-4lz8g\" (UID: \"d7cfe6c1-f519-438c-b6f1-fea0fcc70cc3\") " pod="openstack/heat-api-5d8b84cf9f-4lz8g" Nov 21 14:33:10 crc kubenswrapper[4897]: I1121 14:33:10.401371 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0bf363b7-2fe8-4676-aa3c-4ae74c2c7d05-combined-ca-bundle\") pod \"heat-cfnapi-8459c6d79f-p9s4l\" (UID: \"0bf363b7-2fe8-4676-aa3c-4ae74c2c7d05\") " pod="openstack/heat-cfnapi-8459c6d79f-p9s4l" Nov 21 14:33:10 crc kubenswrapper[4897]: I1121 14:33:10.446098 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-8459c6d79f-p9s4l" Nov 21 14:33:10 crc kubenswrapper[4897]: I1121 14:33:10.464068 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-5d8b84cf9f-4lz8g" Nov 21 14:33:10 crc kubenswrapper[4897]: I1121 14:33:10.536825 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-844b5bbdcf-vkljz" Nov 21 14:33:10 crc kubenswrapper[4897]: I1121 14:33:10.544160 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-844b5bbdcf-vkljz" Nov 21 14:33:11 crc kubenswrapper[4897]: I1121 14:33:11.663691 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7d978555f9-wjc8q" Nov 21 14:33:11 crc kubenswrapper[4897]: I1121 14:33:11.755359 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-674b76c99f-zpcvf"] Nov 21 14:33:11 crc kubenswrapper[4897]: I1121 14:33:11.755661 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-674b76c99f-zpcvf" podUID="881842b1-c889-41be-b92a-4690f665e361" containerName="dnsmasq-dns" containerID="cri-o://ccda44133f50fc08043f235aaf8d793190f2ad7da677e86d25005b5384e7f032" gracePeriod=10 Nov 21 14:33:12 crc kubenswrapper[4897]: I1121 14:33:12.940914 4897 generic.go:334] "Generic (PLEG): container finished" podID="881842b1-c889-41be-b92a-4690f665e361" containerID="ccda44133f50fc08043f235aaf8d793190f2ad7da677e86d25005b5384e7f032" exitCode=0 Nov 21 14:33:12 crc kubenswrapper[4897]: I1121 14:33:12.940966 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-674b76c99f-zpcvf" event={"ID":"881842b1-c889-41be-b92a-4690f665e361","Type":"ContainerDied","Data":"ccda44133f50fc08043f235aaf8d793190f2ad7da677e86d25005b5384e7f032"} Nov 21 14:33:14 crc kubenswrapper[4897]: E1121 14:33:14.431610 4897 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-heat-api:current-podified" Nov 21 14:33:14 crc kubenswrapper[4897]: E1121 14:33:14.432348 4897 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:heat-api,Image:quay.io/podified-antelope-centos9/openstack-heat-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_httpd_setup && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n64chdh7dh5fbh65bh5fdhddh567h67ch545h74h76h66ch5bdh5d7h5d8h544h644h669h99h585h8ch575h5f6h5cfhd7h8bh5b6h66dhfchb5h569q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:heat-api-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-custom,ReadOnly:true,MountPath:/etc/heat/heat.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/default,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4w59b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthcheck,Port:{0 8004 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:10,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthcheck,Port:{0 8004 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:10,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42418,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42418,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-api-56b88658f-4ndrt_openstack(9142a1dd-8c8e-486c-b1f3-d2f1c3fd6728): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 21 14:33:14 crc kubenswrapper[4897]: E1121 14:33:14.433582 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/heat-api-56b88658f-4ndrt" podUID="9142a1dd-8c8e-486c-b1f3-d2f1c3fd6728" Nov 21 14:33:15 crc kubenswrapper[4897]: I1121 14:33:15.402822 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-674b76c99f-zpcvf" podUID="881842b1-c889-41be-b92a-4690f665e361" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.199:5353: connect: connection refused" Nov 21 14:33:16 crc kubenswrapper[4897]: I1121 14:33:16.340119 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-75bd44869-p288j" Nov 21 14:33:17 crc kubenswrapper[4897]: E1121 14:33:17.959192 4897 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-heat-api-cfn:current-podified" Nov 21 14:33:17 crc kubenswrapper[4897]: E1121 14:33:17.959948 4897 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:heat-cfnapi,Image:quay.io/podified-antelope-centos9/openstack-heat-api-cfn:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_httpd_setup && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n687h599h54hdbh5chd8h6dhb9h5bdh596h5dfh587h648h66chcfh7dh56bhc5h84h544h79hd4h55fh68hc9h588h6dh674h9dhbdh644h5fq,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:heat-cfnapi-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-custom,ReadOnly:true,MountPath:/etc/heat/heat.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/default,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-w2w4k,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthcheck,Port:{0 8000 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:10,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthcheck,Port:{0 8000 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:10,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42418,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42418,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-cfnapi-55749dccc7-lbx4x_openstack(2e0a52a6-14c9-4529-b2a8-c26b739b7986): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 21 14:33:17 crc kubenswrapper[4897]: E1121 14:33:17.961117 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/heat-cfnapi-55749dccc7-lbx4x" podUID="2e0a52a6-14c9-4529-b2a8-c26b739b7986" Nov 21 14:33:20 crc kubenswrapper[4897]: I1121 14:33:20.402393 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-674b76c99f-zpcvf" podUID="881842b1-c889-41be-b92a-4690f665e361" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.199:5353: connect: connection refused" Nov 21 14:33:20 crc kubenswrapper[4897]: I1121 14:33:20.893796 4897 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/neutron-7446d77dd8-rg2bn" podUID="bac76d5f-def0-4fe8-babf-2c1a171f04dc" containerName="neutron-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Nov 21 14:33:20 crc kubenswrapper[4897]: I1121 14:33:20.893825 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/neutron-7446d77dd8-rg2bn" podUID="bac76d5f-def0-4fe8-babf-2c1a171f04dc" containerName="neutron-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Nov 21 14:33:20 crc kubenswrapper[4897]: I1121 14:33:20.894227 4897 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/neutron-7446d77dd8-rg2bn" podUID="bac76d5f-def0-4fe8-babf-2c1a171f04dc" containerName="neutron-api" probeResult="failure" output="HTTP probe failed with statuscode: 503" Nov 21 14:33:35 crc kubenswrapper[4897]: W1121 14:33:23.337705 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6d3cad20_fb17_4726_b8fa_9fd5303b5eea.slice/crio-e7308deb1678f8dd48dc93b1c1d499f86812740488e7d5e6f5dba15c1e1f3f3b WatchSource:0}: Error finding container e7308deb1678f8dd48dc93b1c1d499f86812740488e7d5e6f5dba15c1e1f3f3b: Status 404 returned error can't find the container with id e7308deb1678f8dd48dc93b1c1d499f86812740488e7d5e6f5dba15c1e1f3f3b Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:23.394983 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/neutron-f46d49697-qwn8n" podUID="586453ed-44e4-4b53-b205-d4d325a8a702" containerName="neutron-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:23.401396 4897 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/neutron-f46d49697-qwn8n" podUID="586453ed-44e4-4b53-b205-d4d325a8a702" containerName="neutron-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:23.411928 4897 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/neutron-f46d49697-qwn8n" podUID="586453ed-44e4-4b53-b205-d4d325a8a702" containerName="neutron-api" probeResult="failure" output="HTTP probe failed with statuscode: 503" Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:23.723789 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:23.743762 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-56b88658f-4ndrt" Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:23.756012 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-55749dccc7-lbx4x" Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:23.830555 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3dc8b9a3-65bf-436b-b968-20cade83734b-run-httpd\") pod \"3dc8b9a3-65bf-436b-b968-20cade83734b\" (UID: \"3dc8b9a3-65bf-436b-b968-20cade83734b\") " Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:23.830646 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3dc8b9a3-65bf-436b-b968-20cade83734b-sg-core-conf-yaml\") pod \"3dc8b9a3-65bf-436b-b968-20cade83734b\" (UID: \"3dc8b9a3-65bf-436b-b968-20cade83734b\") " Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:23.830725 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3dc8b9a3-65bf-436b-b968-20cade83734b-config-data\") pod \"3dc8b9a3-65bf-436b-b968-20cade83734b\" (UID: \"3dc8b9a3-65bf-436b-b968-20cade83734b\") " Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:23.830819 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rfxtr\" (UniqueName: \"kubernetes.io/projected/3dc8b9a3-65bf-436b-b968-20cade83734b-kube-api-access-rfxtr\") pod \"3dc8b9a3-65bf-436b-b968-20cade83734b\" (UID: \"3dc8b9a3-65bf-436b-b968-20cade83734b\") " Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:23.830886 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3dc8b9a3-65bf-436b-b968-20cade83734b-combined-ca-bundle\") pod \"3dc8b9a3-65bf-436b-b968-20cade83734b\" (UID: \"3dc8b9a3-65bf-436b-b968-20cade83734b\") " Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:23.830996 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3dc8b9a3-65bf-436b-b968-20cade83734b-log-httpd\") pod \"3dc8b9a3-65bf-436b-b968-20cade83734b\" (UID: \"3dc8b9a3-65bf-436b-b968-20cade83734b\") " Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:23.831032 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3dc8b9a3-65bf-436b-b968-20cade83734b-scripts\") pod \"3dc8b9a3-65bf-436b-b968-20cade83734b\" (UID: \"3dc8b9a3-65bf-436b-b968-20cade83734b\") " Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:23.831665 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3dc8b9a3-65bf-436b-b968-20cade83734b-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "3dc8b9a3-65bf-436b-b968-20cade83734b" (UID: "3dc8b9a3-65bf-436b-b968-20cade83734b"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:23.831920 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3dc8b9a3-65bf-436b-b968-20cade83734b-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "3dc8b9a3-65bf-436b-b968-20cade83734b" (UID: "3dc8b9a3-65bf-436b-b968-20cade83734b"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:23.841132 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3dc8b9a3-65bf-436b-b968-20cade83734b-scripts" (OuterVolumeSpecName: "scripts") pod "3dc8b9a3-65bf-436b-b968-20cade83734b" (UID: "3dc8b9a3-65bf-436b-b968-20cade83734b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:23.848396 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3dc8b9a3-65bf-436b-b968-20cade83734b-kube-api-access-rfxtr" (OuterVolumeSpecName: "kube-api-access-rfxtr") pod "3dc8b9a3-65bf-436b-b968-20cade83734b" (UID: "3dc8b9a3-65bf-436b-b968-20cade83734b"). InnerVolumeSpecName "kube-api-access-rfxtr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:23.925867 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3dc8b9a3-65bf-436b-b968-20cade83734b-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "3dc8b9a3-65bf-436b-b968-20cade83734b" (UID: "3dc8b9a3-65bf-436b-b968-20cade83734b"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:23.933094 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9142a1dd-8c8e-486c-b1f3-d2f1c3fd6728-config-data-custom\") pod \"9142a1dd-8c8e-486c-b1f3-d2f1c3fd6728\" (UID: \"9142a1dd-8c8e-486c-b1f3-d2f1c3fd6728\") " Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:23.933174 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9142a1dd-8c8e-486c-b1f3-d2f1c3fd6728-combined-ca-bundle\") pod \"9142a1dd-8c8e-486c-b1f3-d2f1c3fd6728\" (UID: \"9142a1dd-8c8e-486c-b1f3-d2f1c3fd6728\") " Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:23.933230 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9142a1dd-8c8e-486c-b1f3-d2f1c3fd6728-config-data\") pod \"9142a1dd-8c8e-486c-b1f3-d2f1c3fd6728\" (UID: \"9142a1dd-8c8e-486c-b1f3-d2f1c3fd6728\") " Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:23.933261 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e0a52a6-14c9-4529-b2a8-c26b739b7986-config-data\") pod \"2e0a52a6-14c9-4529-b2a8-c26b739b7986\" (UID: \"2e0a52a6-14c9-4529-b2a8-c26b739b7986\") " Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:23.933310 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w2w4k\" (UniqueName: \"kubernetes.io/projected/2e0a52a6-14c9-4529-b2a8-c26b739b7986-kube-api-access-w2w4k\") pod \"2e0a52a6-14c9-4529-b2a8-c26b739b7986\" (UID: \"2e0a52a6-14c9-4529-b2a8-c26b739b7986\") " Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:23.933347 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2e0a52a6-14c9-4529-b2a8-c26b739b7986-config-data-custom\") pod \"2e0a52a6-14c9-4529-b2a8-c26b739b7986\" (UID: \"2e0a52a6-14c9-4529-b2a8-c26b739b7986\") " Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:23.933513 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4w59b\" (UniqueName: \"kubernetes.io/projected/9142a1dd-8c8e-486c-b1f3-d2f1c3fd6728-kube-api-access-4w59b\") pod \"9142a1dd-8c8e-486c-b1f3-d2f1c3fd6728\" (UID: \"9142a1dd-8c8e-486c-b1f3-d2f1c3fd6728\") " Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:23.933547 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e0a52a6-14c9-4529-b2a8-c26b739b7986-combined-ca-bundle\") pod \"2e0a52a6-14c9-4529-b2a8-c26b739b7986\" (UID: \"2e0a52a6-14c9-4529-b2a8-c26b739b7986\") " Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:23.934015 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rfxtr\" (UniqueName: \"kubernetes.io/projected/3dc8b9a3-65bf-436b-b968-20cade83734b-kube-api-access-rfxtr\") on node \"crc\" DevicePath \"\"" Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:23.934027 4897 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3dc8b9a3-65bf-436b-b968-20cade83734b-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:23.934035 4897 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3dc8b9a3-65bf-436b-b968-20cade83734b-scripts\") on node \"crc\" DevicePath \"\"" Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:23.934045 4897 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3dc8b9a3-65bf-436b-b968-20cade83734b-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:23.934052 4897 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3dc8b9a3-65bf-436b-b968-20cade83734b-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:23.944168 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e0a52a6-14c9-4529-b2a8-c26b739b7986-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2e0a52a6-14c9-4529-b2a8-c26b739b7986" (UID: "2e0a52a6-14c9-4529-b2a8-c26b739b7986"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:23.944873 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e0a52a6-14c9-4529-b2a8-c26b739b7986-config-data" (OuterVolumeSpecName: "config-data") pod "2e0a52a6-14c9-4529-b2a8-c26b739b7986" (UID: "2e0a52a6-14c9-4529-b2a8-c26b739b7986"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:23.945263 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9142a1dd-8c8e-486c-b1f3-d2f1c3fd6728-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "9142a1dd-8c8e-486c-b1f3-d2f1c3fd6728" (UID: "9142a1dd-8c8e-486c-b1f3-d2f1c3fd6728"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:23.946415 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e0a52a6-14c9-4529-b2a8-c26b739b7986-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "2e0a52a6-14c9-4529-b2a8-c26b739b7986" (UID: "2e0a52a6-14c9-4529-b2a8-c26b739b7986"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:23.946696 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9142a1dd-8c8e-486c-b1f3-d2f1c3fd6728-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9142a1dd-8c8e-486c-b1f3-d2f1c3fd6728" (UID: "9142a1dd-8c8e-486c-b1f3-d2f1c3fd6728"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:23.947000 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9142a1dd-8c8e-486c-b1f3-d2f1c3fd6728-kube-api-access-4w59b" (OuterVolumeSpecName: "kube-api-access-4w59b") pod "9142a1dd-8c8e-486c-b1f3-d2f1c3fd6728" (UID: "9142a1dd-8c8e-486c-b1f3-d2f1c3fd6728"). InnerVolumeSpecName "kube-api-access-4w59b". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:23.950936 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9142a1dd-8c8e-486c-b1f3-d2f1c3fd6728-config-data" (OuterVolumeSpecName: "config-data") pod "9142a1dd-8c8e-486c-b1f3-d2f1c3fd6728" (UID: "9142a1dd-8c8e-486c-b1f3-d2f1c3fd6728"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:23.951221 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e0a52a6-14c9-4529-b2a8-c26b739b7986-kube-api-access-w2w4k" (OuterVolumeSpecName: "kube-api-access-w2w4k") pod "2e0a52a6-14c9-4529-b2a8-c26b739b7986" (UID: "2e0a52a6-14c9-4529-b2a8-c26b739b7986"). InnerVolumeSpecName "kube-api-access-w2w4k". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:23.979682 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3dc8b9a3-65bf-436b-b968-20cade83734b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3dc8b9a3-65bf-436b-b968-20cade83734b" (UID: "3dc8b9a3-65bf-436b-b968-20cade83734b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:24.036289 4897 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9142a1dd-8c8e-486c-b1f3-d2f1c3fd6728-config-data\") on node \"crc\" DevicePath \"\"" Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:24.036330 4897 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e0a52a6-14c9-4529-b2a8-c26b739b7986-config-data\") on node \"crc\" DevicePath \"\"" Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:24.036344 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w2w4k\" (UniqueName: \"kubernetes.io/projected/2e0a52a6-14c9-4529-b2a8-c26b739b7986-kube-api-access-w2w4k\") on node \"crc\" DevicePath \"\"" Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:24.036371 4897 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2e0a52a6-14c9-4529-b2a8-c26b739b7986-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:24.036385 4897 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3dc8b9a3-65bf-436b-b968-20cade83734b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:24.036397 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4w59b\" (UniqueName: \"kubernetes.io/projected/9142a1dd-8c8e-486c-b1f3-d2f1c3fd6728-kube-api-access-4w59b\") on node \"crc\" DevicePath \"\"" Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:24.036409 4897 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e0a52a6-14c9-4529-b2a8-c26b739b7986-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:24.036421 4897 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9142a1dd-8c8e-486c-b1f3-d2f1c3fd6728-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:24.036439 4897 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9142a1dd-8c8e-486c-b1f3-d2f1c3fd6728-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:24.068698 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-66d9b5b984-fjrkk" event={"ID":"a34d9b31-8c74-47b4-bedc-0cf717bc8f3f","Type":"ContainerStarted","Data":"7f09c45f9512b71f97d0a96c52fcc0c293c49fa290a657bf1d5ced766942e394"} Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:24.070924 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-56b88658f-4ndrt" event={"ID":"9142a1dd-8c8e-486c-b1f3-d2f1c3fd6728","Type":"ContainerDied","Data":"7b690b5df4c8f5c06b4be7e86651a188f861add64001bf4bdc893f9d70cda30f"} Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:24.070992 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-56b88658f-4ndrt" Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:24.085367 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3dc8b9a3-65bf-436b-b968-20cade83734b","Type":"ContainerDied","Data":"4ec1b14863278e91b666cd671229063cc15246cd0bf6c9f0030795d2c16b256e"} Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:24.085408 4897 scope.go:117] "RemoveContainer" containerID="d480947193a7a84560fe60f69d8dfa6bf4beb45e84e6bc5f9af9448e0cd826ab" Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:24.085542 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:24.088119 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-55749dccc7-lbx4x" event={"ID":"2e0a52a6-14c9-4529-b2a8-c26b739b7986","Type":"ContainerDied","Data":"d3434a310a2ffe5692eefe2d9d9c5ed874cd1e06a83f89380323be1bb6181df3"} Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:24.088171 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-55749dccc7-lbx4x" Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:24.114846 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fv2n4" event={"ID":"e7b274ff-8198-43ca-a3bb-b551c97762dc","Type":"ContainerStarted","Data":"ad644f6a2ce4440f96a5a22e67eb9752858380c4b2b1c5d883a0e220659e5765"} Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:24.115067 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-5bc97dfdf4-wdbht" event={"ID":"2ad7667f-2a3d-445b-a582-ad2bbecc23b7","Type":"ContainerStarted","Data":"dce41364e86775aefa9e6523058e1b8a53ad3ab53e5ec385b443645159d74db8"} Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:24.115082 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dtwql" event={"ID":"6d3cad20-fb17-4726-b8fa-9fd5303b5eea","Type":"ContainerStarted","Data":"e7308deb1678f8dd48dc93b1c1d499f86812740488e7d5e6f5dba15c1e1f3f3b"} Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:24.115092 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-fbfff68bc-mgc44" event={"ID":"a31a2ff8-2330-4e11-90cb-83d8b22244b4","Type":"ContainerStarted","Data":"705df789c7edcc79398449bb8e4671d25534bb44738dd71beffd1d28562bf88a"} Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:24.138987 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-56b88658f-4ndrt"] Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:24.157027 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-api-56b88658f-4ndrt"] Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:24.173667 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-55749dccc7-lbx4x"] Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:24.180585 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-cfnapi-55749dccc7-lbx4x"] Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:24.362008 4897 scope.go:117] "RemoveContainer" containerID="665c0ba8bda47e588fbf7269519777bef5776b6faba0458531bf4156232646fb" Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:24.466211 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3dc8b9a3-65bf-436b-b968-20cade83734b-config-data" (OuterVolumeSpecName: "config-data") pod "3dc8b9a3-65bf-436b-b968-20cade83734b" (UID: "3dc8b9a3-65bf-436b-b968-20cade83734b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:24.545971 4897 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3dc8b9a3-65bf-436b-b968-20cade83734b-config-data\") on node \"crc\" DevicePath \"\"" Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:24.559782 4897 scope.go:117] "RemoveContainer" containerID="321b7f3d268272bf59b02dea3780437f7e52fcdd88b5efebf4835bb68e5e9423" Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:24.604085 4897 scope.go:117] "RemoveContainer" containerID="2d00f6a1feaa6c3dd368484437a106b7d628e3481e0c75aed4819ed8a89197b6" Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:24.742447 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:24.758177 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:24.773796 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 21 14:33:35 crc kubenswrapper[4897]: E1121 14:33:24.774273 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3dc8b9a3-65bf-436b-b968-20cade83734b" containerName="ceilometer-notification-agent" Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:24.774285 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="3dc8b9a3-65bf-436b-b968-20cade83734b" containerName="ceilometer-notification-agent" Nov 21 14:33:35 crc kubenswrapper[4897]: E1121 14:33:24.774316 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3dc8b9a3-65bf-436b-b968-20cade83734b" containerName="sg-core" Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:24.774323 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="3dc8b9a3-65bf-436b-b968-20cade83734b" containerName="sg-core" Nov 21 14:33:35 crc kubenswrapper[4897]: E1121 14:33:24.774347 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3dc8b9a3-65bf-436b-b968-20cade83734b" containerName="proxy-httpd" Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:24.774353 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="3dc8b9a3-65bf-436b-b968-20cade83734b" containerName="proxy-httpd" Nov 21 14:33:35 crc kubenswrapper[4897]: E1121 14:33:24.774374 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3dc8b9a3-65bf-436b-b968-20cade83734b" containerName="ceilometer-central-agent" Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:24.774380 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="3dc8b9a3-65bf-436b-b968-20cade83734b" containerName="ceilometer-central-agent" Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:24.774602 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="3dc8b9a3-65bf-436b-b968-20cade83734b" containerName="proxy-httpd" Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:24.774618 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="3dc8b9a3-65bf-436b-b968-20cade83734b" containerName="ceilometer-central-agent" Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:24.774640 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="3dc8b9a3-65bf-436b-b968-20cade83734b" containerName="sg-core" Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:24.774651 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="3dc8b9a3-65bf-436b-b968-20cade83734b" containerName="ceilometer-notification-agent" Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:24.776752 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:24.780493 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:24.781934 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:24.787749 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:24.953899 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9b4dd223-2765-4cd7-92c6-293fadac9520-log-httpd\") pod \"ceilometer-0\" (UID: \"9b4dd223-2765-4cd7-92c6-293fadac9520\") " pod="openstack/ceilometer-0" Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:24.953975 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wfj4h\" (UniqueName: \"kubernetes.io/projected/9b4dd223-2765-4cd7-92c6-293fadac9520-kube-api-access-wfj4h\") pod \"ceilometer-0\" (UID: \"9b4dd223-2765-4cd7-92c6-293fadac9520\") " pod="openstack/ceilometer-0" Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:24.954040 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9b4dd223-2765-4cd7-92c6-293fadac9520-config-data\") pod \"ceilometer-0\" (UID: \"9b4dd223-2765-4cd7-92c6-293fadac9520\") " pod="openstack/ceilometer-0" Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:24.954069 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9b4dd223-2765-4cd7-92c6-293fadac9520-run-httpd\") pod \"ceilometer-0\" (UID: \"9b4dd223-2765-4cd7-92c6-293fadac9520\") " pod="openstack/ceilometer-0" Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:24.954138 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9b4dd223-2765-4cd7-92c6-293fadac9520-scripts\") pod \"ceilometer-0\" (UID: \"9b4dd223-2765-4cd7-92c6-293fadac9520\") " pod="openstack/ceilometer-0" Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:24.954163 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b4dd223-2765-4cd7-92c6-293fadac9520-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9b4dd223-2765-4cd7-92c6-293fadac9520\") " pod="openstack/ceilometer-0" Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:24.954201 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9b4dd223-2765-4cd7-92c6-293fadac9520-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9b4dd223-2765-4cd7-92c6-293fadac9520\") " pod="openstack/ceilometer-0" Nov 21 14:33:35 crc kubenswrapper[4897]: E1121 14:33:24.971539 4897 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-openstackclient:current-podified" Nov 21 14:33:35 crc kubenswrapper[4897]: E1121 14:33:24.971883 4897 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:openstackclient,Image:quay.io/podified-antelope-centos9/openstack-openstackclient:current-podified,Command:[/bin/sleep],Args:[infinity],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n7ch66h5d7h66ch676h658hd6h575h7chddh55h599h574h8fhfhd5h64dhbfh658h67hbh567h54dh594hb6hcdh55ch58ch88h78h699h74q,ValueFrom:nil,},EnvVar{Name:OS_CLOUD,Value:default,ValueFrom:nil,},EnvVar{Name:PROMETHEUS_CA_CERT,Value:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,ValueFrom:nil,},EnvVar{Name:PROMETHEUS_HOST,Value:metric-storage-prometheus.openstack.svc,ValueFrom:nil,},EnvVar{Name:PROMETHEUS_PORT,Value:9090,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:openstack-config,ReadOnly:false,MountPath:/home/cloud-admin/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/home/cloud-admin/.config/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/home/cloud-admin/cloudrc,SubPath:cloudrc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-b2xzx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42401,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*42401,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstackclient_openstack(23323788-338f-42c8-817a-e18aa170417d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 21 14:33:35 crc kubenswrapper[4897]: E1121 14:33:24.973239 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstackclient\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/openstackclient" podUID="23323788-338f-42c8-817a-e18aa170417d" Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:25.055521 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9b4dd223-2765-4cd7-92c6-293fadac9520-config-data\") pod \"ceilometer-0\" (UID: \"9b4dd223-2765-4cd7-92c6-293fadac9520\") " pod="openstack/ceilometer-0" Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:25.055564 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9b4dd223-2765-4cd7-92c6-293fadac9520-run-httpd\") pod \"ceilometer-0\" (UID: \"9b4dd223-2765-4cd7-92c6-293fadac9520\") " pod="openstack/ceilometer-0" Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:25.055626 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9b4dd223-2765-4cd7-92c6-293fadac9520-scripts\") pod \"ceilometer-0\" (UID: \"9b4dd223-2765-4cd7-92c6-293fadac9520\") " pod="openstack/ceilometer-0" Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:25.055647 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b4dd223-2765-4cd7-92c6-293fadac9520-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9b4dd223-2765-4cd7-92c6-293fadac9520\") " pod="openstack/ceilometer-0" Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:25.055678 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9b4dd223-2765-4cd7-92c6-293fadac9520-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9b4dd223-2765-4cd7-92c6-293fadac9520\") " pod="openstack/ceilometer-0" Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:25.055772 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9b4dd223-2765-4cd7-92c6-293fadac9520-log-httpd\") pod \"ceilometer-0\" (UID: \"9b4dd223-2765-4cd7-92c6-293fadac9520\") " pod="openstack/ceilometer-0" Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:25.055814 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wfj4h\" (UniqueName: \"kubernetes.io/projected/9b4dd223-2765-4cd7-92c6-293fadac9520-kube-api-access-wfj4h\") pod \"ceilometer-0\" (UID: \"9b4dd223-2765-4cd7-92c6-293fadac9520\") " pod="openstack/ceilometer-0" Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:25.056183 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9b4dd223-2765-4cd7-92c6-293fadac9520-log-httpd\") pod \"ceilometer-0\" (UID: \"9b4dd223-2765-4cd7-92c6-293fadac9520\") " pod="openstack/ceilometer-0" Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:25.056365 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9b4dd223-2765-4cd7-92c6-293fadac9520-run-httpd\") pod \"ceilometer-0\" (UID: \"9b4dd223-2765-4cd7-92c6-293fadac9520\") " pod="openstack/ceilometer-0" Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:25.061429 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9b4dd223-2765-4cd7-92c6-293fadac9520-scripts\") pod \"ceilometer-0\" (UID: \"9b4dd223-2765-4cd7-92c6-293fadac9520\") " pod="openstack/ceilometer-0" Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:25.066421 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9b4dd223-2765-4cd7-92c6-293fadac9520-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9b4dd223-2765-4cd7-92c6-293fadac9520\") " pod="openstack/ceilometer-0" Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:25.078206 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b4dd223-2765-4cd7-92c6-293fadac9520-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9b4dd223-2765-4cd7-92c6-293fadac9520\") " pod="openstack/ceilometer-0" Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:25.078296 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9b4dd223-2765-4cd7-92c6-293fadac9520-config-data\") pod \"ceilometer-0\" (UID: \"9b4dd223-2765-4cd7-92c6-293fadac9520\") " pod="openstack/ceilometer-0" Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:25.081089 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wfj4h\" (UniqueName: \"kubernetes.io/projected/9b4dd223-2765-4cd7-92c6-293fadac9520-kube-api-access-wfj4h\") pod \"ceilometer-0\" (UID: \"9b4dd223-2765-4cd7-92c6-293fadac9520\") " pod="openstack/ceilometer-0" Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:25.104278 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:25.121135 4897 generic.go:334] "Generic (PLEG): container finished" podID="e7b274ff-8198-43ca-a3bb-b551c97762dc" containerID="54d2d5ffa197abd65ca3d2125bf24efa2801465e8b983a41d30322bf45000e69" exitCode=0 Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:25.121196 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fv2n4" event={"ID":"e7b274ff-8198-43ca-a3bb-b551c97762dc","Type":"ContainerDied","Data":"54d2d5ffa197abd65ca3d2125bf24efa2801465e8b983a41d30322bf45000e69"} Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:25.127096 4897 generic.go:334] "Generic (PLEG): container finished" podID="6d3cad20-fb17-4726-b8fa-9fd5303b5eea" containerID="a1569dddfd2a291372fa4878a85c9efe401ee175a061a9b22c686f0354f98c05" exitCode=0 Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:25.127166 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dtwql" event={"ID":"6d3cad20-fb17-4726-b8fa-9fd5303b5eea","Type":"ContainerDied","Data":"a1569dddfd2a291372fa4878a85c9efe401ee175a061a9b22c686f0354f98c05"} Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:25.133325 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-fbfff68bc-mgc44" event={"ID":"a31a2ff8-2330-4e11-90cb-83d8b22244b4","Type":"ContainerStarted","Data":"a5a70bbd5dd0125b9fb42409f1a2b98e5d87380ea8ab39814d38e7fe71ee27b8"} Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:25.133365 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-fbfff68bc-mgc44" Nov 21 14:33:35 crc kubenswrapper[4897]: E1121 14:33:25.134332 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstackclient\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-openstackclient:current-podified\\\"\"" pod="openstack/openstackclient" podUID="23323788-338f-42c8-817a-e18aa170417d" Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:25.194375 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-fbfff68bc-mgc44" podStartSLOduration=22.194351531 podStartE2EDuration="22.194351531s" podCreationTimestamp="2025-11-21 14:33:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:33:25.18424249 +0000 UTC m=+1482.468835965" watchObservedRunningTime="2025-11-21 14:33:25.194351531 +0000 UTC m=+1482.478945006" Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:25.402197 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-674b76c99f-zpcvf" podUID="881842b1-c889-41be-b92a-4690f665e361" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.199:5353: connect: connection refused" Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:25.402337 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-674b76c99f-zpcvf" Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:26.102576 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e0a52a6-14c9-4529-b2a8-c26b739b7986" path="/var/lib/kubelet/pods/2e0a52a6-14c9-4529-b2a8-c26b739b7986/volumes" Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:26.103287 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3dc8b9a3-65bf-436b-b968-20cade83734b" path="/var/lib/kubelet/pods/3dc8b9a3-65bf-436b-b968-20cade83734b/volumes" Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:26.104209 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9142a1dd-8c8e-486c-b1f3-d2f1c3fd6728" path="/var/lib/kubelet/pods/9142a1dd-8c8e-486c-b1f3-d2f1c3fd6728/volumes" Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:30.403179 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-674b76c99f-zpcvf" podUID="881842b1-c889-41be-b92a-4690f665e361" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.199:5353: connect: connection refused" Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:35.642399 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:35.906734 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-8459c6d79f-p9s4l"] Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:35.916628 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-5d8b84cf9f-4lz8g"] Nov 21 14:33:35 crc kubenswrapper[4897]: I1121 14:33:35.927220 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 21 14:33:38 crc kubenswrapper[4897]: W1121 14:33:38.276766 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd7cfe6c1_f519_438c_b6f1_fea0fcc70cc3.slice/crio-995365c351d53d709c10f9af1ca948be80a0140b54f3a3db2db29fa35ccf582e WatchSource:0}: Error finding container 995365c351d53d709c10f9af1ca948be80a0140b54f3a3db2db29fa35ccf582e: Status 404 returned error can't find the container with id 995365c351d53d709c10f9af1ca948be80a0140b54f3a3db2db29fa35ccf582e Nov 21 14:33:38 crc kubenswrapper[4897]: I1121 14:33:38.290884 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-674b76c99f-zpcvf" event={"ID":"881842b1-c889-41be-b92a-4690f665e361","Type":"ContainerDied","Data":"41554094680b0dd662bd112e1c1b06c186ecb0f5a61bbe9834ce77c7d1b4ff1a"} Nov 21 14:33:38 crc kubenswrapper[4897]: I1121 14:33:38.291215 4897 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="41554094680b0dd662bd112e1c1b06c186ecb0f5a61bbe9834ce77c7d1b4ff1a" Nov 21 14:33:38 crc kubenswrapper[4897]: I1121 14:33:38.292989 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-5d8b84cf9f-4lz8g" event={"ID":"d7cfe6c1-f519-438c-b6f1-fea0fcc70cc3","Type":"ContainerStarted","Data":"995365c351d53d709c10f9af1ca948be80a0140b54f3a3db2db29fa35ccf582e"} Nov 21 14:33:38 crc kubenswrapper[4897]: I1121 14:33:38.294755 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-8459c6d79f-p9s4l" event={"ID":"0bf363b7-2fe8-4676-aa3c-4ae74c2c7d05","Type":"ContainerStarted","Data":"607ee2bb0af6ef3c38af0d5d8727a89ce71ded12cff0147f16381028331f393f"} Nov 21 14:33:38 crc kubenswrapper[4897]: I1121 14:33:38.402756 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-674b76c99f-zpcvf" Nov 21 14:33:38 crc kubenswrapper[4897]: I1121 14:33:38.466327 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/881842b1-c889-41be-b92a-4690f665e361-ovsdbserver-sb\") pod \"881842b1-c889-41be-b92a-4690f665e361\" (UID: \"881842b1-c889-41be-b92a-4690f665e361\") " Nov 21 14:33:38 crc kubenswrapper[4897]: I1121 14:33:38.466524 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/881842b1-c889-41be-b92a-4690f665e361-config\") pod \"881842b1-c889-41be-b92a-4690f665e361\" (UID: \"881842b1-c889-41be-b92a-4690f665e361\") " Nov 21 14:33:38 crc kubenswrapper[4897]: I1121 14:33:38.466678 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/881842b1-c889-41be-b92a-4690f665e361-ovsdbserver-nb\") pod \"881842b1-c889-41be-b92a-4690f665e361\" (UID: \"881842b1-c889-41be-b92a-4690f665e361\") " Nov 21 14:33:38 crc kubenswrapper[4897]: I1121 14:33:38.466743 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/881842b1-c889-41be-b92a-4690f665e361-dns-swift-storage-0\") pod \"881842b1-c889-41be-b92a-4690f665e361\" (UID: \"881842b1-c889-41be-b92a-4690f665e361\") " Nov 21 14:33:38 crc kubenswrapper[4897]: I1121 14:33:38.466802 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vn5cf\" (UniqueName: \"kubernetes.io/projected/881842b1-c889-41be-b92a-4690f665e361-kube-api-access-vn5cf\") pod \"881842b1-c889-41be-b92a-4690f665e361\" (UID: \"881842b1-c889-41be-b92a-4690f665e361\") " Nov 21 14:33:38 crc kubenswrapper[4897]: I1121 14:33:38.466993 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/881842b1-c889-41be-b92a-4690f665e361-dns-svc\") pod \"881842b1-c889-41be-b92a-4690f665e361\" (UID: \"881842b1-c889-41be-b92a-4690f665e361\") " Nov 21 14:33:38 crc kubenswrapper[4897]: I1121 14:33:38.508579 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/881842b1-c889-41be-b92a-4690f665e361-kube-api-access-vn5cf" (OuterVolumeSpecName: "kube-api-access-vn5cf") pod "881842b1-c889-41be-b92a-4690f665e361" (UID: "881842b1-c889-41be-b92a-4690f665e361"). InnerVolumeSpecName "kube-api-access-vn5cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:33:38 crc kubenswrapper[4897]: I1121 14:33:38.574203 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vn5cf\" (UniqueName: \"kubernetes.io/projected/881842b1-c889-41be-b92a-4690f665e361-kube-api-access-vn5cf\") on node \"crc\" DevicePath \"\"" Nov 21 14:33:38 crc kubenswrapper[4897]: I1121 14:33:38.613763 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/881842b1-c889-41be-b92a-4690f665e361-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "881842b1-c889-41be-b92a-4690f665e361" (UID: "881842b1-c889-41be-b92a-4690f665e361"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:33:38 crc kubenswrapper[4897]: I1121 14:33:38.631969 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/881842b1-c889-41be-b92a-4690f665e361-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "881842b1-c889-41be-b92a-4690f665e361" (UID: "881842b1-c889-41be-b92a-4690f665e361"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:33:38 crc kubenswrapper[4897]: I1121 14:33:38.636751 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/881842b1-c889-41be-b92a-4690f665e361-config" (OuterVolumeSpecName: "config") pod "881842b1-c889-41be-b92a-4690f665e361" (UID: "881842b1-c889-41be-b92a-4690f665e361"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:33:38 crc kubenswrapper[4897]: I1121 14:33:38.676300 4897 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/881842b1-c889-41be-b92a-4690f665e361-config\") on node \"crc\" DevicePath \"\"" Nov 21 14:33:38 crc kubenswrapper[4897]: I1121 14:33:38.676703 4897 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/881842b1-c889-41be-b92a-4690f665e361-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 21 14:33:38 crc kubenswrapper[4897]: I1121 14:33:38.676716 4897 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/881842b1-c889-41be-b92a-4690f665e361-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 21 14:33:38 crc kubenswrapper[4897]: I1121 14:33:38.680725 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/881842b1-c889-41be-b92a-4690f665e361-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "881842b1-c889-41be-b92a-4690f665e361" (UID: "881842b1-c889-41be-b92a-4690f665e361"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:33:38 crc kubenswrapper[4897]: I1121 14:33:38.683175 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/881842b1-c889-41be-b92a-4690f665e361-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "881842b1-c889-41be-b92a-4690f665e361" (UID: "881842b1-c889-41be-b92a-4690f665e361"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:33:38 crc kubenswrapper[4897]: I1121 14:33:38.779348 4897 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/881842b1-c889-41be-b92a-4690f665e361-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 21 14:33:38 crc kubenswrapper[4897]: I1121 14:33:38.779385 4897 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/881842b1-c889-41be-b92a-4690f665e361-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 21 14:33:39 crc kubenswrapper[4897]: I1121 14:33:39.303798 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9b4dd223-2765-4cd7-92c6-293fadac9520","Type":"ContainerStarted","Data":"93b9b9e870a5485e47f3d6d422bf6253387084387ce1848166fc66402fb94d10"} Nov 21 14:33:39 crc kubenswrapper[4897]: I1121 14:33:39.305308 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-674b76c99f-zpcvf" Nov 21 14:33:39 crc kubenswrapper[4897]: I1121 14:33:39.316217 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-5bc97dfdf4-wdbht" event={"ID":"2ad7667f-2a3d-445b-a582-ad2bbecc23b7","Type":"ContainerStarted","Data":"f31cb8a4318a2663dea8d64af8163972acc671dfc46b55111d6f6a4c65780c72"} Nov 21 14:33:39 crc kubenswrapper[4897]: I1121 14:33:39.316256 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-5bc97dfdf4-wdbht" Nov 21 14:33:39 crc kubenswrapper[4897]: I1121 14:33:39.347009 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-5bc97dfdf4-wdbht" podStartSLOduration=21.25109985 podStartE2EDuration="36.346988791s" podCreationTimestamp="2025-11-21 14:33:03 +0000 UTC" firstStartedPulling="2025-11-21 14:33:23.407156603 +0000 UTC m=+1480.691750078" lastFinishedPulling="2025-11-21 14:33:38.503045544 +0000 UTC m=+1495.787639019" observedRunningTime="2025-11-21 14:33:39.334730152 +0000 UTC m=+1496.619323627" watchObservedRunningTime="2025-11-21 14:33:39.346988791 +0000 UTC m=+1496.631582266" Nov 21 14:33:39 crc kubenswrapper[4897]: I1121 14:33:39.366224 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-674b76c99f-zpcvf"] Nov 21 14:33:39 crc kubenswrapper[4897]: I1121 14:33:39.375748 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-674b76c99f-zpcvf"] Nov 21 14:33:40 crc kubenswrapper[4897]: I1121 14:33:40.102593 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="881842b1-c889-41be-b92a-4690f665e361" path="/var/lib/kubelet/pods/881842b1-c889-41be-b92a-4690f665e361/volumes" Nov 21 14:33:40 crc kubenswrapper[4897]: I1121 14:33:40.402592 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-674b76c99f-zpcvf" podUID="881842b1-c889-41be-b92a-4690f665e361" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.199:5353: i/o timeout" Nov 21 14:33:41 crc kubenswrapper[4897]: I1121 14:33:41.330828 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-8459c6d79f-p9s4l" event={"ID":"0bf363b7-2fe8-4676-aa3c-4ae74c2c7d05","Type":"ContainerStarted","Data":"8c9411e8019fc5114e4bb8a9ea116f4cf8b427fbab481635f9a469fe2705ff35"} Nov 21 14:33:41 crc kubenswrapper[4897]: I1121 14:33:41.331248 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-8459c6d79f-p9s4l" Nov 21 14:33:41 crc kubenswrapper[4897]: I1121 14:33:41.334656 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-66d9b5b984-fjrkk" event={"ID":"a34d9b31-8c74-47b4-bedc-0cf717bc8f3f","Type":"ContainerStarted","Data":"a011981d7a113776deb3414dc35986f7cdfec3d87d8ecb47c346409b003c0cd3"} Nov 21 14:33:41 crc kubenswrapper[4897]: I1121 14:33:41.334852 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-66d9b5b984-fjrkk" Nov 21 14:33:41 crc kubenswrapper[4897]: I1121 14:33:41.355745 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-8459c6d79f-p9s4l" podStartSLOduration=32.355723912 podStartE2EDuration="32.355723912s" podCreationTimestamp="2025-11-21 14:33:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:33:41.349929186 +0000 UTC m=+1498.634522681" watchObservedRunningTime="2025-11-21 14:33:41.355723912 +0000 UTC m=+1498.640317377" Nov 21 14:33:41 crc kubenswrapper[4897]: I1121 14:33:41.376467 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-66d9b5b984-fjrkk" podStartSLOduration=23.280434672 podStartE2EDuration="38.376442787s" podCreationTimestamp="2025-11-21 14:33:03 +0000 UTC" firstStartedPulling="2025-11-21 14:33:23.406805923 +0000 UTC m=+1480.691399398" lastFinishedPulling="2025-11-21 14:33:38.502814038 +0000 UTC m=+1495.787407513" observedRunningTime="2025-11-21 14:33:41.37016667 +0000 UTC m=+1498.654760155" watchObservedRunningTime="2025-11-21 14:33:41.376442787 +0000 UTC m=+1498.661036262" Nov 21 14:33:42 crc kubenswrapper[4897]: I1121 14:33:42.348835 4897 generic.go:334] "Generic (PLEG): container finished" podID="2ad7667f-2a3d-445b-a582-ad2bbecc23b7" containerID="f31cb8a4318a2663dea8d64af8163972acc671dfc46b55111d6f6a4c65780c72" exitCode=1 Nov 21 14:33:42 crc kubenswrapper[4897]: I1121 14:33:42.349155 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-5bc97dfdf4-wdbht" event={"ID":"2ad7667f-2a3d-445b-a582-ad2bbecc23b7","Type":"ContainerDied","Data":"f31cb8a4318a2663dea8d64af8163972acc671dfc46b55111d6f6a4c65780c72"} Nov 21 14:33:42 crc kubenswrapper[4897]: I1121 14:33:42.349995 4897 scope.go:117] "RemoveContainer" containerID="f31cb8a4318a2663dea8d64af8163972acc671dfc46b55111d6f6a4c65780c72" Nov 21 14:33:42 crc kubenswrapper[4897]: I1121 14:33:42.353846 4897 generic.go:334] "Generic (PLEG): container finished" podID="a34d9b31-8c74-47b4-bedc-0cf717bc8f3f" containerID="a011981d7a113776deb3414dc35986f7cdfec3d87d8ecb47c346409b003c0cd3" exitCode=1 Nov 21 14:33:42 crc kubenswrapper[4897]: I1121 14:33:42.353924 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-66d9b5b984-fjrkk" event={"ID":"a34d9b31-8c74-47b4-bedc-0cf717bc8f3f","Type":"ContainerDied","Data":"a011981d7a113776deb3414dc35986f7cdfec3d87d8ecb47c346409b003c0cd3"} Nov 21 14:33:42 crc kubenswrapper[4897]: I1121 14:33:42.354641 4897 scope.go:117] "RemoveContainer" containerID="a011981d7a113776deb3414dc35986f7cdfec3d87d8ecb47c346409b003c0cd3" Nov 21 14:33:42 crc kubenswrapper[4897]: I1121 14:33:42.355255 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9b4dd223-2765-4cd7-92c6-293fadac9520","Type":"ContainerStarted","Data":"45dea6b4b4b005a5c41b3ad4e7113041cad0ffea101d5022396b3fbc4e7c2bf8"} Nov 21 14:33:42 crc kubenswrapper[4897]: I1121 14:33:42.357301 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"23323788-338f-42c8-817a-e18aa170417d","Type":"ContainerStarted","Data":"0aa52499fbc13b116970dd9e694e390bff5a2bf04ac55ca5b081a2f8cc27357b"} Nov 21 14:33:42 crc kubenswrapper[4897]: I1121 14:33:42.359558 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-5d8b84cf9f-4lz8g" event={"ID":"d7cfe6c1-f519-438c-b6f1-fea0fcc70cc3","Type":"ContainerStarted","Data":"762160a7ac40ed22f4f6b9f64c7579c5e61b7da769018e0e19d8fedd5c0382ff"} Nov 21 14:33:42 crc kubenswrapper[4897]: I1121 14:33:42.360173 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-5d8b84cf9f-4lz8g" Nov 21 14:33:42 crc kubenswrapper[4897]: I1121 14:33:42.362584 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fv2n4" event={"ID":"e7b274ff-8198-43ca-a3bb-b551c97762dc","Type":"ContainerStarted","Data":"3dfe33c8352b4bc1ba944517a9f5429db25ada1f9a96f4b7dbb3e9911a14b5a2"} Nov 21 14:33:42 crc kubenswrapper[4897]: I1121 14:33:42.366034 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dtwql" event={"ID":"6d3cad20-fb17-4726-b8fa-9fd5303b5eea","Type":"ContainerStarted","Data":"684b374923ebdf191516f8820f70c1d389b071397175c74c286552f78600a36a"} Nov 21 14:33:42 crc kubenswrapper[4897]: I1121 14:33:42.408071 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=1.815028547 podStartE2EDuration="53.408008238s" podCreationTimestamp="2025-11-21 14:32:49 +0000 UTC" firstStartedPulling="2025-11-21 14:32:50.300743532 +0000 UTC m=+1447.585337007" lastFinishedPulling="2025-11-21 14:33:41.893723223 +0000 UTC m=+1499.178316698" observedRunningTime="2025-11-21 14:33:42.387378794 +0000 UTC m=+1499.671972269" watchObservedRunningTime="2025-11-21 14:33:42.408008238 +0000 UTC m=+1499.692601713" Nov 21 14:33:42 crc kubenswrapper[4897]: I1121 14:33:42.472448 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-5d8b84cf9f-4lz8g" podStartSLOduration=30.163215781 podStartE2EDuration="33.472427256s" podCreationTimestamp="2025-11-21 14:33:09 +0000 UTC" firstStartedPulling="2025-11-21 14:33:38.396622949 +0000 UTC m=+1495.681216424" lastFinishedPulling="2025-11-21 14:33:41.705834424 +0000 UTC m=+1498.990427899" observedRunningTime="2025-11-21 14:33:42.455730078 +0000 UTC m=+1499.740323543" watchObservedRunningTime="2025-11-21 14:33:42.472427256 +0000 UTC m=+1499.757020731" Nov 21 14:33:43 crc kubenswrapper[4897]: I1121 14:33:43.380360 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-66d9b5b984-fjrkk" event={"ID":"a34d9b31-8c74-47b4-bedc-0cf717bc8f3f","Type":"ContainerStarted","Data":"9e784c349a2efe265bcb29e65ad848445b6a70577b7b72146b82f2f52881ed39"} Nov 21 14:33:43 crc kubenswrapper[4897]: I1121 14:33:43.380930 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-66d9b5b984-fjrkk" Nov 21 14:33:43 crc kubenswrapper[4897]: I1121 14:33:43.384382 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-5bc97dfdf4-wdbht" event={"ID":"2ad7667f-2a3d-445b-a582-ad2bbecc23b7","Type":"ContainerStarted","Data":"ab05fe230e0ca004bd61d7d2d263e9a096650f2228713753995c51b68da6867e"} Nov 21 14:33:43 crc kubenswrapper[4897]: I1121 14:33:43.760971 4897 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/heat-api-5bc97dfdf4-wdbht" Nov 21 14:33:44 crc kubenswrapper[4897]: I1121 14:33:44.000810 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-fbfff68bc-mgc44" Nov 21 14:33:44 crc kubenswrapper[4897]: I1121 14:33:44.061329 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-75bd44869-p288j"] Nov 21 14:33:44 crc kubenswrapper[4897]: I1121 14:33:44.061618 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-engine-75bd44869-p288j" podUID="7b546c9e-dba8-4806-a2ed-8023271b448a" containerName="heat-engine" containerID="cri-o://8e41603de28669b082f8f5511f675dd2ebd4d7fd9fc98678968408620410fbb2" gracePeriod=60 Nov 21 14:33:44 crc kubenswrapper[4897]: I1121 14:33:44.397077 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-5bc97dfdf4-wdbht" Nov 21 14:33:46 crc kubenswrapper[4897]: E1121 14:33:46.160905 4897 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="8e41603de28669b082f8f5511f675dd2ebd4d7fd9fc98678968408620410fbb2" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Nov 21 14:33:46 crc kubenswrapper[4897]: E1121 14:33:46.162784 4897 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="8e41603de28669b082f8f5511f675dd2ebd4d7fd9fc98678968408620410fbb2" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Nov 21 14:33:46 crc kubenswrapper[4897]: E1121 14:33:46.164585 4897 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="8e41603de28669b082f8f5511f675dd2ebd4d7fd9fc98678968408620410fbb2" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Nov 21 14:33:46 crc kubenswrapper[4897]: E1121 14:33:46.164646 4897 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/heat-engine-75bd44869-p288j" podUID="7b546c9e-dba8-4806-a2ed-8023271b448a" containerName="heat-engine" Nov 21 14:33:46 crc kubenswrapper[4897]: I1121 14:33:46.428948 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-v4ws9"] Nov 21 14:33:46 crc kubenswrapper[4897]: E1121 14:33:46.429804 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="881842b1-c889-41be-b92a-4690f665e361" containerName="dnsmasq-dns" Nov 21 14:33:46 crc kubenswrapper[4897]: I1121 14:33:46.429820 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="881842b1-c889-41be-b92a-4690f665e361" containerName="dnsmasq-dns" Nov 21 14:33:46 crc kubenswrapper[4897]: E1121 14:33:46.429848 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="881842b1-c889-41be-b92a-4690f665e361" containerName="init" Nov 21 14:33:46 crc kubenswrapper[4897]: I1121 14:33:46.429855 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="881842b1-c889-41be-b92a-4690f665e361" containerName="init" Nov 21 14:33:46 crc kubenswrapper[4897]: I1121 14:33:46.430109 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="881842b1-c889-41be-b92a-4690f665e361" containerName="dnsmasq-dns" Nov 21 14:33:46 crc kubenswrapper[4897]: I1121 14:33:46.430935 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-v4ws9" Nov 21 14:33:46 crc kubenswrapper[4897]: I1121 14:33:46.442294 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-v4ws9"] Nov 21 14:33:46 crc kubenswrapper[4897]: I1121 14:33:46.481252 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xt8l2\" (UniqueName: \"kubernetes.io/projected/9b100fec-a276-44ac-b14d-febed2079580-kube-api-access-xt8l2\") pod \"nova-api-db-create-v4ws9\" (UID: \"9b100fec-a276-44ac-b14d-febed2079580\") " pod="openstack/nova-api-db-create-v4ws9" Nov 21 14:33:46 crc kubenswrapper[4897]: I1121 14:33:46.481296 4897 generic.go:334] "Generic (PLEG): container finished" podID="2ad7667f-2a3d-445b-a582-ad2bbecc23b7" containerID="ab05fe230e0ca004bd61d7d2d263e9a096650f2228713753995c51b68da6867e" exitCode=1 Nov 21 14:33:46 crc kubenswrapper[4897]: I1121 14:33:46.481390 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-5bc97dfdf4-wdbht" event={"ID":"2ad7667f-2a3d-445b-a582-ad2bbecc23b7","Type":"ContainerDied","Data":"ab05fe230e0ca004bd61d7d2d263e9a096650f2228713753995c51b68da6867e"} Nov 21 14:33:46 crc kubenswrapper[4897]: I1121 14:33:46.481424 4897 scope.go:117] "RemoveContainer" containerID="f31cb8a4318a2663dea8d64af8163972acc671dfc46b55111d6f6a4c65780c72" Nov 21 14:33:46 crc kubenswrapper[4897]: I1121 14:33:46.481472 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9b100fec-a276-44ac-b14d-febed2079580-operator-scripts\") pod \"nova-api-db-create-v4ws9\" (UID: \"9b100fec-a276-44ac-b14d-febed2079580\") " pod="openstack/nova-api-db-create-v4ws9" Nov 21 14:33:46 crc kubenswrapper[4897]: I1121 14:33:46.482277 4897 scope.go:117] "RemoveContainer" containerID="ab05fe230e0ca004bd61d7d2d263e9a096650f2228713753995c51b68da6867e" Nov 21 14:33:46 crc kubenswrapper[4897]: E1121 14:33:46.485681 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-5bc97dfdf4-wdbht_openstack(2ad7667f-2a3d-445b-a582-ad2bbecc23b7)\"" pod="openstack/heat-api-5bc97dfdf4-wdbht" podUID="2ad7667f-2a3d-445b-a582-ad2bbecc23b7" Nov 21 14:33:46 crc kubenswrapper[4897]: I1121 14:33:46.512523 4897 generic.go:334] "Generic (PLEG): container finished" podID="a34d9b31-8c74-47b4-bedc-0cf717bc8f3f" containerID="9e784c349a2efe265bcb29e65ad848445b6a70577b7b72146b82f2f52881ed39" exitCode=1 Nov 21 14:33:46 crc kubenswrapper[4897]: I1121 14:33:46.512768 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-66d9b5b984-fjrkk" event={"ID":"a34d9b31-8c74-47b4-bedc-0cf717bc8f3f","Type":"ContainerDied","Data":"9e784c349a2efe265bcb29e65ad848445b6a70577b7b72146b82f2f52881ed39"} Nov 21 14:33:46 crc kubenswrapper[4897]: I1121 14:33:46.513637 4897 scope.go:117] "RemoveContainer" containerID="9e784c349a2efe265bcb29e65ad848445b6a70577b7b72146b82f2f52881ed39" Nov 21 14:33:46 crc kubenswrapper[4897]: E1121 14:33:46.513968 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-66d9b5b984-fjrkk_openstack(a34d9b31-8c74-47b4-bedc-0cf717bc8f3f)\"" pod="openstack/heat-cfnapi-66d9b5b984-fjrkk" podUID="a34d9b31-8c74-47b4-bedc-0cf717bc8f3f" Nov 21 14:33:46 crc kubenswrapper[4897]: I1121 14:33:46.531388 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-vs69w"] Nov 21 14:33:46 crc kubenswrapper[4897]: I1121 14:33:46.534028 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-vs69w" Nov 21 14:33:46 crc kubenswrapper[4897]: I1121 14:33:46.547613 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-5f35-account-create-v8krn"] Nov 21 14:33:46 crc kubenswrapper[4897]: I1121 14:33:46.550616 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-5f35-account-create-v8krn" Nov 21 14:33:46 crc kubenswrapper[4897]: I1121 14:33:46.559613 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Nov 21 14:33:46 crc kubenswrapper[4897]: I1121 14:33:46.560799 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-vs69w"] Nov 21 14:33:46 crc kubenswrapper[4897]: I1121 14:33:46.619798 4897 scope.go:117] "RemoveContainer" containerID="a011981d7a113776deb3414dc35986f7cdfec3d87d8ecb47c346409b003c0cd3" Nov 21 14:33:46 crc kubenswrapper[4897]: I1121 14:33:46.632115 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xt8l2\" (UniqueName: \"kubernetes.io/projected/9b100fec-a276-44ac-b14d-febed2079580-kube-api-access-xt8l2\") pod \"nova-api-db-create-v4ws9\" (UID: \"9b100fec-a276-44ac-b14d-febed2079580\") " pod="openstack/nova-api-db-create-v4ws9" Nov 21 14:33:46 crc kubenswrapper[4897]: I1121 14:33:46.632583 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/56d6af89-a8f0-4b64-9426-04536dedf90c-operator-scripts\") pod \"nova-api-5f35-account-create-v8krn\" (UID: \"56d6af89-a8f0-4b64-9426-04536dedf90c\") " pod="openstack/nova-api-5f35-account-create-v8krn" Nov 21 14:33:46 crc kubenswrapper[4897]: I1121 14:33:46.633418 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9b100fec-a276-44ac-b14d-febed2079580-operator-scripts\") pod \"nova-api-db-create-v4ws9\" (UID: \"9b100fec-a276-44ac-b14d-febed2079580\") " pod="openstack/nova-api-db-create-v4ws9" Nov 21 14:33:46 crc kubenswrapper[4897]: I1121 14:33:46.633777 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t4nkd\" (UniqueName: \"kubernetes.io/projected/56d6af89-a8f0-4b64-9426-04536dedf90c-kube-api-access-t4nkd\") pod \"nova-api-5f35-account-create-v8krn\" (UID: \"56d6af89-a8f0-4b64-9426-04536dedf90c\") " pod="openstack/nova-api-5f35-account-create-v8krn" Nov 21 14:33:46 crc kubenswrapper[4897]: I1121 14:33:46.633880 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-krcxj\" (UniqueName: \"kubernetes.io/projected/6f516d79-7775-40cc-9d50-4167d9b1afbb-kube-api-access-krcxj\") pod \"nova-cell0-db-create-vs69w\" (UID: \"6f516d79-7775-40cc-9d50-4167d9b1afbb\") " pod="openstack/nova-cell0-db-create-vs69w" Nov 21 14:33:46 crc kubenswrapper[4897]: I1121 14:33:46.634400 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6f516d79-7775-40cc-9d50-4167d9b1afbb-operator-scripts\") pod \"nova-cell0-db-create-vs69w\" (UID: \"6f516d79-7775-40cc-9d50-4167d9b1afbb\") " pod="openstack/nova-cell0-db-create-vs69w" Nov 21 14:33:46 crc kubenswrapper[4897]: I1121 14:33:46.662617 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9b100fec-a276-44ac-b14d-febed2079580-operator-scripts\") pod \"nova-api-db-create-v4ws9\" (UID: \"9b100fec-a276-44ac-b14d-febed2079580\") " pod="openstack/nova-api-db-create-v4ws9" Nov 21 14:33:46 crc kubenswrapper[4897]: I1121 14:33:46.692720 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-5f35-account-create-v8krn"] Nov 21 14:33:46 crc kubenswrapper[4897]: I1121 14:33:46.697673 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xt8l2\" (UniqueName: \"kubernetes.io/projected/9b100fec-a276-44ac-b14d-febed2079580-kube-api-access-xt8l2\") pod \"nova-api-db-create-v4ws9\" (UID: \"9b100fec-a276-44ac-b14d-febed2079580\") " pod="openstack/nova-api-db-create-v4ws9" Nov 21 14:33:46 crc kubenswrapper[4897]: I1121 14:33:46.754153 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-7qv9b"] Nov 21 14:33:46 crc kubenswrapper[4897]: I1121 14:33:46.756013 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-7qv9b" Nov 21 14:33:46 crc kubenswrapper[4897]: I1121 14:33:46.769312 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-v4ws9" Nov 21 14:33:46 crc kubenswrapper[4897]: I1121 14:33:46.772216 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/56d6af89-a8f0-4b64-9426-04536dedf90c-operator-scripts\") pod \"nova-api-5f35-account-create-v8krn\" (UID: \"56d6af89-a8f0-4b64-9426-04536dedf90c\") " pod="openstack/nova-api-5f35-account-create-v8krn" Nov 21 14:33:46 crc kubenswrapper[4897]: I1121 14:33:46.772294 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t4nkd\" (UniqueName: \"kubernetes.io/projected/56d6af89-a8f0-4b64-9426-04536dedf90c-kube-api-access-t4nkd\") pod \"nova-api-5f35-account-create-v8krn\" (UID: \"56d6af89-a8f0-4b64-9426-04536dedf90c\") " pod="openstack/nova-api-5f35-account-create-v8krn" Nov 21 14:33:46 crc kubenswrapper[4897]: I1121 14:33:46.772323 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-krcxj\" (UniqueName: \"kubernetes.io/projected/6f516d79-7775-40cc-9d50-4167d9b1afbb-kube-api-access-krcxj\") pod \"nova-cell0-db-create-vs69w\" (UID: \"6f516d79-7775-40cc-9d50-4167d9b1afbb\") " pod="openstack/nova-cell0-db-create-vs69w" Nov 21 14:33:46 crc kubenswrapper[4897]: I1121 14:33:46.772377 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6f516d79-7775-40cc-9d50-4167d9b1afbb-operator-scripts\") pod \"nova-cell0-db-create-vs69w\" (UID: \"6f516d79-7775-40cc-9d50-4167d9b1afbb\") " pod="openstack/nova-cell0-db-create-vs69w" Nov 21 14:33:46 crc kubenswrapper[4897]: I1121 14:33:46.778075 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/56d6af89-a8f0-4b64-9426-04536dedf90c-operator-scripts\") pod \"nova-api-5f35-account-create-v8krn\" (UID: \"56d6af89-a8f0-4b64-9426-04536dedf90c\") " pod="openstack/nova-api-5f35-account-create-v8krn" Nov 21 14:33:46 crc kubenswrapper[4897]: I1121 14:33:46.784105 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6f516d79-7775-40cc-9d50-4167d9b1afbb-operator-scripts\") pod \"nova-cell0-db-create-vs69w\" (UID: \"6f516d79-7775-40cc-9d50-4167d9b1afbb\") " pod="openstack/nova-cell0-db-create-vs69w" Nov 21 14:33:46 crc kubenswrapper[4897]: I1121 14:33:46.789587 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-7qv9b"] Nov 21 14:33:46 crc kubenswrapper[4897]: I1121 14:33:46.800330 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t4nkd\" (UniqueName: \"kubernetes.io/projected/56d6af89-a8f0-4b64-9426-04536dedf90c-kube-api-access-t4nkd\") pod \"nova-api-5f35-account-create-v8krn\" (UID: \"56d6af89-a8f0-4b64-9426-04536dedf90c\") " pod="openstack/nova-api-5f35-account-create-v8krn" Nov 21 14:33:46 crc kubenswrapper[4897]: I1121 14:33:46.803104 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-krcxj\" (UniqueName: \"kubernetes.io/projected/6f516d79-7775-40cc-9d50-4167d9b1afbb-kube-api-access-krcxj\") pod \"nova-cell0-db-create-vs69w\" (UID: \"6f516d79-7775-40cc-9d50-4167d9b1afbb\") " pod="openstack/nova-cell0-db-create-vs69w" Nov 21 14:33:46 crc kubenswrapper[4897]: I1121 14:33:46.827433 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-1c5d-account-create-lx2qd"] Nov 21 14:33:46 crc kubenswrapper[4897]: I1121 14:33:46.829455 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-1c5d-account-create-lx2qd" Nov 21 14:33:46 crc kubenswrapper[4897]: I1121 14:33:46.831644 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Nov 21 14:33:46 crc kubenswrapper[4897]: I1121 14:33:46.844017 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-1c5d-account-create-lx2qd"] Nov 21 14:33:46 crc kubenswrapper[4897]: I1121 14:33:46.889422 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0d0c865b-9c9b-49cf-ba2f-730f98adce04-operator-scripts\") pod \"nova-cell1-db-create-7qv9b\" (UID: \"0d0c865b-9c9b-49cf-ba2f-730f98adce04\") " pod="openstack/nova-cell1-db-create-7qv9b" Nov 21 14:33:46 crc kubenswrapper[4897]: I1121 14:33:46.889709 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9jggk\" (UniqueName: \"kubernetes.io/projected/0d0c865b-9c9b-49cf-ba2f-730f98adce04-kube-api-access-9jggk\") pod \"nova-cell1-db-create-7qv9b\" (UID: \"0d0c865b-9c9b-49cf-ba2f-730f98adce04\") " pod="openstack/nova-cell1-db-create-7qv9b" Nov 21 14:33:46 crc kubenswrapper[4897]: I1121 14:33:46.915064 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-vs69w" Nov 21 14:33:46 crc kubenswrapper[4897]: I1121 14:33:46.940186 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-5e1d-account-create-wkbm2"] Nov 21 14:33:46 crc kubenswrapper[4897]: I1121 14:33:46.941886 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-5e1d-account-create-wkbm2" Nov 21 14:33:46 crc kubenswrapper[4897]: I1121 14:33:46.947774 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Nov 21 14:33:46 crc kubenswrapper[4897]: I1121 14:33:46.952313 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-5f35-account-create-v8krn" Nov 21 14:33:46 crc kubenswrapper[4897]: I1121 14:33:46.961287 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-5e1d-account-create-wkbm2"] Nov 21 14:33:46 crc kubenswrapper[4897]: I1121 14:33:46.992163 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0d0c865b-9c9b-49cf-ba2f-730f98adce04-operator-scripts\") pod \"nova-cell1-db-create-7qv9b\" (UID: \"0d0c865b-9c9b-49cf-ba2f-730f98adce04\") " pod="openstack/nova-cell1-db-create-7qv9b" Nov 21 14:33:46 crc kubenswrapper[4897]: I1121 14:33:46.992212 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f5440e5f-8e45-4348-ba61-e9d3e56832c6-operator-scripts\") pod \"nova-cell0-1c5d-account-create-lx2qd\" (UID: \"f5440e5f-8e45-4348-ba61-e9d3e56832c6\") " pod="openstack/nova-cell0-1c5d-account-create-lx2qd" Nov 21 14:33:46 crc kubenswrapper[4897]: I1121 14:33:46.992253 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z6mz6\" (UniqueName: \"kubernetes.io/projected/f5440e5f-8e45-4348-ba61-e9d3e56832c6-kube-api-access-z6mz6\") pod \"nova-cell0-1c5d-account-create-lx2qd\" (UID: \"f5440e5f-8e45-4348-ba61-e9d3e56832c6\") " pod="openstack/nova-cell0-1c5d-account-create-lx2qd" Nov 21 14:33:46 crc kubenswrapper[4897]: I1121 14:33:46.992349 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9jggk\" (UniqueName: \"kubernetes.io/projected/0d0c865b-9c9b-49cf-ba2f-730f98adce04-kube-api-access-9jggk\") pod \"nova-cell1-db-create-7qv9b\" (UID: \"0d0c865b-9c9b-49cf-ba2f-730f98adce04\") " pod="openstack/nova-cell1-db-create-7qv9b" Nov 21 14:33:46 crc kubenswrapper[4897]: I1121 14:33:46.993351 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0d0c865b-9c9b-49cf-ba2f-730f98adce04-operator-scripts\") pod \"nova-cell1-db-create-7qv9b\" (UID: \"0d0c865b-9c9b-49cf-ba2f-730f98adce04\") " pod="openstack/nova-cell1-db-create-7qv9b" Nov 21 14:33:47 crc kubenswrapper[4897]: I1121 14:33:47.014086 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9jggk\" (UniqueName: \"kubernetes.io/projected/0d0c865b-9c9b-49cf-ba2f-730f98adce04-kube-api-access-9jggk\") pod \"nova-cell1-db-create-7qv9b\" (UID: \"0d0c865b-9c9b-49cf-ba2f-730f98adce04\") " pod="openstack/nova-cell1-db-create-7qv9b" Nov 21 14:33:47 crc kubenswrapper[4897]: I1121 14:33:47.102779 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f5440e5f-8e45-4348-ba61-e9d3e56832c6-operator-scripts\") pod \"nova-cell0-1c5d-account-create-lx2qd\" (UID: \"f5440e5f-8e45-4348-ba61-e9d3e56832c6\") " pod="openstack/nova-cell0-1c5d-account-create-lx2qd" Nov 21 14:33:47 crc kubenswrapper[4897]: I1121 14:33:47.103098 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vw6v9\" (UniqueName: \"kubernetes.io/projected/548045ab-1444-4b7a-a663-3fa90b7fdb22-kube-api-access-vw6v9\") pod \"nova-cell1-5e1d-account-create-wkbm2\" (UID: \"548045ab-1444-4b7a-a663-3fa90b7fdb22\") " pod="openstack/nova-cell1-5e1d-account-create-wkbm2" Nov 21 14:33:47 crc kubenswrapper[4897]: I1121 14:33:47.103140 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z6mz6\" (UniqueName: \"kubernetes.io/projected/f5440e5f-8e45-4348-ba61-e9d3e56832c6-kube-api-access-z6mz6\") pod \"nova-cell0-1c5d-account-create-lx2qd\" (UID: \"f5440e5f-8e45-4348-ba61-e9d3e56832c6\") " pod="openstack/nova-cell0-1c5d-account-create-lx2qd" Nov 21 14:33:47 crc kubenswrapper[4897]: I1121 14:33:47.103205 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/548045ab-1444-4b7a-a663-3fa90b7fdb22-operator-scripts\") pod \"nova-cell1-5e1d-account-create-wkbm2\" (UID: \"548045ab-1444-4b7a-a663-3fa90b7fdb22\") " pod="openstack/nova-cell1-5e1d-account-create-wkbm2" Nov 21 14:33:47 crc kubenswrapper[4897]: I1121 14:33:47.104393 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f5440e5f-8e45-4348-ba61-e9d3e56832c6-operator-scripts\") pod \"nova-cell0-1c5d-account-create-lx2qd\" (UID: \"f5440e5f-8e45-4348-ba61-e9d3e56832c6\") " pod="openstack/nova-cell0-1c5d-account-create-lx2qd" Nov 21 14:33:47 crc kubenswrapper[4897]: I1121 14:33:47.113068 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-7qv9b" Nov 21 14:33:47 crc kubenswrapper[4897]: I1121 14:33:47.129012 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z6mz6\" (UniqueName: \"kubernetes.io/projected/f5440e5f-8e45-4348-ba61-e9d3e56832c6-kube-api-access-z6mz6\") pod \"nova-cell0-1c5d-account-create-lx2qd\" (UID: \"f5440e5f-8e45-4348-ba61-e9d3e56832c6\") " pod="openstack/nova-cell0-1c5d-account-create-lx2qd" Nov 21 14:33:47 crc kubenswrapper[4897]: I1121 14:33:47.161824 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-1c5d-account-create-lx2qd" Nov 21 14:33:47 crc kubenswrapper[4897]: I1121 14:33:47.213299 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vw6v9\" (UniqueName: \"kubernetes.io/projected/548045ab-1444-4b7a-a663-3fa90b7fdb22-kube-api-access-vw6v9\") pod \"nova-cell1-5e1d-account-create-wkbm2\" (UID: \"548045ab-1444-4b7a-a663-3fa90b7fdb22\") " pod="openstack/nova-cell1-5e1d-account-create-wkbm2" Nov 21 14:33:47 crc kubenswrapper[4897]: I1121 14:33:47.213429 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/548045ab-1444-4b7a-a663-3fa90b7fdb22-operator-scripts\") pod \"nova-cell1-5e1d-account-create-wkbm2\" (UID: \"548045ab-1444-4b7a-a663-3fa90b7fdb22\") " pod="openstack/nova-cell1-5e1d-account-create-wkbm2" Nov 21 14:33:47 crc kubenswrapper[4897]: I1121 14:33:47.224863 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/548045ab-1444-4b7a-a663-3fa90b7fdb22-operator-scripts\") pod \"nova-cell1-5e1d-account-create-wkbm2\" (UID: \"548045ab-1444-4b7a-a663-3fa90b7fdb22\") " pod="openstack/nova-cell1-5e1d-account-create-wkbm2" Nov 21 14:33:47 crc kubenswrapper[4897]: I1121 14:33:47.255324 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vw6v9\" (UniqueName: \"kubernetes.io/projected/548045ab-1444-4b7a-a663-3fa90b7fdb22-kube-api-access-vw6v9\") pod \"nova-cell1-5e1d-account-create-wkbm2\" (UID: \"548045ab-1444-4b7a-a663-3fa90b7fdb22\") " pod="openstack/nova-cell1-5e1d-account-create-wkbm2" Nov 21 14:33:47 crc kubenswrapper[4897]: I1121 14:33:47.300054 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-5e1d-account-create-wkbm2" Nov 21 14:33:47 crc kubenswrapper[4897]: I1121 14:33:47.401669 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-v4ws9"] Nov 21 14:33:47 crc kubenswrapper[4897]: W1121 14:33:47.462275 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9b100fec_a276_44ac_b14d_febed2079580.slice/crio-cb54327671f948511b3cebab1a454c2ef652423c6580fb03b9e18ec9bf16b3a0 WatchSource:0}: Error finding container cb54327671f948511b3cebab1a454c2ef652423c6580fb03b9e18ec9bf16b3a0: Status 404 returned error can't find the container with id cb54327671f948511b3cebab1a454c2ef652423c6580fb03b9e18ec9bf16b3a0 Nov 21 14:33:47 crc kubenswrapper[4897]: I1121 14:33:47.616201 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-v4ws9" event={"ID":"9b100fec-a276-44ac-b14d-febed2079580","Type":"ContainerStarted","Data":"cb54327671f948511b3cebab1a454c2ef652423c6580fb03b9e18ec9bf16b3a0"} Nov 21 14:33:47 crc kubenswrapper[4897]: I1121 14:33:47.635535 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9b4dd223-2765-4cd7-92c6-293fadac9520","Type":"ContainerStarted","Data":"361fc6578a5f944b99422cc41c214a4fce5a6bf5a72887df4dfbaf9250665d89"} Nov 21 14:33:47 crc kubenswrapper[4897]: I1121 14:33:47.744409 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-5f35-account-create-v8krn"] Nov 21 14:33:47 crc kubenswrapper[4897]: W1121 14:33:47.762630 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod56d6af89_a8f0_4b64_9426_04536dedf90c.slice/crio-b0cf47a8e9fd983e55be10cb4450b554acc2443c63343d1b8020b7e3bd8398e4 WatchSource:0}: Error finding container b0cf47a8e9fd983e55be10cb4450b554acc2443c63343d1b8020b7e3bd8398e4: Status 404 returned error can't find the container with id b0cf47a8e9fd983e55be10cb4450b554acc2443c63343d1b8020b7e3bd8398e4 Nov 21 14:33:47 crc kubenswrapper[4897]: I1121 14:33:47.809012 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-vs69w"] Nov 21 14:33:47 crc kubenswrapper[4897]: W1121 14:33:47.821623 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6f516d79_7775_40cc_9d50_4167d9b1afbb.slice/crio-8a2f7a46ca48c7b562758523cd1b759253f3e1a92515291ee9eab59bf832b419 WatchSource:0}: Error finding container 8a2f7a46ca48c7b562758523cd1b759253f3e1a92515291ee9eab59bf832b419: Status 404 returned error can't find the container with id 8a2f7a46ca48c7b562758523cd1b759253f3e1a92515291ee9eab59bf832b419 Nov 21 14:33:48 crc kubenswrapper[4897]: I1121 14:33:48.013678 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-7qv9b"] Nov 21 14:33:48 crc kubenswrapper[4897]: I1121 14:33:48.240038 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-1c5d-account-create-lx2qd"] Nov 21 14:33:48 crc kubenswrapper[4897]: I1121 14:33:48.270810 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-5e1d-account-create-wkbm2"] Nov 21 14:33:48 crc kubenswrapper[4897]: W1121 14:33:48.281562 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf5440e5f_8e45_4348_ba61_e9d3e56832c6.slice/crio-5cf355386733bfe447cc5ed3d1441a294495370e82764112f7f91ddc3f8c77a7 WatchSource:0}: Error finding container 5cf355386733bfe447cc5ed3d1441a294495370e82764112f7f91ddc3f8c77a7: Status 404 returned error can't find the container with id 5cf355386733bfe447cc5ed3d1441a294495370e82764112f7f91ddc3f8c77a7 Nov 21 14:33:48 crc kubenswrapper[4897]: I1121 14:33:48.656705 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-1c5d-account-create-lx2qd" event={"ID":"f5440e5f-8e45-4348-ba61-e9d3e56832c6","Type":"ContainerStarted","Data":"5cf355386733bfe447cc5ed3d1441a294495370e82764112f7f91ddc3f8c77a7"} Nov 21 14:33:48 crc kubenswrapper[4897]: I1121 14:33:48.665034 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-v4ws9" event={"ID":"9b100fec-a276-44ac-b14d-febed2079580","Type":"ContainerStarted","Data":"98c4e1657f5fe0d5be7978356a94be35f57ab572a6d8aaf9109027d63a7df60a"} Nov 21 14:33:48 crc kubenswrapper[4897]: I1121 14:33:48.682483 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-5f35-account-create-v8krn" event={"ID":"56d6af89-a8f0-4b64-9426-04536dedf90c","Type":"ContainerStarted","Data":"b0cf47a8e9fd983e55be10cb4450b554acc2443c63343d1b8020b7e3bd8398e4"} Nov 21 14:33:48 crc kubenswrapper[4897]: I1121 14:33:48.684103 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-5e1d-account-create-wkbm2" event={"ID":"548045ab-1444-4b7a-a663-3fa90b7fdb22","Type":"ContainerStarted","Data":"2ed927d8c9b9a85090c3266e9c202389f6b3ea278c7937306553595f6b129fd3"} Nov 21 14:33:48 crc kubenswrapper[4897]: I1121 14:33:48.687560 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-7qv9b" event={"ID":"0d0c865b-9c9b-49cf-ba2f-730f98adce04","Type":"ContainerStarted","Data":"f80422cd5de0a3925602a9066a04976357e147eecac35a6dd3587b1a447ed20f"} Nov 21 14:33:48 crc kubenswrapper[4897]: I1121 14:33:48.692154 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-vs69w" event={"ID":"6f516d79-7775-40cc-9d50-4167d9b1afbb","Type":"ContainerStarted","Data":"8a2f7a46ca48c7b562758523cd1b759253f3e1a92515291ee9eab59bf832b419"} Nov 21 14:33:48 crc kubenswrapper[4897]: I1121 14:33:48.718367 4897 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/heat-cfnapi-66d9b5b984-fjrkk" Nov 21 14:33:48 crc kubenswrapper[4897]: I1121 14:33:48.719278 4897 scope.go:117] "RemoveContainer" containerID="9e784c349a2efe265bcb29e65ad848445b6a70577b7b72146b82f2f52881ed39" Nov 21 14:33:48 crc kubenswrapper[4897]: E1121 14:33:48.719530 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-66d9b5b984-fjrkk_openstack(a34d9b31-8c74-47b4-bedc-0cf717bc8f3f)\"" pod="openstack/heat-cfnapi-66d9b5b984-fjrkk" podUID="a34d9b31-8c74-47b4-bedc-0cf717bc8f3f" Nov 21 14:33:48 crc kubenswrapper[4897]: I1121 14:33:48.761603 4897 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/heat-api-5bc97dfdf4-wdbht" Nov 21 14:33:48 crc kubenswrapper[4897]: I1121 14:33:48.762380 4897 scope.go:117] "RemoveContainer" containerID="ab05fe230e0ca004bd61d7d2d263e9a096650f2228713753995c51b68da6867e" Nov 21 14:33:48 crc kubenswrapper[4897]: E1121 14:33:48.762616 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-5bc97dfdf4-wdbht_openstack(2ad7667f-2a3d-445b-a582-ad2bbecc23b7)\"" pod="openstack/heat-api-5bc97dfdf4-wdbht" podUID="2ad7667f-2a3d-445b-a582-ad2bbecc23b7" Nov 21 14:33:49 crc kubenswrapper[4897]: I1121 14:33:49.705659 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-1c5d-account-create-lx2qd" event={"ID":"f5440e5f-8e45-4348-ba61-e9d3e56832c6","Type":"ContainerStarted","Data":"c0b618bf5ae4ce49aef42bd5f9fd1db2a9749f5add75cb36a9b75ad9708b75ca"} Nov 21 14:33:49 crc kubenswrapper[4897]: I1121 14:33:49.708633 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-5f35-account-create-v8krn" event={"ID":"56d6af89-a8f0-4b64-9426-04536dedf90c","Type":"ContainerStarted","Data":"ff653de54f2f14431bfbdca9aaceb1a64697a7c40805a7f359cb44ee5c297a09"} Nov 21 14:33:49 crc kubenswrapper[4897]: I1121 14:33:49.710989 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-5e1d-account-create-wkbm2" event={"ID":"548045ab-1444-4b7a-a663-3fa90b7fdb22","Type":"ContainerStarted","Data":"08ac180150001e919e507fa3ef316521456df000f2841292a09dafc0dc32f60f"} Nov 21 14:33:49 crc kubenswrapper[4897]: I1121 14:33:49.713691 4897 generic.go:334] "Generic (PLEG): container finished" podID="6d3cad20-fb17-4726-b8fa-9fd5303b5eea" containerID="684b374923ebdf191516f8820f70c1d389b071397175c74c286552f78600a36a" exitCode=0 Nov 21 14:33:49 crc kubenswrapper[4897]: I1121 14:33:49.713758 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dtwql" event={"ID":"6d3cad20-fb17-4726-b8fa-9fd5303b5eea","Type":"ContainerDied","Data":"684b374923ebdf191516f8820f70c1d389b071397175c74c286552f78600a36a"} Nov 21 14:33:49 crc kubenswrapper[4897]: I1121 14:33:49.715795 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-7qv9b" event={"ID":"0d0c865b-9c9b-49cf-ba2f-730f98adce04","Type":"ContainerStarted","Data":"92f1d630304a62d838c8406625e9d7c47364f3e598d34992f5a62a5858181229"} Nov 21 14:33:49 crc kubenswrapper[4897]: I1121 14:33:49.719971 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-vs69w" event={"ID":"6f516d79-7775-40cc-9d50-4167d9b1afbb","Type":"ContainerStarted","Data":"f5690c0c986bc84e69fdd7b3c70b9bd8e55ba08bd1c98561ae2c687488b4cd32"} Nov 21 14:33:49 crc kubenswrapper[4897]: I1121 14:33:49.734380 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-1c5d-account-create-lx2qd" podStartSLOduration=3.7343605650000002 podStartE2EDuration="3.734360565s" podCreationTimestamp="2025-11-21 14:33:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:33:49.721570522 +0000 UTC m=+1507.006164027" watchObservedRunningTime="2025-11-21 14:33:49.734360565 +0000 UTC m=+1507.018954060" Nov 21 14:33:49 crc kubenswrapper[4897]: I1121 14:33:49.747885 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-5e1d-account-create-wkbm2" podStartSLOduration=3.747864366 podStartE2EDuration="3.747864366s" podCreationTimestamp="2025-11-21 14:33:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:33:49.744302001 +0000 UTC m=+1507.028895476" watchObservedRunningTime="2025-11-21 14:33:49.747864366 +0000 UTC m=+1507.032457861" Nov 21 14:33:49 crc kubenswrapper[4897]: I1121 14:33:49.765568 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-db-create-7qv9b" podStartSLOduration=3.7655480409999997 podStartE2EDuration="3.765548041s" podCreationTimestamp="2025-11-21 14:33:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:33:49.763909587 +0000 UTC m=+1507.048503062" watchObservedRunningTime="2025-11-21 14:33:49.765548041 +0000 UTC m=+1507.050141516" Nov 21 14:33:49 crc kubenswrapper[4897]: I1121 14:33:49.801673 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-db-create-v4ws9" podStartSLOduration=3.801636029 podStartE2EDuration="3.801636029s" podCreationTimestamp="2025-11-21 14:33:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:33:49.799496021 +0000 UTC m=+1507.084089496" watchObservedRunningTime="2025-11-21 14:33:49.801636029 +0000 UTC m=+1507.086229514" Nov 21 14:33:49 crc kubenswrapper[4897]: I1121 14:33:49.829350 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-5f35-account-create-v8krn" podStartSLOduration=3.829325862 podStartE2EDuration="3.829325862s" podCreationTimestamp="2025-11-21 14:33:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:33:49.816987091 +0000 UTC m=+1507.101580576" watchObservedRunningTime="2025-11-21 14:33:49.829325862 +0000 UTC m=+1507.113919357" Nov 21 14:33:49 crc kubenswrapper[4897]: I1121 14:33:49.841687 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-db-create-vs69w" podStartSLOduration=3.841668982 podStartE2EDuration="3.841668982s" podCreationTimestamp="2025-11-21 14:33:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:33:49.827525233 +0000 UTC m=+1507.112118708" watchObservedRunningTime="2025-11-21 14:33:49.841668982 +0000 UTC m=+1507.126262457" Nov 21 14:33:50 crc kubenswrapper[4897]: I1121 14:33:50.892851 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/neutron-7446d77dd8-rg2bn" podUID="bac76d5f-def0-4fe8-babf-2c1a171f04dc" containerName="neutron-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Nov 21 14:33:50 crc kubenswrapper[4897]: I1121 14:33:50.892904 4897 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/neutron-7446d77dd8-rg2bn" podUID="bac76d5f-def0-4fe8-babf-2c1a171f04dc" containerName="neutron-api" probeResult="failure" output="HTTP probe failed with statuscode: 503" Nov 21 14:33:50 crc kubenswrapper[4897]: I1121 14:33:50.893379 4897 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/neutron-7446d77dd8-rg2bn" podUID="bac76d5f-def0-4fe8-babf-2c1a171f04dc" containerName="neutron-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Nov 21 14:33:51 crc kubenswrapper[4897]: I1121 14:33:51.824323 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9b4dd223-2765-4cd7-92c6-293fadac9520","Type":"ContainerStarted","Data":"84989f9793378edd4304f2cd8abc4845242c3a1d4738690d303837d51a1a273a"} Nov 21 14:33:52 crc kubenswrapper[4897]: I1121 14:33:52.839555 4897 generic.go:334] "Generic (PLEG): container finished" podID="e7b274ff-8198-43ca-a3bb-b551c97762dc" containerID="3dfe33c8352b4bc1ba944517a9f5429db25ada1f9a96f4b7dbb3e9911a14b5a2" exitCode=0 Nov 21 14:33:52 crc kubenswrapper[4897]: I1121 14:33:52.839931 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fv2n4" event={"ID":"e7b274ff-8198-43ca-a3bb-b551c97762dc","Type":"ContainerDied","Data":"3dfe33c8352b4bc1ba944517a9f5429db25ada1f9a96f4b7dbb3e9911a14b5a2"} Nov 21 14:33:52 crc kubenswrapper[4897]: I1121 14:33:52.848783 4897 generic.go:334] "Generic (PLEG): container finished" podID="0d0c865b-9c9b-49cf-ba2f-730f98adce04" containerID="92f1d630304a62d838c8406625e9d7c47364f3e598d34992f5a62a5858181229" exitCode=0 Nov 21 14:33:52 crc kubenswrapper[4897]: I1121 14:33:52.848833 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-7qv9b" event={"ID":"0d0c865b-9c9b-49cf-ba2f-730f98adce04","Type":"ContainerDied","Data":"92f1d630304a62d838c8406625e9d7c47364f3e598d34992f5a62a5858181229"} Nov 21 14:33:53 crc kubenswrapper[4897]: I1121 14:33:53.345675 4897 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/neutron-f46d49697-qwn8n" podUID="586453ed-44e4-4b53-b205-d4d325a8a702" containerName="neutron-api" probeResult="failure" output="HTTP probe failed with statuscode: 503" Nov 21 14:33:53 crc kubenswrapper[4897]: I1121 14:33:53.346310 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/neutron-f46d49697-qwn8n" podUID="586453ed-44e4-4b53-b205-d4d325a8a702" containerName="neutron-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Nov 21 14:33:53 crc kubenswrapper[4897]: I1121 14:33:53.358301 4897 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/neutron-f46d49697-qwn8n" podUID="586453ed-44e4-4b53-b205-d4d325a8a702" containerName="neutron-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Nov 21 14:33:53 crc kubenswrapper[4897]: I1121 14:33:53.863358 4897 generic.go:334] "Generic (PLEG): container finished" podID="9b100fec-a276-44ac-b14d-febed2079580" containerID="98c4e1657f5fe0d5be7978356a94be35f57ab572a6d8aaf9109027d63a7df60a" exitCode=0 Nov 21 14:33:53 crc kubenswrapper[4897]: I1121 14:33:53.863464 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-v4ws9" event={"ID":"9b100fec-a276-44ac-b14d-febed2079580","Type":"ContainerDied","Data":"98c4e1657f5fe0d5be7978356a94be35f57ab572a6d8aaf9109027d63a7df60a"} Nov 21 14:33:53 crc kubenswrapper[4897]: I1121 14:33:53.870364 4897 generic.go:334] "Generic (PLEG): container finished" podID="548045ab-1444-4b7a-a663-3fa90b7fdb22" containerID="08ac180150001e919e507fa3ef316521456df000f2841292a09dafc0dc32f60f" exitCode=0 Nov 21 14:33:53 crc kubenswrapper[4897]: I1121 14:33:53.870448 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-5e1d-account-create-wkbm2" event={"ID":"548045ab-1444-4b7a-a663-3fa90b7fdb22","Type":"ContainerDied","Data":"08ac180150001e919e507fa3ef316521456df000f2841292a09dafc0dc32f60f"} Nov 21 14:33:53 crc kubenswrapper[4897]: I1121 14:33:53.872175 4897 generic.go:334] "Generic (PLEG): container finished" podID="6f516d79-7775-40cc-9d50-4167d9b1afbb" containerID="f5690c0c986bc84e69fdd7b3c70b9bd8e55ba08bd1c98561ae2c687488b4cd32" exitCode=0 Nov 21 14:33:53 crc kubenswrapper[4897]: I1121 14:33:53.872261 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-vs69w" event={"ID":"6f516d79-7775-40cc-9d50-4167d9b1afbb","Type":"ContainerDied","Data":"f5690c0c986bc84e69fdd7b3c70b9bd8e55ba08bd1c98561ae2c687488b4cd32"} Nov 21 14:33:54 crc kubenswrapper[4897]: I1121 14:33:54.067171 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 21 14:33:54 crc kubenswrapper[4897]: I1121 14:33:54.069775 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756" containerName="glance-log" containerID="cri-o://fbffe9604199c2c6d9800e7f1823e0e982f185c924f967e3770744b01d341b13" gracePeriod=30 Nov 21 14:33:54 crc kubenswrapper[4897]: I1121 14:33:54.070355 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756" containerName="glance-httpd" containerID="cri-o://b40036e131ce501b96971cc68d11d31dd29c34c53797f7d0d45c993eaf89d3ec" gracePeriod=30 Nov 21 14:33:54 crc kubenswrapper[4897]: I1121 14:33:54.509954 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-api-5d8b84cf9f-4lz8g" Nov 21 14:33:54 crc kubenswrapper[4897]: I1121 14:33:54.580404 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-5bc97dfdf4-wdbht"] Nov 21 14:33:54 crc kubenswrapper[4897]: I1121 14:33:54.668425 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-7qv9b" Nov 21 14:33:54 crc kubenswrapper[4897]: I1121 14:33:54.726876 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-cfnapi-8459c6d79f-p9s4l" Nov 21 14:33:54 crc kubenswrapper[4897]: I1121 14:33:54.762457 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9jggk\" (UniqueName: \"kubernetes.io/projected/0d0c865b-9c9b-49cf-ba2f-730f98adce04-kube-api-access-9jggk\") pod \"0d0c865b-9c9b-49cf-ba2f-730f98adce04\" (UID: \"0d0c865b-9c9b-49cf-ba2f-730f98adce04\") " Nov 21 14:33:54 crc kubenswrapper[4897]: I1121 14:33:54.762674 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0d0c865b-9c9b-49cf-ba2f-730f98adce04-operator-scripts\") pod \"0d0c865b-9c9b-49cf-ba2f-730f98adce04\" (UID: \"0d0c865b-9c9b-49cf-ba2f-730f98adce04\") " Nov 21 14:33:54 crc kubenswrapper[4897]: I1121 14:33:54.764745 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0d0c865b-9c9b-49cf-ba2f-730f98adce04-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0d0c865b-9c9b-49cf-ba2f-730f98adce04" (UID: "0d0c865b-9c9b-49cf-ba2f-730f98adce04"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:33:54 crc kubenswrapper[4897]: I1121 14:33:54.809689 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d0c865b-9c9b-49cf-ba2f-730f98adce04-kube-api-access-9jggk" (OuterVolumeSpecName: "kube-api-access-9jggk") pod "0d0c865b-9c9b-49cf-ba2f-730f98adce04" (UID: "0d0c865b-9c9b-49cf-ba2f-730f98adce04"). InnerVolumeSpecName "kube-api-access-9jggk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:33:54 crc kubenswrapper[4897]: I1121 14:33:54.830167 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-66d9b5b984-fjrkk"] Nov 21 14:33:54 crc kubenswrapper[4897]: I1121 14:33:54.867948 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9jggk\" (UniqueName: \"kubernetes.io/projected/0d0c865b-9c9b-49cf-ba2f-730f98adce04-kube-api-access-9jggk\") on node \"crc\" DevicePath \"\"" Nov 21 14:33:54 crc kubenswrapper[4897]: I1121 14:33:54.867979 4897 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0d0c865b-9c9b-49cf-ba2f-730f98adce04-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 21 14:33:54 crc kubenswrapper[4897]: I1121 14:33:54.930702 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dtwql" event={"ID":"6d3cad20-fb17-4726-b8fa-9fd5303b5eea","Type":"ContainerStarted","Data":"36d1ba740cdac40a73ac8d07a9d49188d3f951e9e6e9797089e9005483687829"} Nov 21 14:33:54 crc kubenswrapper[4897]: I1121 14:33:54.943518 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-7qv9b" event={"ID":"0d0c865b-9c9b-49cf-ba2f-730f98adce04","Type":"ContainerDied","Data":"f80422cd5de0a3925602a9066a04976357e147eecac35a6dd3587b1a447ed20f"} Nov 21 14:33:54 crc kubenswrapper[4897]: I1121 14:33:54.944847 4897 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f80422cd5de0a3925602a9066a04976357e147eecac35a6dd3587b1a447ed20f" Nov 21 14:33:54 crc kubenswrapper[4897]: I1121 14:33:54.944980 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-7qv9b" Nov 21 14:33:54 crc kubenswrapper[4897]: I1121 14:33:54.973767 4897 generic.go:334] "Generic (PLEG): container finished" podID="f5440e5f-8e45-4348-ba61-e9d3e56832c6" containerID="c0b618bf5ae4ce49aef42bd5f9fd1db2a9749f5add75cb36a9b75ad9708b75ca" exitCode=0 Nov 21 14:33:54 crc kubenswrapper[4897]: I1121 14:33:54.973829 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-1c5d-account-create-lx2qd" event={"ID":"f5440e5f-8e45-4348-ba61-e9d3e56832c6","Type":"ContainerDied","Data":"c0b618bf5ae4ce49aef42bd5f9fd1db2a9749f5add75cb36a9b75ad9708b75ca"} Nov 21 14:33:54 crc kubenswrapper[4897]: I1121 14:33:54.975106 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-dtwql" podStartSLOduration=22.90325357 podStartE2EDuration="50.975090858s" podCreationTimestamp="2025-11-21 14:33:04 +0000 UTC" firstStartedPulling="2025-11-21 14:33:25.134159317 +0000 UTC m=+1482.418752802" lastFinishedPulling="2025-11-21 14:33:53.205996625 +0000 UTC m=+1510.490590090" observedRunningTime="2025-11-21 14:33:54.949155042 +0000 UTC m=+1512.233748517" watchObservedRunningTime="2025-11-21 14:33:54.975090858 +0000 UTC m=+1512.259684333" Nov 21 14:33:54 crc kubenswrapper[4897]: I1121 14:33:54.992537 4897 generic.go:334] "Generic (PLEG): container finished" podID="56d6af89-a8f0-4b64-9426-04536dedf90c" containerID="ff653de54f2f14431bfbdca9aaceb1a64697a7c40805a7f359cb44ee5c297a09" exitCode=0 Nov 21 14:33:54 crc kubenswrapper[4897]: I1121 14:33:54.992601 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-5f35-account-create-v8krn" event={"ID":"56d6af89-a8f0-4b64-9426-04536dedf90c","Type":"ContainerDied","Data":"ff653de54f2f14431bfbdca9aaceb1a64697a7c40805a7f359cb44ee5c297a09"} Nov 21 14:33:55 crc kubenswrapper[4897]: I1121 14:33:54.998675 4897 generic.go:334] "Generic (PLEG): container finished" podID="cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756" containerID="fbffe9604199c2c6d9800e7f1823e0e982f185c924f967e3770744b01d341b13" exitCode=143 Nov 21 14:33:55 crc kubenswrapper[4897]: I1121 14:33:54.998759 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756","Type":"ContainerDied","Data":"fbffe9604199c2c6d9800e7f1823e0e982f185c924f967e3770744b01d341b13"} Nov 21 14:33:55 crc kubenswrapper[4897]: I1121 14:33:55.010480 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fv2n4" event={"ID":"e7b274ff-8198-43ca-a3bb-b551c97762dc","Type":"ContainerStarted","Data":"2516c86ebae5b237dc067031873de6ca1e6174e5c88ce3c1b1dfe8b5d6a057f2"} Nov 21 14:33:55 crc kubenswrapper[4897]: I1121 14:33:55.066744 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-fv2n4" podStartSLOduration=23.995923523 podStartE2EDuration="53.066717695s" podCreationTimestamp="2025-11-21 14:33:02 +0000 UTC" firstStartedPulling="2025-11-21 14:33:25.125147416 +0000 UTC m=+1482.409740891" lastFinishedPulling="2025-11-21 14:33:54.195941588 +0000 UTC m=+1511.480535063" observedRunningTime="2025-11-21 14:33:55.049189846 +0000 UTC m=+1512.333783321" watchObservedRunningTime="2025-11-21 14:33:55.066717695 +0000 UTC m=+1512.351311170" Nov 21 14:33:55 crc kubenswrapper[4897]: I1121 14:33:55.113771 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-dtwql" Nov 21 14:33:55 crc kubenswrapper[4897]: I1121 14:33:55.113815 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-dtwql" Nov 21 14:33:55 crc kubenswrapper[4897]: I1121 14:33:55.690347 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-66d9b5b984-fjrkk" Nov 21 14:33:55 crc kubenswrapper[4897]: I1121 14:33:55.714980 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-5bc97dfdf4-wdbht" Nov 21 14:33:55 crc kubenswrapper[4897]: I1121 14:33:55.813360 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a34d9b31-8c74-47b4-bedc-0cf717bc8f3f-config-data\") pod \"a34d9b31-8c74-47b4-bedc-0cf717bc8f3f\" (UID: \"a34d9b31-8c74-47b4-bedc-0cf717bc8f3f\") " Nov 21 14:33:55 crc kubenswrapper[4897]: I1121 14:33:55.813814 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a34d9b31-8c74-47b4-bedc-0cf717bc8f3f-combined-ca-bundle\") pod \"a34d9b31-8c74-47b4-bedc-0cf717bc8f3f\" (UID: \"a34d9b31-8c74-47b4-bedc-0cf717bc8f3f\") " Nov 21 14:33:55 crc kubenswrapper[4897]: I1121 14:33:55.814825 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wf2fn\" (UniqueName: \"kubernetes.io/projected/a34d9b31-8c74-47b4-bedc-0cf717bc8f3f-kube-api-access-wf2fn\") pod \"a34d9b31-8c74-47b4-bedc-0cf717bc8f3f\" (UID: \"a34d9b31-8c74-47b4-bedc-0cf717bc8f3f\") " Nov 21 14:33:55 crc kubenswrapper[4897]: I1121 14:33:55.815650 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ad7667f-2a3d-445b-a582-ad2bbecc23b7-config-data\") pod \"2ad7667f-2a3d-445b-a582-ad2bbecc23b7\" (UID: \"2ad7667f-2a3d-445b-a582-ad2bbecc23b7\") " Nov 21 14:33:55 crc kubenswrapper[4897]: I1121 14:33:55.817309 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ad7667f-2a3d-445b-a582-ad2bbecc23b7-combined-ca-bundle\") pod \"2ad7667f-2a3d-445b-a582-ad2bbecc23b7\" (UID: \"2ad7667f-2a3d-445b-a582-ad2bbecc23b7\") " Nov 21 14:33:55 crc kubenswrapper[4897]: I1121 14:33:55.817873 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2ad7667f-2a3d-445b-a582-ad2bbecc23b7-config-data-custom\") pod \"2ad7667f-2a3d-445b-a582-ad2bbecc23b7\" (UID: \"2ad7667f-2a3d-445b-a582-ad2bbecc23b7\") " Nov 21 14:33:55 crc kubenswrapper[4897]: I1121 14:33:55.820329 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a34d9b31-8c74-47b4-bedc-0cf717bc8f3f-config-data-custom\") pod \"a34d9b31-8c74-47b4-bedc-0cf717bc8f3f\" (UID: \"a34d9b31-8c74-47b4-bedc-0cf717bc8f3f\") " Nov 21 14:33:55 crc kubenswrapper[4897]: I1121 14:33:55.822608 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lwb8g\" (UniqueName: \"kubernetes.io/projected/2ad7667f-2a3d-445b-a582-ad2bbecc23b7-kube-api-access-lwb8g\") pod \"2ad7667f-2a3d-445b-a582-ad2bbecc23b7\" (UID: \"2ad7667f-2a3d-445b-a582-ad2bbecc23b7\") " Nov 21 14:33:55 crc kubenswrapper[4897]: I1121 14:33:55.851268 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a34d9b31-8c74-47b4-bedc-0cf717bc8f3f-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "a34d9b31-8c74-47b4-bedc-0cf717bc8f3f" (UID: "a34d9b31-8c74-47b4-bedc-0cf717bc8f3f"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:33:55 crc kubenswrapper[4897]: I1121 14:33:55.862239 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ad7667f-2a3d-445b-a582-ad2bbecc23b7-kube-api-access-lwb8g" (OuterVolumeSpecName: "kube-api-access-lwb8g") pod "2ad7667f-2a3d-445b-a582-ad2bbecc23b7" (UID: "2ad7667f-2a3d-445b-a582-ad2bbecc23b7"). InnerVolumeSpecName "kube-api-access-lwb8g". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:33:55 crc kubenswrapper[4897]: I1121 14:33:55.869309 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lwb8g\" (UniqueName: \"kubernetes.io/projected/2ad7667f-2a3d-445b-a582-ad2bbecc23b7-kube-api-access-lwb8g\") on node \"crc\" DevicePath \"\"" Nov 21 14:33:55 crc kubenswrapper[4897]: I1121 14:33:55.869340 4897 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a34d9b31-8c74-47b4-bedc-0cf717bc8f3f-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 21 14:33:55 crc kubenswrapper[4897]: I1121 14:33:55.931741 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a34d9b31-8c74-47b4-bedc-0cf717bc8f3f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a34d9b31-8c74-47b4-bedc-0cf717bc8f3f" (UID: "a34d9b31-8c74-47b4-bedc-0cf717bc8f3f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:33:55 crc kubenswrapper[4897]: I1121 14:33:55.943951 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ad7667f-2a3d-445b-a582-ad2bbecc23b7-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "2ad7667f-2a3d-445b-a582-ad2bbecc23b7" (UID: "2ad7667f-2a3d-445b-a582-ad2bbecc23b7"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:33:55 crc kubenswrapper[4897]: I1121 14:33:55.952819 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-5e1d-account-create-wkbm2" Nov 21 14:33:55 crc kubenswrapper[4897]: I1121 14:33:55.954912 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a34d9b31-8c74-47b4-bedc-0cf717bc8f3f-kube-api-access-wf2fn" (OuterVolumeSpecName: "kube-api-access-wf2fn") pod "a34d9b31-8c74-47b4-bedc-0cf717bc8f3f" (UID: "a34d9b31-8c74-47b4-bedc-0cf717bc8f3f"). InnerVolumeSpecName "kube-api-access-wf2fn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:33:55 crc kubenswrapper[4897]: I1121 14:33:55.970422 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/548045ab-1444-4b7a-a663-3fa90b7fdb22-operator-scripts\") pod \"548045ab-1444-4b7a-a663-3fa90b7fdb22\" (UID: \"548045ab-1444-4b7a-a663-3fa90b7fdb22\") " Nov 21 14:33:55 crc kubenswrapper[4897]: I1121 14:33:55.970481 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vw6v9\" (UniqueName: \"kubernetes.io/projected/548045ab-1444-4b7a-a663-3fa90b7fdb22-kube-api-access-vw6v9\") pod \"548045ab-1444-4b7a-a663-3fa90b7fdb22\" (UID: \"548045ab-1444-4b7a-a663-3fa90b7fdb22\") " Nov 21 14:33:55 crc kubenswrapper[4897]: I1121 14:33:55.970965 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/548045ab-1444-4b7a-a663-3fa90b7fdb22-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "548045ab-1444-4b7a-a663-3fa90b7fdb22" (UID: "548045ab-1444-4b7a-a663-3fa90b7fdb22"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:33:55 crc kubenswrapper[4897]: I1121 14:33:55.971281 4897 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/548045ab-1444-4b7a-a663-3fa90b7fdb22-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 21 14:33:55 crc kubenswrapper[4897]: I1121 14:33:55.971295 4897 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a34d9b31-8c74-47b4-bedc-0cf717bc8f3f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 21 14:33:55 crc kubenswrapper[4897]: I1121 14:33:55.971303 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wf2fn\" (UniqueName: \"kubernetes.io/projected/a34d9b31-8c74-47b4-bedc-0cf717bc8f3f-kube-api-access-wf2fn\") on node \"crc\" DevicePath \"\"" Nov 21 14:33:55 crc kubenswrapper[4897]: I1121 14:33:55.971314 4897 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2ad7667f-2a3d-445b-a582-ad2bbecc23b7-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 21 14:33:55 crc kubenswrapper[4897]: I1121 14:33:55.974771 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a34d9b31-8c74-47b4-bedc-0cf717bc8f3f-config-data" (OuterVolumeSpecName: "config-data") pod "a34d9b31-8c74-47b4-bedc-0cf717bc8f3f" (UID: "a34d9b31-8c74-47b4-bedc-0cf717bc8f3f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:33:55 crc kubenswrapper[4897]: I1121 14:33:55.978009 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/548045ab-1444-4b7a-a663-3fa90b7fdb22-kube-api-access-vw6v9" (OuterVolumeSpecName: "kube-api-access-vw6v9") pod "548045ab-1444-4b7a-a663-3fa90b7fdb22" (UID: "548045ab-1444-4b7a-a663-3fa90b7fdb22"). InnerVolumeSpecName "kube-api-access-vw6v9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:33:56 crc kubenswrapper[4897]: I1121 14:33:55.997563 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ad7667f-2a3d-445b-a582-ad2bbecc23b7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2ad7667f-2a3d-445b-a582-ad2bbecc23b7" (UID: "2ad7667f-2a3d-445b-a582-ad2bbecc23b7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:33:56 crc kubenswrapper[4897]: I1121 14:33:56.045107 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ad7667f-2a3d-445b-a582-ad2bbecc23b7-config-data" (OuterVolumeSpecName: "config-data") pod "2ad7667f-2a3d-445b-a582-ad2bbecc23b7" (UID: "2ad7667f-2a3d-445b-a582-ad2bbecc23b7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:33:56 crc kubenswrapper[4897]: I1121 14:33:56.046581 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-5e1d-account-create-wkbm2" event={"ID":"548045ab-1444-4b7a-a663-3fa90b7fdb22","Type":"ContainerDied","Data":"2ed927d8c9b9a85090c3266e9c202389f6b3ea278c7937306553595f6b129fd3"} Nov 21 14:33:56 crc kubenswrapper[4897]: I1121 14:33:56.046606 4897 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2ed927d8c9b9a85090c3266e9c202389f6b3ea278c7937306553595f6b129fd3" Nov 21 14:33:56 crc kubenswrapper[4897]: I1121 14:33:56.046611 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-5e1d-account-create-wkbm2" Nov 21 14:33:56 crc kubenswrapper[4897]: I1121 14:33:56.047890 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-66d9b5b984-fjrkk" event={"ID":"a34d9b31-8c74-47b4-bedc-0cf717bc8f3f","Type":"ContainerDied","Data":"7f09c45f9512b71f97d0a96c52fcc0c293c49fa290a657bf1d5ced766942e394"} Nov 21 14:33:56 crc kubenswrapper[4897]: I1121 14:33:56.047915 4897 scope.go:117] "RemoveContainer" containerID="9e784c349a2efe265bcb29e65ad848445b6a70577b7b72146b82f2f52881ed39" Nov 21 14:33:56 crc kubenswrapper[4897]: I1121 14:33:56.047985 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-66d9b5b984-fjrkk" Nov 21 14:33:56 crc kubenswrapper[4897]: I1121 14:33:56.052982 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-5bc97dfdf4-wdbht" Nov 21 14:33:56 crc kubenswrapper[4897]: I1121 14:33:56.055351 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-5bc97dfdf4-wdbht" event={"ID":"2ad7667f-2a3d-445b-a582-ad2bbecc23b7","Type":"ContainerDied","Data":"dce41364e86775aefa9e6523058e1b8a53ad3ab53e5ec385b443645159d74db8"} Nov 21 14:33:56 crc kubenswrapper[4897]: I1121 14:33:56.089069 4897 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ad7667f-2a3d-445b-a582-ad2bbecc23b7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 21 14:33:56 crc kubenswrapper[4897]: I1121 14:33:56.089122 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vw6v9\" (UniqueName: \"kubernetes.io/projected/548045ab-1444-4b7a-a663-3fa90b7fdb22-kube-api-access-vw6v9\") on node \"crc\" DevicePath \"\"" Nov 21 14:33:56 crc kubenswrapper[4897]: I1121 14:33:56.089171 4897 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a34d9b31-8c74-47b4-bedc-0cf717bc8f3f-config-data\") on node \"crc\" DevicePath \"\"" Nov 21 14:33:56 crc kubenswrapper[4897]: I1121 14:33:56.089186 4897 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ad7667f-2a3d-445b-a582-ad2bbecc23b7-config-data\") on node \"crc\" DevicePath \"\"" Nov 21 14:33:56 crc kubenswrapper[4897]: I1121 14:33:56.121474 4897 trace.go:236] Trace[1805090524]: "Calculate volume metrics of catalog-content for pod openshift-marketplace/certified-operators-s8gkj" (21-Nov-2025 14:33:55.022) (total time: 1098ms): Nov 21 14:33:56 crc kubenswrapper[4897]: Trace[1805090524]: [1.098626768s] [1.098626768s] END Nov 21 14:33:56 crc kubenswrapper[4897]: E1121 14:33:56.166697 4897 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="8e41603de28669b082f8f5511f675dd2ebd4d7fd9fc98678968408620410fbb2" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Nov 21 14:33:56 crc kubenswrapper[4897]: E1121 14:33:56.188642 4897 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="8e41603de28669b082f8f5511f675dd2ebd4d7fd9fc98678968408620410fbb2" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Nov 21 14:33:56 crc kubenswrapper[4897]: I1121 14:33:56.203149 4897 scope.go:117] "RemoveContainer" containerID="ab05fe230e0ca004bd61d7d2d263e9a096650f2228713753995c51b68da6867e" Nov 21 14:33:56 crc kubenswrapper[4897]: E1121 14:33:56.203261 4897 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="8e41603de28669b082f8f5511f675dd2ebd4d7fd9fc98678968408620410fbb2" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Nov 21 14:33:56 crc kubenswrapper[4897]: E1121 14:33:56.203298 4897 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/heat-engine-75bd44869-p288j" podUID="7b546c9e-dba8-4806-a2ed-8023271b448a" containerName="heat-engine" Nov 21 14:33:56 crc kubenswrapper[4897]: I1121 14:33:56.529837 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-dtwql" podUID="6d3cad20-fb17-4726-b8fa-9fd5303b5eea" containerName="registry-server" probeResult="failure" output=< Nov 21 14:33:56 crc kubenswrapper[4897]: timeout: failed to connect service ":50051" within 1s Nov 21 14:33:56 crc kubenswrapper[4897]: > Nov 21 14:33:57 crc kubenswrapper[4897]: I1121 14:33:57.347023 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-internal-api-0" podUID="cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756" containerName="glance-log" probeResult="failure" output="Get \"https://10.217.0.195:9292/healthcheck\": read tcp 10.217.0.2:45694->10.217.0.195:9292: read: connection reset by peer" Nov 21 14:33:57 crc kubenswrapper[4897]: I1121 14:33:57.347065 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-internal-api-0" podUID="cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756" containerName="glance-httpd" probeResult="failure" output="Get \"https://10.217.0.195:9292/healthcheck\": read tcp 10.217.0.2:45692->10.217.0.195:9292: read: connection reset by peer" Nov 21 14:33:57 crc kubenswrapper[4897]: E1121 14:33:57.476348 4897 kubelet.go:2526] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="1.388s" Nov 21 14:33:57 crc kubenswrapper[4897]: I1121 14:33:57.476443 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-vs69w" event={"ID":"6f516d79-7775-40cc-9d50-4167d9b1afbb","Type":"ContainerDied","Data":"8a2f7a46ca48c7b562758523cd1b759253f3e1a92515291ee9eab59bf832b419"} Nov 21 14:33:57 crc kubenswrapper[4897]: I1121 14:33:57.476468 4897 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8a2f7a46ca48c7b562758523cd1b759253f3e1a92515291ee9eab59bf832b419" Nov 21 14:33:57 crc kubenswrapper[4897]: I1121 14:33:57.476478 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-v4ws9" event={"ID":"9b100fec-a276-44ac-b14d-febed2079580","Type":"ContainerDied","Data":"cb54327671f948511b3cebab1a454c2ef652423c6580fb03b9e18ec9bf16b3a0"} Nov 21 14:33:57 crc kubenswrapper[4897]: I1121 14:33:57.476489 4897 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cb54327671f948511b3cebab1a454c2ef652423c6580fb03b9e18ec9bf16b3a0" Nov 21 14:33:57 crc kubenswrapper[4897]: I1121 14:33:57.671332 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 21 14:33:57 crc kubenswrapper[4897]: I1121 14:33:57.672160 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="7155191e-f228-4639-8a75-ed8ff3dce81a" containerName="glance-log" containerID="cri-o://e3f7f26e10be327c1a46809d507483a15eb0aca60ab79afe47a1fa3b5e10885f" gracePeriod=30 Nov 21 14:33:57 crc kubenswrapper[4897]: I1121 14:33:57.672445 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="7155191e-f228-4639-8a75-ed8ff3dce81a" containerName="glance-httpd" containerID="cri-o://dc5724ef2786b81a902bc7373794e6066c91ce62272f7e949534d19dee9f3fe0" gracePeriod=30 Nov 21 14:33:57 crc kubenswrapper[4897]: I1121 14:33:57.744909 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-v4ws9" Nov 21 14:33:57 crc kubenswrapper[4897]: I1121 14:33:57.758263 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-5bc97dfdf4-wdbht"] Nov 21 14:33:57 crc kubenswrapper[4897]: I1121 14:33:57.768270 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-vs69w" Nov 21 14:33:57 crc kubenswrapper[4897]: I1121 14:33:57.774867 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-1c5d-account-create-lx2qd" Nov 21 14:33:57 crc kubenswrapper[4897]: I1121 14:33:57.811567 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-api-5bc97dfdf4-wdbht"] Nov 21 14:33:57 crc kubenswrapper[4897]: I1121 14:33:57.830669 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-66d9b5b984-fjrkk"] Nov 21 14:33:57 crc kubenswrapper[4897]: I1121 14:33:57.855575 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-cfnapi-66d9b5b984-fjrkk"] Nov 21 14:33:57 crc kubenswrapper[4897]: I1121 14:33:57.877607 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9b100fec-a276-44ac-b14d-febed2079580-operator-scripts\") pod \"9b100fec-a276-44ac-b14d-febed2079580\" (UID: \"9b100fec-a276-44ac-b14d-febed2079580\") " Nov 21 14:33:57 crc kubenswrapper[4897]: I1121 14:33:57.877741 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z6mz6\" (UniqueName: \"kubernetes.io/projected/f5440e5f-8e45-4348-ba61-e9d3e56832c6-kube-api-access-z6mz6\") pod \"f5440e5f-8e45-4348-ba61-e9d3e56832c6\" (UID: \"f5440e5f-8e45-4348-ba61-e9d3e56832c6\") " Nov 21 14:33:57 crc kubenswrapper[4897]: I1121 14:33:57.877818 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xt8l2\" (UniqueName: \"kubernetes.io/projected/9b100fec-a276-44ac-b14d-febed2079580-kube-api-access-xt8l2\") pod \"9b100fec-a276-44ac-b14d-febed2079580\" (UID: \"9b100fec-a276-44ac-b14d-febed2079580\") " Nov 21 14:33:57 crc kubenswrapper[4897]: I1121 14:33:57.877862 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-krcxj\" (UniqueName: \"kubernetes.io/projected/6f516d79-7775-40cc-9d50-4167d9b1afbb-kube-api-access-krcxj\") pod \"6f516d79-7775-40cc-9d50-4167d9b1afbb\" (UID: \"6f516d79-7775-40cc-9d50-4167d9b1afbb\") " Nov 21 14:33:57 crc kubenswrapper[4897]: I1121 14:33:57.877916 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f5440e5f-8e45-4348-ba61-e9d3e56832c6-operator-scripts\") pod \"f5440e5f-8e45-4348-ba61-e9d3e56832c6\" (UID: \"f5440e5f-8e45-4348-ba61-e9d3e56832c6\") " Nov 21 14:33:57 crc kubenswrapper[4897]: I1121 14:33:57.877960 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6f516d79-7775-40cc-9d50-4167d9b1afbb-operator-scripts\") pod \"6f516d79-7775-40cc-9d50-4167d9b1afbb\" (UID: \"6f516d79-7775-40cc-9d50-4167d9b1afbb\") " Nov 21 14:33:57 crc kubenswrapper[4897]: I1121 14:33:57.879756 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9b100fec-a276-44ac-b14d-febed2079580-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9b100fec-a276-44ac-b14d-febed2079580" (UID: "9b100fec-a276-44ac-b14d-febed2079580"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:33:57 crc kubenswrapper[4897]: I1121 14:33:57.882610 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6f516d79-7775-40cc-9d50-4167d9b1afbb-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "6f516d79-7775-40cc-9d50-4167d9b1afbb" (UID: "6f516d79-7775-40cc-9d50-4167d9b1afbb"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:33:57 crc kubenswrapper[4897]: I1121 14:33:57.883382 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f5440e5f-8e45-4348-ba61-e9d3e56832c6-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f5440e5f-8e45-4348-ba61-e9d3e56832c6" (UID: "f5440e5f-8e45-4348-ba61-e9d3e56832c6"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:33:57 crc kubenswrapper[4897]: I1121 14:33:57.885280 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-5f35-account-create-v8krn" Nov 21 14:33:57 crc kubenswrapper[4897]: I1121 14:33:57.888937 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5440e5f-8e45-4348-ba61-e9d3e56832c6-kube-api-access-z6mz6" (OuterVolumeSpecName: "kube-api-access-z6mz6") pod "f5440e5f-8e45-4348-ba61-e9d3e56832c6" (UID: "f5440e5f-8e45-4348-ba61-e9d3e56832c6"). InnerVolumeSpecName "kube-api-access-z6mz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:33:57 crc kubenswrapper[4897]: I1121 14:33:57.899020 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9b100fec-a276-44ac-b14d-febed2079580-kube-api-access-xt8l2" (OuterVolumeSpecName: "kube-api-access-xt8l2") pod "9b100fec-a276-44ac-b14d-febed2079580" (UID: "9b100fec-a276-44ac-b14d-febed2079580"). InnerVolumeSpecName "kube-api-access-xt8l2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:33:57 crc kubenswrapper[4897]: I1121 14:33:57.902132 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6f516d79-7775-40cc-9d50-4167d9b1afbb-kube-api-access-krcxj" (OuterVolumeSpecName: "kube-api-access-krcxj") pod "6f516d79-7775-40cc-9d50-4167d9b1afbb" (UID: "6f516d79-7775-40cc-9d50-4167d9b1afbb"). InnerVolumeSpecName "kube-api-access-krcxj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:33:57 crc kubenswrapper[4897]: I1121 14:33:57.995027 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/56d6af89-a8f0-4b64-9426-04536dedf90c-operator-scripts\") pod \"56d6af89-a8f0-4b64-9426-04536dedf90c\" (UID: \"56d6af89-a8f0-4b64-9426-04536dedf90c\") " Nov 21 14:33:57 crc kubenswrapper[4897]: I1121 14:33:57.995411 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t4nkd\" (UniqueName: \"kubernetes.io/projected/56d6af89-a8f0-4b64-9426-04536dedf90c-kube-api-access-t4nkd\") pod \"56d6af89-a8f0-4b64-9426-04536dedf90c\" (UID: \"56d6af89-a8f0-4b64-9426-04536dedf90c\") " Nov 21 14:33:57 crc kubenswrapper[4897]: I1121 14:33:57.997821 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/56d6af89-a8f0-4b64-9426-04536dedf90c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "56d6af89-a8f0-4b64-9426-04536dedf90c" (UID: "56d6af89-a8f0-4b64-9426-04536dedf90c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:33:58 crc kubenswrapper[4897]: I1121 14:33:58.015913 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56d6af89-a8f0-4b64-9426-04536dedf90c-kube-api-access-t4nkd" (OuterVolumeSpecName: "kube-api-access-t4nkd") pod "56d6af89-a8f0-4b64-9426-04536dedf90c" (UID: "56d6af89-a8f0-4b64-9426-04536dedf90c"). InnerVolumeSpecName "kube-api-access-t4nkd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:33:58 crc kubenswrapper[4897]: I1121 14:33:58.025079 4897 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f5440e5f-8e45-4348-ba61-e9d3e56832c6-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 21 14:33:58 crc kubenswrapper[4897]: I1121 14:33:58.025106 4897 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6f516d79-7775-40cc-9d50-4167d9b1afbb-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 21 14:33:58 crc kubenswrapper[4897]: I1121 14:33:58.025116 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t4nkd\" (UniqueName: \"kubernetes.io/projected/56d6af89-a8f0-4b64-9426-04536dedf90c-kube-api-access-t4nkd\") on node \"crc\" DevicePath \"\"" Nov 21 14:33:58 crc kubenswrapper[4897]: I1121 14:33:58.025138 4897 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/56d6af89-a8f0-4b64-9426-04536dedf90c-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 21 14:33:58 crc kubenswrapper[4897]: I1121 14:33:58.025148 4897 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9b100fec-a276-44ac-b14d-febed2079580-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 21 14:33:58 crc kubenswrapper[4897]: I1121 14:33:58.025158 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z6mz6\" (UniqueName: \"kubernetes.io/projected/f5440e5f-8e45-4348-ba61-e9d3e56832c6-kube-api-access-z6mz6\") on node \"crc\" DevicePath \"\"" Nov 21 14:33:58 crc kubenswrapper[4897]: I1121 14:33:58.025167 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xt8l2\" (UniqueName: \"kubernetes.io/projected/9b100fec-a276-44ac-b14d-febed2079580-kube-api-access-xt8l2\") on node \"crc\" DevicePath \"\"" Nov 21 14:33:58 crc kubenswrapper[4897]: I1121 14:33:58.025175 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-krcxj\" (UniqueName: \"kubernetes.io/projected/6f516d79-7775-40cc-9d50-4167d9b1afbb-kube-api-access-krcxj\") on node \"crc\" DevicePath \"\"" Nov 21 14:33:58 crc kubenswrapper[4897]: I1121 14:33:58.150032 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2ad7667f-2a3d-445b-a582-ad2bbecc23b7" path="/var/lib/kubelet/pods/2ad7667f-2a3d-445b-a582-ad2bbecc23b7/volumes" Nov 21 14:33:58 crc kubenswrapper[4897]: I1121 14:33:58.151068 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a34d9b31-8c74-47b4-bedc-0cf717bc8f3f" path="/var/lib/kubelet/pods/a34d9b31-8c74-47b4-bedc-0cf717bc8f3f/volumes" Nov 21 14:33:58 crc kubenswrapper[4897]: I1121 14:33:58.223728 4897 generic.go:334] "Generic (PLEG): container finished" podID="cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756" containerID="b40036e131ce501b96971cc68d11d31dd29c34c53797f7d0d45c993eaf89d3ec" exitCode=0 Nov 21 14:33:58 crc kubenswrapper[4897]: I1121 14:33:58.224236 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756","Type":"ContainerDied","Data":"b40036e131ce501b96971cc68d11d31dd29c34c53797f7d0d45c993eaf89d3ec"} Nov 21 14:33:58 crc kubenswrapper[4897]: I1121 14:33:58.281696 4897 generic.go:334] "Generic (PLEG): container finished" podID="7b546c9e-dba8-4806-a2ed-8023271b448a" containerID="8e41603de28669b082f8f5511f675dd2ebd4d7fd9fc98678968408620410fbb2" exitCode=0 Nov 21 14:33:58 crc kubenswrapper[4897]: I1121 14:33:58.281845 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-75bd44869-p288j" event={"ID":"7b546c9e-dba8-4806-a2ed-8023271b448a","Type":"ContainerDied","Data":"8e41603de28669b082f8f5511f675dd2ebd4d7fd9fc98678968408620410fbb2"} Nov 21 14:33:58 crc kubenswrapper[4897]: I1121 14:33:58.312058 4897 generic.go:334] "Generic (PLEG): container finished" podID="7155191e-f228-4639-8a75-ed8ff3dce81a" containerID="e3f7f26e10be327c1a46809d507483a15eb0aca60ab79afe47a1fa3b5e10885f" exitCode=143 Nov 21 14:33:58 crc kubenswrapper[4897]: I1121 14:33:58.312168 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"7155191e-f228-4639-8a75-ed8ff3dce81a","Type":"ContainerDied","Data":"e3f7f26e10be327c1a46809d507483a15eb0aca60ab79afe47a1fa3b5e10885f"} Nov 21 14:33:58 crc kubenswrapper[4897]: I1121 14:33:58.325735 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-1c5d-account-create-lx2qd" event={"ID":"f5440e5f-8e45-4348-ba61-e9d3e56832c6","Type":"ContainerDied","Data":"5cf355386733bfe447cc5ed3d1441a294495370e82764112f7f91ddc3f8c77a7"} Nov 21 14:33:58 crc kubenswrapper[4897]: I1121 14:33:58.325775 4897 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5cf355386733bfe447cc5ed3d1441a294495370e82764112f7f91ddc3f8c77a7" Nov 21 14:33:58 crc kubenswrapper[4897]: I1121 14:33:58.325838 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-1c5d-account-create-lx2qd" Nov 21 14:33:58 crc kubenswrapper[4897]: I1121 14:33:58.360260 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9b4dd223-2765-4cd7-92c6-293fadac9520","Type":"ContainerStarted","Data":"2c786aefd013c99dc5c1d324fa90173506b4fc5af2323d0c2bdbc16bc6ad25d1"} Nov 21 14:33:58 crc kubenswrapper[4897]: I1121 14:33:58.360951 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9b4dd223-2765-4cd7-92c6-293fadac9520" containerName="ceilometer-central-agent" containerID="cri-o://45dea6b4b4b005a5c41b3ad4e7113041cad0ffea101d5022396b3fbc4e7c2bf8" gracePeriod=30 Nov 21 14:33:58 crc kubenswrapper[4897]: I1121 14:33:58.361234 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 21 14:33:58 crc kubenswrapper[4897]: I1121 14:33:58.361397 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9b4dd223-2765-4cd7-92c6-293fadac9520" containerName="proxy-httpd" containerID="cri-o://2c786aefd013c99dc5c1d324fa90173506b4fc5af2323d0c2bdbc16bc6ad25d1" gracePeriod=30 Nov 21 14:33:58 crc kubenswrapper[4897]: I1121 14:33:58.361466 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9b4dd223-2765-4cd7-92c6-293fadac9520" containerName="sg-core" containerID="cri-o://84989f9793378edd4304f2cd8abc4845242c3a1d4738690d303837d51a1a273a" gracePeriod=30 Nov 21 14:33:58 crc kubenswrapper[4897]: I1121 14:33:58.361535 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9b4dd223-2765-4cd7-92c6-293fadac9520" containerName="ceilometer-notification-agent" containerID="cri-o://361fc6578a5f944b99422cc41c214a4fce5a6bf5a72887df4dfbaf9250665d89" gracePeriod=30 Nov 21 14:33:58 crc kubenswrapper[4897]: I1121 14:33:58.387628 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-5f35-account-create-v8krn" event={"ID":"56d6af89-a8f0-4b64-9426-04536dedf90c","Type":"ContainerDied","Data":"b0cf47a8e9fd983e55be10cb4450b554acc2443c63343d1b8020b7e3bd8398e4"} Nov 21 14:33:58 crc kubenswrapper[4897]: I1121 14:33:58.387703 4897 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b0cf47a8e9fd983e55be10cb4450b554acc2443c63343d1b8020b7e3bd8398e4" Nov 21 14:33:58 crc kubenswrapper[4897]: I1121 14:33:58.387805 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-v4ws9" Nov 21 14:33:58 crc kubenswrapper[4897]: I1121 14:33:58.387888 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-5f35-account-create-v8krn" Nov 21 14:33:58 crc kubenswrapper[4897]: I1121 14:33:58.393738 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-vs69w" Nov 21 14:33:58 crc kubenswrapper[4897]: I1121 14:33:58.397202 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 21 14:33:58 crc kubenswrapper[4897]: I1121 14:33:58.400393 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=16.692025109 podStartE2EDuration="34.400379305s" podCreationTimestamp="2025-11-21 14:33:24 +0000 UTC" firstStartedPulling="2025-11-21 14:33:38.494551176 +0000 UTC m=+1495.779144651" lastFinishedPulling="2025-11-21 14:33:56.202905372 +0000 UTC m=+1513.487498847" observedRunningTime="2025-11-21 14:33:58.38788346 +0000 UTC m=+1515.672476945" watchObservedRunningTime="2025-11-21 14:33:58.400379305 +0000 UTC m=+1515.684972780" Nov 21 14:33:58 crc kubenswrapper[4897]: I1121 14:33:58.472099 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756-scripts\") pod \"cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756\" (UID: \"cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756\") " Nov 21 14:33:58 crc kubenswrapper[4897]: I1121 14:33:58.472131 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756-combined-ca-bundle\") pod \"cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756\" (UID: \"cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756\") " Nov 21 14:33:58 crc kubenswrapper[4897]: I1121 14:33:58.472228 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756-logs\") pod \"cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756\" (UID: \"cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756\") " Nov 21 14:33:58 crc kubenswrapper[4897]: I1121 14:33:58.472307 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756-config-data\") pod \"cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756\" (UID: \"cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756\") " Nov 21 14:33:58 crc kubenswrapper[4897]: I1121 14:33:58.472381 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756\" (UID: \"cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756\") " Nov 21 14:33:58 crc kubenswrapper[4897]: I1121 14:33:58.472415 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ct675\" (UniqueName: \"kubernetes.io/projected/cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756-kube-api-access-ct675\") pod \"cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756\" (UID: \"cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756\") " Nov 21 14:33:58 crc kubenswrapper[4897]: I1121 14:33:58.472475 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756-internal-tls-certs\") pod \"cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756\" (UID: \"cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756\") " Nov 21 14:33:58 crc kubenswrapper[4897]: I1121 14:33:58.472552 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756-httpd-run\") pod \"cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756\" (UID: \"cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756\") " Nov 21 14:33:58 crc kubenswrapper[4897]: I1121 14:33:58.473618 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756" (UID: "cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:33:58 crc kubenswrapper[4897]: I1121 14:33:58.479873 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756-logs" (OuterVolumeSpecName: "logs") pod "cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756" (UID: "cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:33:58 crc kubenswrapper[4897]: I1121 14:33:58.487463 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756-kube-api-access-ct675" (OuterVolumeSpecName: "kube-api-access-ct675") pod "cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756" (UID: "cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756"). InnerVolumeSpecName "kube-api-access-ct675". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:33:58 crc kubenswrapper[4897]: I1121 14:33:58.497832 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756-scripts" (OuterVolumeSpecName: "scripts") pod "cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756" (UID: "cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:33:58 crc kubenswrapper[4897]: I1121 14:33:58.515104 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "glance") pod "cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756" (UID: "cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 21 14:33:58 crc kubenswrapper[4897]: I1121 14:33:58.568013 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756" (UID: "cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:33:58 crc kubenswrapper[4897]: I1121 14:33:58.576211 4897 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Nov 21 14:33:58 crc kubenswrapper[4897]: I1121 14:33:58.576250 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ct675\" (UniqueName: \"kubernetes.io/projected/cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756-kube-api-access-ct675\") on node \"crc\" DevicePath \"\"" Nov 21 14:33:58 crc kubenswrapper[4897]: I1121 14:33:58.576262 4897 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 21 14:33:58 crc kubenswrapper[4897]: I1121 14:33:58.576270 4897 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756-scripts\") on node \"crc\" DevicePath \"\"" Nov 21 14:33:58 crc kubenswrapper[4897]: I1121 14:33:58.576279 4897 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 21 14:33:58 crc kubenswrapper[4897]: I1121 14:33:58.576288 4897 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756-logs\") on node \"crc\" DevicePath \"\"" Nov 21 14:33:58 crc kubenswrapper[4897]: I1121 14:33:58.649102 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756" (UID: "cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:33:58 crc kubenswrapper[4897]: I1121 14:33:58.655783 4897 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Nov 21 14:33:58 crc kubenswrapper[4897]: I1121 14:33:58.657445 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-75bd44869-p288j" Nov 21 14:33:58 crc kubenswrapper[4897]: I1121 14:33:58.684561 4897 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Nov 21 14:33:58 crc kubenswrapper[4897]: I1121 14:33:58.685311 4897 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 21 14:33:58 crc kubenswrapper[4897]: I1121 14:33:58.758217 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756-config-data" (OuterVolumeSpecName: "config-data") pod "cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756" (UID: "cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:33:58 crc kubenswrapper[4897]: I1121 14:33:58.786561 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7b546c9e-dba8-4806-a2ed-8023271b448a-config-data-custom\") pod \"7b546c9e-dba8-4806-a2ed-8023271b448a\" (UID: \"7b546c9e-dba8-4806-a2ed-8023271b448a\") " Nov 21 14:33:58 crc kubenswrapper[4897]: I1121 14:33:58.786745 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b546c9e-dba8-4806-a2ed-8023271b448a-config-data\") pod \"7b546c9e-dba8-4806-a2ed-8023271b448a\" (UID: \"7b546c9e-dba8-4806-a2ed-8023271b448a\") " Nov 21 14:33:58 crc kubenswrapper[4897]: I1121 14:33:58.786812 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hc55h\" (UniqueName: \"kubernetes.io/projected/7b546c9e-dba8-4806-a2ed-8023271b448a-kube-api-access-hc55h\") pod \"7b546c9e-dba8-4806-a2ed-8023271b448a\" (UID: \"7b546c9e-dba8-4806-a2ed-8023271b448a\") " Nov 21 14:33:58 crc kubenswrapper[4897]: I1121 14:33:58.786942 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b546c9e-dba8-4806-a2ed-8023271b448a-combined-ca-bundle\") pod \"7b546c9e-dba8-4806-a2ed-8023271b448a\" (UID: \"7b546c9e-dba8-4806-a2ed-8023271b448a\") " Nov 21 14:33:58 crc kubenswrapper[4897]: I1121 14:33:58.788228 4897 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756-config-data\") on node \"crc\" DevicePath \"\"" Nov 21 14:33:58 crc kubenswrapper[4897]: I1121 14:33:58.810914 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b546c9e-dba8-4806-a2ed-8023271b448a-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "7b546c9e-dba8-4806-a2ed-8023271b448a" (UID: "7b546c9e-dba8-4806-a2ed-8023271b448a"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:33:58 crc kubenswrapper[4897]: I1121 14:33:58.842982 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b546c9e-dba8-4806-a2ed-8023271b448a-kube-api-access-hc55h" (OuterVolumeSpecName: "kube-api-access-hc55h") pod "7b546c9e-dba8-4806-a2ed-8023271b448a" (UID: "7b546c9e-dba8-4806-a2ed-8023271b448a"). InnerVolumeSpecName "kube-api-access-hc55h". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:33:58 crc kubenswrapper[4897]: I1121 14:33:58.850034 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b546c9e-dba8-4806-a2ed-8023271b448a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7b546c9e-dba8-4806-a2ed-8023271b448a" (UID: "7b546c9e-dba8-4806-a2ed-8023271b448a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:33:58 crc kubenswrapper[4897]: I1121 14:33:58.890437 4897 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7b546c9e-dba8-4806-a2ed-8023271b448a-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 21 14:33:58 crc kubenswrapper[4897]: I1121 14:33:58.890468 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hc55h\" (UniqueName: \"kubernetes.io/projected/7b546c9e-dba8-4806-a2ed-8023271b448a-kube-api-access-hc55h\") on node \"crc\" DevicePath \"\"" Nov 21 14:33:58 crc kubenswrapper[4897]: I1121 14:33:58.890479 4897 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b546c9e-dba8-4806-a2ed-8023271b448a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 21 14:33:58 crc kubenswrapper[4897]: I1121 14:33:58.906722 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b546c9e-dba8-4806-a2ed-8023271b448a-config-data" (OuterVolumeSpecName: "config-data") pod "7b546c9e-dba8-4806-a2ed-8023271b448a" (UID: "7b546c9e-dba8-4806-a2ed-8023271b448a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:33:58 crc kubenswrapper[4897]: I1121 14:33:58.992881 4897 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b546c9e-dba8-4806-a2ed-8023271b448a-config-data\") on node \"crc\" DevicePath \"\"" Nov 21 14:33:59 crc kubenswrapper[4897]: I1121 14:33:59.403226 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756","Type":"ContainerDied","Data":"2d3f77279621bedaaaa2adc858cb4b8df929d9fc05eb69b73803167dbae2a405"} Nov 21 14:33:59 crc kubenswrapper[4897]: I1121 14:33:59.403583 4897 scope.go:117] "RemoveContainer" containerID="b40036e131ce501b96971cc68d11d31dd29c34c53797f7d0d45c993eaf89d3ec" Nov 21 14:33:59 crc kubenswrapper[4897]: I1121 14:33:59.404048 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 21 14:33:59 crc kubenswrapper[4897]: I1121 14:33:59.406184 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-75bd44869-p288j" event={"ID":"7b546c9e-dba8-4806-a2ed-8023271b448a","Type":"ContainerDied","Data":"541dc73ecc166fc2449316454225cc83f3c390225f4a3426e6e554d0f34d67d0"} Nov 21 14:33:59 crc kubenswrapper[4897]: I1121 14:33:59.406261 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-75bd44869-p288j" Nov 21 14:33:59 crc kubenswrapper[4897]: I1121 14:33:59.422413 4897 generic.go:334] "Generic (PLEG): container finished" podID="9b4dd223-2765-4cd7-92c6-293fadac9520" containerID="84989f9793378edd4304f2cd8abc4845242c3a1d4738690d303837d51a1a273a" exitCode=2 Nov 21 14:33:59 crc kubenswrapper[4897]: I1121 14:33:59.422453 4897 generic.go:334] "Generic (PLEG): container finished" podID="9b4dd223-2765-4cd7-92c6-293fadac9520" containerID="361fc6578a5f944b99422cc41c214a4fce5a6bf5a72887df4dfbaf9250665d89" exitCode=0 Nov 21 14:33:59 crc kubenswrapper[4897]: I1121 14:33:59.422476 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9b4dd223-2765-4cd7-92c6-293fadac9520","Type":"ContainerDied","Data":"84989f9793378edd4304f2cd8abc4845242c3a1d4738690d303837d51a1a273a"} Nov 21 14:33:59 crc kubenswrapper[4897]: I1121 14:33:59.422521 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9b4dd223-2765-4cd7-92c6-293fadac9520","Type":"ContainerDied","Data":"361fc6578a5f944b99422cc41c214a4fce5a6bf5a72887df4dfbaf9250665d89"} Nov 21 14:33:59 crc kubenswrapper[4897]: I1121 14:33:59.478814 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-75bd44869-p288j"] Nov 21 14:33:59 crc kubenswrapper[4897]: I1121 14:33:59.491758 4897 scope.go:117] "RemoveContainer" containerID="fbffe9604199c2c6d9800e7f1823e0e982f185c924f967e3770744b01d341b13" Nov 21 14:33:59 crc kubenswrapper[4897]: I1121 14:33:59.491918 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-engine-75bd44869-p288j"] Nov 21 14:33:59 crc kubenswrapper[4897]: I1121 14:33:59.506168 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 21 14:33:59 crc kubenswrapper[4897]: I1121 14:33:59.513886 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 21 14:33:59 crc kubenswrapper[4897]: I1121 14:33:59.524738 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 21 14:33:59 crc kubenswrapper[4897]: E1121 14:33:59.525306 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b546c9e-dba8-4806-a2ed-8023271b448a" containerName="heat-engine" Nov 21 14:33:59 crc kubenswrapper[4897]: I1121 14:33:59.525324 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b546c9e-dba8-4806-a2ed-8023271b448a" containerName="heat-engine" Nov 21 14:33:59 crc kubenswrapper[4897]: E1121 14:33:59.525340 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756" containerName="glance-httpd" Nov 21 14:33:59 crc kubenswrapper[4897]: I1121 14:33:59.525348 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756" containerName="glance-httpd" Nov 21 14:33:59 crc kubenswrapper[4897]: E1121 14:33:59.525361 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b100fec-a276-44ac-b14d-febed2079580" containerName="mariadb-database-create" Nov 21 14:33:59 crc kubenswrapper[4897]: I1121 14:33:59.525367 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b100fec-a276-44ac-b14d-febed2079580" containerName="mariadb-database-create" Nov 21 14:33:59 crc kubenswrapper[4897]: E1121 14:33:59.525380 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5440e5f-8e45-4348-ba61-e9d3e56832c6" containerName="mariadb-account-create" Nov 21 14:33:59 crc kubenswrapper[4897]: I1121 14:33:59.525386 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5440e5f-8e45-4348-ba61-e9d3e56832c6" containerName="mariadb-account-create" Nov 21 14:33:59 crc kubenswrapper[4897]: E1121 14:33:59.525401 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f516d79-7775-40cc-9d50-4167d9b1afbb" containerName="mariadb-database-create" Nov 21 14:33:59 crc kubenswrapper[4897]: I1121 14:33:59.525416 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f516d79-7775-40cc-9d50-4167d9b1afbb" containerName="mariadb-database-create" Nov 21 14:33:59 crc kubenswrapper[4897]: E1121 14:33:59.525441 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a34d9b31-8c74-47b4-bedc-0cf717bc8f3f" containerName="heat-cfnapi" Nov 21 14:33:59 crc kubenswrapper[4897]: I1121 14:33:59.525447 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="a34d9b31-8c74-47b4-bedc-0cf717bc8f3f" containerName="heat-cfnapi" Nov 21 14:33:59 crc kubenswrapper[4897]: E1121 14:33:59.525462 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ad7667f-2a3d-445b-a582-ad2bbecc23b7" containerName="heat-api" Nov 21 14:33:59 crc kubenswrapper[4897]: I1121 14:33:59.525467 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ad7667f-2a3d-445b-a582-ad2bbecc23b7" containerName="heat-api" Nov 21 14:33:59 crc kubenswrapper[4897]: E1121 14:33:59.525476 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756" containerName="glance-log" Nov 21 14:33:59 crc kubenswrapper[4897]: I1121 14:33:59.525482 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756" containerName="glance-log" Nov 21 14:33:59 crc kubenswrapper[4897]: E1121 14:33:59.525501 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="548045ab-1444-4b7a-a663-3fa90b7fdb22" containerName="mariadb-account-create" Nov 21 14:33:59 crc kubenswrapper[4897]: I1121 14:33:59.525524 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="548045ab-1444-4b7a-a663-3fa90b7fdb22" containerName="mariadb-account-create" Nov 21 14:33:59 crc kubenswrapper[4897]: E1121 14:33:59.525535 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ad7667f-2a3d-445b-a582-ad2bbecc23b7" containerName="heat-api" Nov 21 14:33:59 crc kubenswrapper[4897]: I1121 14:33:59.525541 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ad7667f-2a3d-445b-a582-ad2bbecc23b7" containerName="heat-api" Nov 21 14:33:59 crc kubenswrapper[4897]: E1121 14:33:59.525554 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56d6af89-a8f0-4b64-9426-04536dedf90c" containerName="mariadb-account-create" Nov 21 14:33:59 crc kubenswrapper[4897]: I1121 14:33:59.525560 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="56d6af89-a8f0-4b64-9426-04536dedf90c" containerName="mariadb-account-create" Nov 21 14:33:59 crc kubenswrapper[4897]: E1121 14:33:59.525579 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d0c865b-9c9b-49cf-ba2f-730f98adce04" containerName="mariadb-database-create" Nov 21 14:33:59 crc kubenswrapper[4897]: I1121 14:33:59.525585 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d0c865b-9c9b-49cf-ba2f-730f98adce04" containerName="mariadb-database-create" Nov 21 14:33:59 crc kubenswrapper[4897]: I1121 14:33:59.525834 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f516d79-7775-40cc-9d50-4167d9b1afbb" containerName="mariadb-database-create" Nov 21 14:33:59 crc kubenswrapper[4897]: I1121 14:33:59.525859 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="548045ab-1444-4b7a-a663-3fa90b7fdb22" containerName="mariadb-account-create" Nov 21 14:33:59 crc kubenswrapper[4897]: I1121 14:33:59.525878 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756" containerName="glance-httpd" Nov 21 14:33:59 crc kubenswrapper[4897]: I1121 14:33:59.525893 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="a34d9b31-8c74-47b4-bedc-0cf717bc8f3f" containerName="heat-cfnapi" Nov 21 14:33:59 crc kubenswrapper[4897]: I1121 14:33:59.525905 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ad7667f-2a3d-445b-a582-ad2bbecc23b7" containerName="heat-api" Nov 21 14:33:59 crc kubenswrapper[4897]: I1121 14:33:59.525914 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5440e5f-8e45-4348-ba61-e9d3e56832c6" containerName="mariadb-account-create" Nov 21 14:33:59 crc kubenswrapper[4897]: I1121 14:33:59.525925 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b546c9e-dba8-4806-a2ed-8023271b448a" containerName="heat-engine" Nov 21 14:33:59 crc kubenswrapper[4897]: I1121 14:33:59.525936 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="56d6af89-a8f0-4b64-9426-04536dedf90c" containerName="mariadb-account-create" Nov 21 14:33:59 crc kubenswrapper[4897]: I1121 14:33:59.525956 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b100fec-a276-44ac-b14d-febed2079580" containerName="mariadb-database-create" Nov 21 14:33:59 crc kubenswrapper[4897]: I1121 14:33:59.525970 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ad7667f-2a3d-445b-a582-ad2bbecc23b7" containerName="heat-api" Nov 21 14:33:59 crc kubenswrapper[4897]: I1121 14:33:59.525982 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756" containerName="glance-log" Nov 21 14:33:59 crc kubenswrapper[4897]: I1121 14:33:59.525994 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d0c865b-9c9b-49cf-ba2f-730f98adce04" containerName="mariadb-database-create" Nov 21 14:33:59 crc kubenswrapper[4897]: E1121 14:33:59.526270 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a34d9b31-8c74-47b4-bedc-0cf717bc8f3f" containerName="heat-cfnapi" Nov 21 14:33:59 crc kubenswrapper[4897]: I1121 14:33:59.526281 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="a34d9b31-8c74-47b4-bedc-0cf717bc8f3f" containerName="heat-cfnapi" Nov 21 14:33:59 crc kubenswrapper[4897]: I1121 14:33:59.526561 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="a34d9b31-8c74-47b4-bedc-0cf717bc8f3f" containerName="heat-cfnapi" Nov 21 14:33:59 crc kubenswrapper[4897]: I1121 14:33:59.527587 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 21 14:33:59 crc kubenswrapper[4897]: I1121 14:33:59.534356 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Nov 21 14:33:59 crc kubenswrapper[4897]: I1121 14:33:59.534495 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 21 14:33:59 crc kubenswrapper[4897]: I1121 14:33:59.542835 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 21 14:33:59 crc kubenswrapper[4897]: I1121 14:33:59.578555 4897 scope.go:117] "RemoveContainer" containerID="8e41603de28669b082f8f5511f675dd2ebd4d7fd9fc98678968408620410fbb2" Nov 21 14:33:59 crc kubenswrapper[4897]: I1121 14:33:59.613022 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72e6ab2c-5802-4b3c-9be7-b3082407df28-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"72e6ab2c-5802-4b3c-9be7-b3082407df28\") " pod="openstack/glance-default-internal-api-0" Nov 21 14:33:59 crc kubenswrapper[4897]: I1121 14:33:59.613090 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t8j55\" (UniqueName: \"kubernetes.io/projected/72e6ab2c-5802-4b3c-9be7-b3082407df28-kube-api-access-t8j55\") pod \"glance-default-internal-api-0\" (UID: \"72e6ab2c-5802-4b3c-9be7-b3082407df28\") " pod="openstack/glance-default-internal-api-0" Nov 21 14:33:59 crc kubenswrapper[4897]: I1121 14:33:59.613143 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/72e6ab2c-5802-4b3c-9be7-b3082407df28-scripts\") pod \"glance-default-internal-api-0\" (UID: \"72e6ab2c-5802-4b3c-9be7-b3082407df28\") " pod="openstack/glance-default-internal-api-0" Nov 21 14:33:59 crc kubenswrapper[4897]: I1121 14:33:59.613183 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"72e6ab2c-5802-4b3c-9be7-b3082407df28\") " pod="openstack/glance-default-internal-api-0" Nov 21 14:33:59 crc kubenswrapper[4897]: I1121 14:33:59.613200 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/72e6ab2c-5802-4b3c-9be7-b3082407df28-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"72e6ab2c-5802-4b3c-9be7-b3082407df28\") " pod="openstack/glance-default-internal-api-0" Nov 21 14:33:59 crc kubenswrapper[4897]: I1121 14:33:59.613230 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/72e6ab2c-5802-4b3c-9be7-b3082407df28-logs\") pod \"glance-default-internal-api-0\" (UID: \"72e6ab2c-5802-4b3c-9be7-b3082407df28\") " pod="openstack/glance-default-internal-api-0" Nov 21 14:33:59 crc kubenswrapper[4897]: I1121 14:33:59.613274 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/72e6ab2c-5802-4b3c-9be7-b3082407df28-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"72e6ab2c-5802-4b3c-9be7-b3082407df28\") " pod="openstack/glance-default-internal-api-0" Nov 21 14:33:59 crc kubenswrapper[4897]: I1121 14:33:59.613322 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72e6ab2c-5802-4b3c-9be7-b3082407df28-config-data\") pod \"glance-default-internal-api-0\" (UID: \"72e6ab2c-5802-4b3c-9be7-b3082407df28\") " pod="openstack/glance-default-internal-api-0" Nov 21 14:33:59 crc kubenswrapper[4897]: I1121 14:33:59.715404 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t8j55\" (UniqueName: \"kubernetes.io/projected/72e6ab2c-5802-4b3c-9be7-b3082407df28-kube-api-access-t8j55\") pod \"glance-default-internal-api-0\" (UID: \"72e6ab2c-5802-4b3c-9be7-b3082407df28\") " pod="openstack/glance-default-internal-api-0" Nov 21 14:33:59 crc kubenswrapper[4897]: I1121 14:33:59.715489 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/72e6ab2c-5802-4b3c-9be7-b3082407df28-scripts\") pod \"glance-default-internal-api-0\" (UID: \"72e6ab2c-5802-4b3c-9be7-b3082407df28\") " pod="openstack/glance-default-internal-api-0" Nov 21 14:33:59 crc kubenswrapper[4897]: I1121 14:33:59.715662 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"72e6ab2c-5802-4b3c-9be7-b3082407df28\") " pod="openstack/glance-default-internal-api-0" Nov 21 14:33:59 crc kubenswrapper[4897]: I1121 14:33:59.715687 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/72e6ab2c-5802-4b3c-9be7-b3082407df28-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"72e6ab2c-5802-4b3c-9be7-b3082407df28\") " pod="openstack/glance-default-internal-api-0" Nov 21 14:33:59 crc kubenswrapper[4897]: I1121 14:33:59.715727 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/72e6ab2c-5802-4b3c-9be7-b3082407df28-logs\") pod \"glance-default-internal-api-0\" (UID: \"72e6ab2c-5802-4b3c-9be7-b3082407df28\") " pod="openstack/glance-default-internal-api-0" Nov 21 14:33:59 crc kubenswrapper[4897]: I1121 14:33:59.715793 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/72e6ab2c-5802-4b3c-9be7-b3082407df28-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"72e6ab2c-5802-4b3c-9be7-b3082407df28\") " pod="openstack/glance-default-internal-api-0" Nov 21 14:33:59 crc kubenswrapper[4897]: I1121 14:33:59.715858 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72e6ab2c-5802-4b3c-9be7-b3082407df28-config-data\") pod \"glance-default-internal-api-0\" (UID: \"72e6ab2c-5802-4b3c-9be7-b3082407df28\") " pod="openstack/glance-default-internal-api-0" Nov 21 14:33:59 crc kubenswrapper[4897]: I1121 14:33:59.715933 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72e6ab2c-5802-4b3c-9be7-b3082407df28-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"72e6ab2c-5802-4b3c-9be7-b3082407df28\") " pod="openstack/glance-default-internal-api-0" Nov 21 14:33:59 crc kubenswrapper[4897]: I1121 14:33:59.716491 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/72e6ab2c-5802-4b3c-9be7-b3082407df28-logs\") pod \"glance-default-internal-api-0\" (UID: \"72e6ab2c-5802-4b3c-9be7-b3082407df28\") " pod="openstack/glance-default-internal-api-0" Nov 21 14:33:59 crc kubenswrapper[4897]: I1121 14:33:59.716572 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/72e6ab2c-5802-4b3c-9be7-b3082407df28-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"72e6ab2c-5802-4b3c-9be7-b3082407df28\") " pod="openstack/glance-default-internal-api-0" Nov 21 14:33:59 crc kubenswrapper[4897]: I1121 14:33:59.717314 4897 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"72e6ab2c-5802-4b3c-9be7-b3082407df28\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/glance-default-internal-api-0" Nov 21 14:33:59 crc kubenswrapper[4897]: I1121 14:33:59.722213 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/72e6ab2c-5802-4b3c-9be7-b3082407df28-scripts\") pod \"glance-default-internal-api-0\" (UID: \"72e6ab2c-5802-4b3c-9be7-b3082407df28\") " pod="openstack/glance-default-internal-api-0" Nov 21 14:33:59 crc kubenswrapper[4897]: I1121 14:33:59.722605 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72e6ab2c-5802-4b3c-9be7-b3082407df28-config-data\") pod \"glance-default-internal-api-0\" (UID: \"72e6ab2c-5802-4b3c-9be7-b3082407df28\") " pod="openstack/glance-default-internal-api-0" Nov 21 14:33:59 crc kubenswrapper[4897]: I1121 14:33:59.724426 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72e6ab2c-5802-4b3c-9be7-b3082407df28-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"72e6ab2c-5802-4b3c-9be7-b3082407df28\") " pod="openstack/glance-default-internal-api-0" Nov 21 14:33:59 crc kubenswrapper[4897]: I1121 14:33:59.747344 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/72e6ab2c-5802-4b3c-9be7-b3082407df28-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"72e6ab2c-5802-4b3c-9be7-b3082407df28\") " pod="openstack/glance-default-internal-api-0" Nov 21 14:33:59 crc kubenswrapper[4897]: I1121 14:33:59.748405 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t8j55\" (UniqueName: \"kubernetes.io/projected/72e6ab2c-5802-4b3c-9be7-b3082407df28-kube-api-access-t8j55\") pod \"glance-default-internal-api-0\" (UID: \"72e6ab2c-5802-4b3c-9be7-b3082407df28\") " pod="openstack/glance-default-internal-api-0" Nov 21 14:33:59 crc kubenswrapper[4897]: I1121 14:33:59.766337 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"72e6ab2c-5802-4b3c-9be7-b3082407df28\") " pod="openstack/glance-default-internal-api-0" Nov 21 14:33:59 crc kubenswrapper[4897]: I1121 14:33:59.883689 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 21 14:34:00 crc kubenswrapper[4897]: I1121 14:34:00.120319 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7b546c9e-dba8-4806-a2ed-8023271b448a" path="/var/lib/kubelet/pods/7b546c9e-dba8-4806-a2ed-8023271b448a/volumes" Nov 21 14:34:00 crc kubenswrapper[4897]: I1121 14:34:00.122584 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756" path="/var/lib/kubelet/pods/cdc9a9c2-dd55-4fe9-85cb-f4f2bf4ab756/volumes" Nov 21 14:34:00 crc kubenswrapper[4897]: I1121 14:34:00.510138 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 21 14:34:01 crc kubenswrapper[4897]: I1121 14:34:01.479448 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"72e6ab2c-5802-4b3c-9be7-b3082407df28","Type":"ContainerStarted","Data":"c0bb609f29f729523759bce39f788341871838d2f8e5470517de9f3fb93284d6"} Nov 21 14:34:01 crc kubenswrapper[4897]: I1121 14:34:01.484738 4897 generic.go:334] "Generic (PLEG): container finished" podID="7155191e-f228-4639-8a75-ed8ff3dce81a" containerID="dc5724ef2786b81a902bc7373794e6066c91ce62272f7e949534d19dee9f3fe0" exitCode=0 Nov 21 14:34:01 crc kubenswrapper[4897]: I1121 14:34:01.484770 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"7155191e-f228-4639-8a75-ed8ff3dce81a","Type":"ContainerDied","Data":"dc5724ef2786b81a902bc7373794e6066c91ce62272f7e949534d19dee9f3fe0"} Nov 21 14:34:01 crc kubenswrapper[4897]: I1121 14:34:01.484797 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"7155191e-f228-4639-8a75-ed8ff3dce81a","Type":"ContainerDied","Data":"f46e3df66f856439d496964b0db493d27a14089e32fff851de3b1b81d913b3e4"} Nov 21 14:34:01 crc kubenswrapper[4897]: I1121 14:34:01.484810 4897 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f46e3df66f856439d496964b0db493d27a14089e32fff851de3b1b81d913b3e4" Nov 21 14:34:01 crc kubenswrapper[4897]: I1121 14:34:01.600702 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 21 14:34:01 crc kubenswrapper[4897]: I1121 14:34:01.786258 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7155191e-f228-4639-8a75-ed8ff3dce81a-public-tls-certs\") pod \"7155191e-f228-4639-8a75-ed8ff3dce81a\" (UID: \"7155191e-f228-4639-8a75-ed8ff3dce81a\") " Nov 21 14:34:01 crc kubenswrapper[4897]: I1121 14:34:01.786696 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7155191e-f228-4639-8a75-ed8ff3dce81a-logs\") pod \"7155191e-f228-4639-8a75-ed8ff3dce81a\" (UID: \"7155191e-f228-4639-8a75-ed8ff3dce81a\") " Nov 21 14:34:01 crc kubenswrapper[4897]: I1121 14:34:01.786743 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7155191e-f228-4639-8a75-ed8ff3dce81a-combined-ca-bundle\") pod \"7155191e-f228-4639-8a75-ed8ff3dce81a\" (UID: \"7155191e-f228-4639-8a75-ed8ff3dce81a\") " Nov 21 14:34:01 crc kubenswrapper[4897]: I1121 14:34:01.786822 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7155191e-f228-4639-8a75-ed8ff3dce81a-config-data\") pod \"7155191e-f228-4639-8a75-ed8ff3dce81a\" (UID: \"7155191e-f228-4639-8a75-ed8ff3dce81a\") " Nov 21 14:34:01 crc kubenswrapper[4897]: I1121 14:34:01.786931 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dpbk2\" (UniqueName: \"kubernetes.io/projected/7155191e-f228-4639-8a75-ed8ff3dce81a-kube-api-access-dpbk2\") pod \"7155191e-f228-4639-8a75-ed8ff3dce81a\" (UID: \"7155191e-f228-4639-8a75-ed8ff3dce81a\") " Nov 21 14:34:01 crc kubenswrapper[4897]: I1121 14:34:01.787046 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7155191e-f228-4639-8a75-ed8ff3dce81a-httpd-run\") pod \"7155191e-f228-4639-8a75-ed8ff3dce81a\" (UID: \"7155191e-f228-4639-8a75-ed8ff3dce81a\") " Nov 21 14:34:01 crc kubenswrapper[4897]: I1121 14:34:01.787065 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7155191e-f228-4639-8a75-ed8ff3dce81a-scripts\") pod \"7155191e-f228-4639-8a75-ed8ff3dce81a\" (UID: \"7155191e-f228-4639-8a75-ed8ff3dce81a\") " Nov 21 14:34:01 crc kubenswrapper[4897]: I1121 14:34:01.787104 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"7155191e-f228-4639-8a75-ed8ff3dce81a\" (UID: \"7155191e-f228-4639-8a75-ed8ff3dce81a\") " Nov 21 14:34:01 crc kubenswrapper[4897]: I1121 14:34:01.788391 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7155191e-f228-4639-8a75-ed8ff3dce81a-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "7155191e-f228-4639-8a75-ed8ff3dce81a" (UID: "7155191e-f228-4639-8a75-ed8ff3dce81a"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:34:01 crc kubenswrapper[4897]: I1121 14:34:01.788759 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7155191e-f228-4639-8a75-ed8ff3dce81a-logs" (OuterVolumeSpecName: "logs") pod "7155191e-f228-4639-8a75-ed8ff3dce81a" (UID: "7155191e-f228-4639-8a75-ed8ff3dce81a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:34:01 crc kubenswrapper[4897]: I1121 14:34:01.794840 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7155191e-f228-4639-8a75-ed8ff3dce81a-scripts" (OuterVolumeSpecName: "scripts") pod "7155191e-f228-4639-8a75-ed8ff3dce81a" (UID: "7155191e-f228-4639-8a75-ed8ff3dce81a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:34:01 crc kubenswrapper[4897]: I1121 14:34:01.803715 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7155191e-f228-4639-8a75-ed8ff3dce81a-kube-api-access-dpbk2" (OuterVolumeSpecName: "kube-api-access-dpbk2") pod "7155191e-f228-4639-8a75-ed8ff3dce81a" (UID: "7155191e-f228-4639-8a75-ed8ff3dce81a"). InnerVolumeSpecName "kube-api-access-dpbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:34:01 crc kubenswrapper[4897]: I1121 14:34:01.812963 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance") pod "7155191e-f228-4639-8a75-ed8ff3dce81a" (UID: "7155191e-f228-4639-8a75-ed8ff3dce81a"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 21 14:34:01 crc kubenswrapper[4897]: I1121 14:34:01.832293 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7155191e-f228-4639-8a75-ed8ff3dce81a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7155191e-f228-4639-8a75-ed8ff3dce81a" (UID: "7155191e-f228-4639-8a75-ed8ff3dce81a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:34:01 crc kubenswrapper[4897]: I1121 14:34:01.880129 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7155191e-f228-4639-8a75-ed8ff3dce81a-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "7155191e-f228-4639-8a75-ed8ff3dce81a" (UID: "7155191e-f228-4639-8a75-ed8ff3dce81a"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:34:01 crc kubenswrapper[4897]: I1121 14:34:01.889252 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7155191e-f228-4639-8a75-ed8ff3dce81a-config-data" (OuterVolumeSpecName: "config-data") pod "7155191e-f228-4639-8a75-ed8ff3dce81a" (UID: "7155191e-f228-4639-8a75-ed8ff3dce81a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:34:01 crc kubenswrapper[4897]: I1121 14:34:01.890192 4897 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7155191e-f228-4639-8a75-ed8ff3dce81a-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 21 14:34:01 crc kubenswrapper[4897]: I1121 14:34:01.890228 4897 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7155191e-f228-4639-8a75-ed8ff3dce81a-scripts\") on node \"crc\" DevicePath \"\"" Nov 21 14:34:01 crc kubenswrapper[4897]: I1121 14:34:01.890259 4897 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Nov 21 14:34:01 crc kubenswrapper[4897]: I1121 14:34:01.890272 4897 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7155191e-f228-4639-8a75-ed8ff3dce81a-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 21 14:34:01 crc kubenswrapper[4897]: I1121 14:34:01.890286 4897 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7155191e-f228-4639-8a75-ed8ff3dce81a-logs\") on node \"crc\" DevicePath \"\"" Nov 21 14:34:01 crc kubenswrapper[4897]: I1121 14:34:01.890296 4897 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7155191e-f228-4639-8a75-ed8ff3dce81a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 21 14:34:01 crc kubenswrapper[4897]: I1121 14:34:01.890307 4897 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7155191e-f228-4639-8a75-ed8ff3dce81a-config-data\") on node \"crc\" DevicePath \"\"" Nov 21 14:34:01 crc kubenswrapper[4897]: I1121 14:34:01.890320 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dpbk2\" (UniqueName: \"kubernetes.io/projected/7155191e-f228-4639-8a75-ed8ff3dce81a-kube-api-access-dpbk2\") on node \"crc\" DevicePath \"\"" Nov 21 14:34:01 crc kubenswrapper[4897]: I1121 14:34:01.945281 4897 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Nov 21 14:34:01 crc kubenswrapper[4897]: I1121 14:34:01.995090 4897 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Nov 21 14:34:02 crc kubenswrapper[4897]: I1121 14:34:02.076605 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-cv5pw"] Nov 21 14:34:02 crc kubenswrapper[4897]: E1121 14:34:02.077188 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7155191e-f228-4639-8a75-ed8ff3dce81a" containerName="glance-httpd" Nov 21 14:34:02 crc kubenswrapper[4897]: I1121 14:34:02.077206 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="7155191e-f228-4639-8a75-ed8ff3dce81a" containerName="glance-httpd" Nov 21 14:34:02 crc kubenswrapper[4897]: E1121 14:34:02.077224 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7155191e-f228-4639-8a75-ed8ff3dce81a" containerName="glance-log" Nov 21 14:34:02 crc kubenswrapper[4897]: I1121 14:34:02.077231 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="7155191e-f228-4639-8a75-ed8ff3dce81a" containerName="glance-log" Nov 21 14:34:02 crc kubenswrapper[4897]: I1121 14:34:02.083862 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="7155191e-f228-4639-8a75-ed8ff3dce81a" containerName="glance-httpd" Nov 21 14:34:02 crc kubenswrapper[4897]: I1121 14:34:02.083952 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="7155191e-f228-4639-8a75-ed8ff3dce81a" containerName="glance-log" Nov 21 14:34:02 crc kubenswrapper[4897]: I1121 14:34:02.085918 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-cv5pw" Nov 21 14:34:02 crc kubenswrapper[4897]: I1121 14:34:02.094414 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-26nl9" Nov 21 14:34:02 crc kubenswrapper[4897]: I1121 14:34:02.094715 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Nov 21 14:34:02 crc kubenswrapper[4897]: I1121 14:34:02.095152 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Nov 21 14:34:02 crc kubenswrapper[4897]: I1121 14:34:02.188241 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-cv5pw"] Nov 21 14:34:02 crc kubenswrapper[4897]: I1121 14:34:02.211292 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6e90788-4f97-45c0-aacc-cfa848ac0f2f-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-cv5pw\" (UID: \"c6e90788-4f97-45c0-aacc-cfa848ac0f2f\") " pod="openstack/nova-cell0-conductor-db-sync-cv5pw" Nov 21 14:34:02 crc kubenswrapper[4897]: I1121 14:34:02.211400 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c6e90788-4f97-45c0-aacc-cfa848ac0f2f-scripts\") pod \"nova-cell0-conductor-db-sync-cv5pw\" (UID: \"c6e90788-4f97-45c0-aacc-cfa848ac0f2f\") " pod="openstack/nova-cell0-conductor-db-sync-cv5pw" Nov 21 14:34:02 crc kubenswrapper[4897]: I1121 14:34:02.211471 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6e90788-4f97-45c0-aacc-cfa848ac0f2f-config-data\") pod \"nova-cell0-conductor-db-sync-cv5pw\" (UID: \"c6e90788-4f97-45c0-aacc-cfa848ac0f2f\") " pod="openstack/nova-cell0-conductor-db-sync-cv5pw" Nov 21 14:34:02 crc kubenswrapper[4897]: I1121 14:34:02.211531 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tqxjv\" (UniqueName: \"kubernetes.io/projected/c6e90788-4f97-45c0-aacc-cfa848ac0f2f-kube-api-access-tqxjv\") pod \"nova-cell0-conductor-db-sync-cv5pw\" (UID: \"c6e90788-4f97-45c0-aacc-cfa848ac0f2f\") " pod="openstack/nova-cell0-conductor-db-sync-cv5pw" Nov 21 14:34:02 crc kubenswrapper[4897]: I1121 14:34:02.313159 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tqxjv\" (UniqueName: \"kubernetes.io/projected/c6e90788-4f97-45c0-aacc-cfa848ac0f2f-kube-api-access-tqxjv\") pod \"nova-cell0-conductor-db-sync-cv5pw\" (UID: \"c6e90788-4f97-45c0-aacc-cfa848ac0f2f\") " pod="openstack/nova-cell0-conductor-db-sync-cv5pw" Nov 21 14:34:02 crc kubenswrapper[4897]: I1121 14:34:02.313425 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6e90788-4f97-45c0-aacc-cfa848ac0f2f-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-cv5pw\" (UID: \"c6e90788-4f97-45c0-aacc-cfa848ac0f2f\") " pod="openstack/nova-cell0-conductor-db-sync-cv5pw" Nov 21 14:34:02 crc kubenswrapper[4897]: I1121 14:34:02.313546 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c6e90788-4f97-45c0-aacc-cfa848ac0f2f-scripts\") pod \"nova-cell0-conductor-db-sync-cv5pw\" (UID: \"c6e90788-4f97-45c0-aacc-cfa848ac0f2f\") " pod="openstack/nova-cell0-conductor-db-sync-cv5pw" Nov 21 14:34:02 crc kubenswrapper[4897]: I1121 14:34:02.313658 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6e90788-4f97-45c0-aacc-cfa848ac0f2f-config-data\") pod \"nova-cell0-conductor-db-sync-cv5pw\" (UID: \"c6e90788-4f97-45c0-aacc-cfa848ac0f2f\") " pod="openstack/nova-cell0-conductor-db-sync-cv5pw" Nov 21 14:34:02 crc kubenswrapper[4897]: I1121 14:34:02.318090 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6e90788-4f97-45c0-aacc-cfa848ac0f2f-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-cv5pw\" (UID: \"c6e90788-4f97-45c0-aacc-cfa848ac0f2f\") " pod="openstack/nova-cell0-conductor-db-sync-cv5pw" Nov 21 14:34:02 crc kubenswrapper[4897]: I1121 14:34:02.318298 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c6e90788-4f97-45c0-aacc-cfa848ac0f2f-scripts\") pod \"nova-cell0-conductor-db-sync-cv5pw\" (UID: \"c6e90788-4f97-45c0-aacc-cfa848ac0f2f\") " pod="openstack/nova-cell0-conductor-db-sync-cv5pw" Nov 21 14:34:02 crc kubenswrapper[4897]: I1121 14:34:02.319209 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6e90788-4f97-45c0-aacc-cfa848ac0f2f-config-data\") pod \"nova-cell0-conductor-db-sync-cv5pw\" (UID: \"c6e90788-4f97-45c0-aacc-cfa848ac0f2f\") " pod="openstack/nova-cell0-conductor-db-sync-cv5pw" Nov 21 14:34:02 crc kubenswrapper[4897]: I1121 14:34:02.339956 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tqxjv\" (UniqueName: \"kubernetes.io/projected/c6e90788-4f97-45c0-aacc-cfa848ac0f2f-kube-api-access-tqxjv\") pod \"nova-cell0-conductor-db-sync-cv5pw\" (UID: \"c6e90788-4f97-45c0-aacc-cfa848ac0f2f\") " pod="openstack/nova-cell0-conductor-db-sync-cv5pw" Nov 21 14:34:02 crc kubenswrapper[4897]: I1121 14:34:02.462304 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-cv5pw" Nov 21 14:34:02 crc kubenswrapper[4897]: I1121 14:34:02.505653 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 21 14:34:02 crc kubenswrapper[4897]: I1121 14:34:02.506874 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"72e6ab2c-5802-4b3c-9be7-b3082407df28","Type":"ContainerStarted","Data":"21c4b9305c40b407bbedba22f06da151a11b95f3d76cb53706c2b458295e79ad"} Nov 21 14:34:02 crc kubenswrapper[4897]: I1121 14:34:02.636722 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 21 14:34:02 crc kubenswrapper[4897]: I1121 14:34:02.647093 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 21 14:34:02 crc kubenswrapper[4897]: I1121 14:34:02.664677 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 21 14:34:02 crc kubenswrapper[4897]: I1121 14:34:02.666608 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 21 14:34:02 crc kubenswrapper[4897]: I1121 14:34:02.676142 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 21 14:34:02 crc kubenswrapper[4897]: I1121 14:34:02.676300 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Nov 21 14:34:02 crc kubenswrapper[4897]: I1121 14:34:02.697734 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 21 14:34:02 crc kubenswrapper[4897]: I1121 14:34:02.802020 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-fv2n4" Nov 21 14:34:02 crc kubenswrapper[4897]: I1121 14:34:02.803271 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-fv2n4" Nov 21 14:34:02 crc kubenswrapper[4897]: I1121 14:34:02.844231 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/daf7e654-043d-4394-a210-c841975c7de7-scripts\") pod \"glance-default-external-api-0\" (UID: \"daf7e654-043d-4394-a210-c841975c7de7\") " pod="openstack/glance-default-external-api-0" Nov 21 14:34:02 crc kubenswrapper[4897]: I1121 14:34:02.844312 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7fqqz\" (UniqueName: \"kubernetes.io/projected/daf7e654-043d-4394-a210-c841975c7de7-kube-api-access-7fqqz\") pod \"glance-default-external-api-0\" (UID: \"daf7e654-043d-4394-a210-c841975c7de7\") " pod="openstack/glance-default-external-api-0" Nov 21 14:34:02 crc kubenswrapper[4897]: I1121 14:34:02.844603 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"daf7e654-043d-4394-a210-c841975c7de7\") " pod="openstack/glance-default-external-api-0" Nov 21 14:34:02 crc kubenswrapper[4897]: I1121 14:34:02.844884 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/daf7e654-043d-4394-a210-c841975c7de7-config-data\") pod \"glance-default-external-api-0\" (UID: \"daf7e654-043d-4394-a210-c841975c7de7\") " pod="openstack/glance-default-external-api-0" Nov 21 14:34:02 crc kubenswrapper[4897]: I1121 14:34:02.845000 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/daf7e654-043d-4394-a210-c841975c7de7-logs\") pod \"glance-default-external-api-0\" (UID: \"daf7e654-043d-4394-a210-c841975c7de7\") " pod="openstack/glance-default-external-api-0" Nov 21 14:34:02 crc kubenswrapper[4897]: I1121 14:34:02.845081 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/daf7e654-043d-4394-a210-c841975c7de7-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"daf7e654-043d-4394-a210-c841975c7de7\") " pod="openstack/glance-default-external-api-0" Nov 21 14:34:02 crc kubenswrapper[4897]: I1121 14:34:02.845142 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/daf7e654-043d-4394-a210-c841975c7de7-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"daf7e654-043d-4394-a210-c841975c7de7\") " pod="openstack/glance-default-external-api-0" Nov 21 14:34:02 crc kubenswrapper[4897]: I1121 14:34:02.845171 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/daf7e654-043d-4394-a210-c841975c7de7-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"daf7e654-043d-4394-a210-c841975c7de7\") " pod="openstack/glance-default-external-api-0" Nov 21 14:34:02 crc kubenswrapper[4897]: I1121 14:34:02.869880 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-fv2n4" Nov 21 14:34:02 crc kubenswrapper[4897]: I1121 14:34:02.947920 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/daf7e654-043d-4394-a210-c841975c7de7-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"daf7e654-043d-4394-a210-c841975c7de7\") " pod="openstack/glance-default-external-api-0" Nov 21 14:34:02 crc kubenswrapper[4897]: I1121 14:34:02.948040 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/daf7e654-043d-4394-a210-c841975c7de7-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"daf7e654-043d-4394-a210-c841975c7de7\") " pod="openstack/glance-default-external-api-0" Nov 21 14:34:02 crc kubenswrapper[4897]: I1121 14:34:02.948136 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/daf7e654-043d-4394-a210-c841975c7de7-scripts\") pod \"glance-default-external-api-0\" (UID: \"daf7e654-043d-4394-a210-c841975c7de7\") " pod="openstack/glance-default-external-api-0" Nov 21 14:34:02 crc kubenswrapper[4897]: I1121 14:34:02.948218 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7fqqz\" (UniqueName: \"kubernetes.io/projected/daf7e654-043d-4394-a210-c841975c7de7-kube-api-access-7fqqz\") pod \"glance-default-external-api-0\" (UID: \"daf7e654-043d-4394-a210-c841975c7de7\") " pod="openstack/glance-default-external-api-0" Nov 21 14:34:02 crc kubenswrapper[4897]: I1121 14:34:02.948489 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"daf7e654-043d-4394-a210-c841975c7de7\") " pod="openstack/glance-default-external-api-0" Nov 21 14:34:02 crc kubenswrapper[4897]: I1121 14:34:02.948781 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/daf7e654-043d-4394-a210-c841975c7de7-config-data\") pod \"glance-default-external-api-0\" (UID: \"daf7e654-043d-4394-a210-c841975c7de7\") " pod="openstack/glance-default-external-api-0" Nov 21 14:34:02 crc kubenswrapper[4897]: I1121 14:34:02.948845 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/daf7e654-043d-4394-a210-c841975c7de7-logs\") pod \"glance-default-external-api-0\" (UID: \"daf7e654-043d-4394-a210-c841975c7de7\") " pod="openstack/glance-default-external-api-0" Nov 21 14:34:02 crc kubenswrapper[4897]: I1121 14:34:02.948935 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/daf7e654-043d-4394-a210-c841975c7de7-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"daf7e654-043d-4394-a210-c841975c7de7\") " pod="openstack/glance-default-external-api-0" Nov 21 14:34:02 crc kubenswrapper[4897]: I1121 14:34:02.949624 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/daf7e654-043d-4394-a210-c841975c7de7-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"daf7e654-043d-4394-a210-c841975c7de7\") " pod="openstack/glance-default-external-api-0" Nov 21 14:34:02 crc kubenswrapper[4897]: I1121 14:34:02.952006 4897 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"daf7e654-043d-4394-a210-c841975c7de7\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/glance-default-external-api-0" Nov 21 14:34:02 crc kubenswrapper[4897]: I1121 14:34:02.958144 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/daf7e654-043d-4394-a210-c841975c7de7-logs\") pod \"glance-default-external-api-0\" (UID: \"daf7e654-043d-4394-a210-c841975c7de7\") " pod="openstack/glance-default-external-api-0" Nov 21 14:34:02 crc kubenswrapper[4897]: I1121 14:34:02.963472 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/daf7e654-043d-4394-a210-c841975c7de7-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"daf7e654-043d-4394-a210-c841975c7de7\") " pod="openstack/glance-default-external-api-0" Nov 21 14:34:02 crc kubenswrapper[4897]: I1121 14:34:02.964853 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/daf7e654-043d-4394-a210-c841975c7de7-config-data\") pod \"glance-default-external-api-0\" (UID: \"daf7e654-043d-4394-a210-c841975c7de7\") " pod="openstack/glance-default-external-api-0" Nov 21 14:34:02 crc kubenswrapper[4897]: I1121 14:34:02.983245 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/daf7e654-043d-4394-a210-c841975c7de7-scripts\") pod \"glance-default-external-api-0\" (UID: \"daf7e654-043d-4394-a210-c841975c7de7\") " pod="openstack/glance-default-external-api-0" Nov 21 14:34:02 crc kubenswrapper[4897]: I1121 14:34:02.998421 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7fqqz\" (UniqueName: \"kubernetes.io/projected/daf7e654-043d-4394-a210-c841975c7de7-kube-api-access-7fqqz\") pod \"glance-default-external-api-0\" (UID: \"daf7e654-043d-4394-a210-c841975c7de7\") " pod="openstack/glance-default-external-api-0" Nov 21 14:34:02 crc kubenswrapper[4897]: I1121 14:34:02.998797 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/daf7e654-043d-4394-a210-c841975c7de7-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"daf7e654-043d-4394-a210-c841975c7de7\") " pod="openstack/glance-default-external-api-0" Nov 21 14:34:03 crc kubenswrapper[4897]: I1121 14:34:03.021773 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"daf7e654-043d-4394-a210-c841975c7de7\") " pod="openstack/glance-default-external-api-0" Nov 21 14:34:03 crc kubenswrapper[4897]: I1121 14:34:03.169157 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-cv5pw"] Nov 21 14:34:03 crc kubenswrapper[4897]: I1121 14:34:03.307182 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 21 14:34:03 crc kubenswrapper[4897]: I1121 14:34:03.534379 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-cv5pw" event={"ID":"c6e90788-4f97-45c0-aacc-cfa848ac0f2f","Type":"ContainerStarted","Data":"08b721dc40c171056df1d3e5579a86b1a03c8a8b48881d1df0c2c5cd1e3bba32"} Nov 21 14:34:03 crc kubenswrapper[4897]: I1121 14:34:03.541686 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"72e6ab2c-5802-4b3c-9be7-b3082407df28","Type":"ContainerStarted","Data":"dfac7df6c24896f1faa4e5c9a50fac8f76a3b732f765367813af7aa10194e77c"} Nov 21 14:34:03 crc kubenswrapper[4897]: I1121 14:34:03.673759 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-fv2n4" Nov 21 14:34:03 crc kubenswrapper[4897]: I1121 14:34:03.720043 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.720021325 podStartE2EDuration="4.720021325s" podCreationTimestamp="2025-11-21 14:33:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:34:03.602932335 +0000 UTC m=+1520.887525820" watchObservedRunningTime="2025-11-21 14:34:03.720021325 +0000 UTC m=+1521.004614800" Nov 21 14:34:03 crc kubenswrapper[4897]: I1121 14:34:03.801826 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fv2n4"] Nov 21 14:34:03 crc kubenswrapper[4897]: I1121 14:34:03.980015 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 21 14:34:04 crc kubenswrapper[4897]: I1121 14:34:04.107399 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7155191e-f228-4639-8a75-ed8ff3dce81a" path="/var/lib/kubelet/pods/7155191e-f228-4639-8a75-ed8ff3dce81a/volumes" Nov 21 14:34:04 crc kubenswrapper[4897]: I1121 14:34:04.371603 4897 patch_prober.go:28] interesting pod/machine-config-daemon-krv5b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 21 14:34:04 crc kubenswrapper[4897]: I1121 14:34:04.371959 4897 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 21 14:34:04 crc kubenswrapper[4897]: I1121 14:34:04.560668 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"daf7e654-043d-4394-a210-c841975c7de7","Type":"ContainerStarted","Data":"8ee262fdef745a867b2db2c931d71c7341c5a60683843679e194d8cf498dc8f6"} Nov 21 14:34:05 crc kubenswrapper[4897]: I1121 14:34:05.574441 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"daf7e654-043d-4394-a210-c841975c7de7","Type":"ContainerStarted","Data":"f6119f4be7fab38b7108d968d0c126ce6d4b63b0c2f9dcf622ab0c985a99d226"} Nov 21 14:34:05 crc kubenswrapper[4897]: I1121 14:34:05.575038 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"daf7e654-043d-4394-a210-c841975c7de7","Type":"ContainerStarted","Data":"7d1d066481d256bc361367d2b3dc2c5966a1a9510f2e1ea5b4f7eb9a4521f7b0"} Nov 21 14:34:05 crc kubenswrapper[4897]: I1121 14:34:05.574607 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-fv2n4" podUID="e7b274ff-8198-43ca-a3bb-b551c97762dc" containerName="registry-server" containerID="cri-o://2516c86ebae5b237dc067031873de6ca1e6174e5c88ce3c1b1dfe8b5d6a057f2" gracePeriod=2 Nov 21 14:34:06 crc kubenswrapper[4897]: I1121 14:34:06.181095 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-dtwql" podUID="6d3cad20-fb17-4726-b8fa-9fd5303b5eea" containerName="registry-server" probeResult="failure" output=< Nov 21 14:34:06 crc kubenswrapper[4897]: timeout: failed to connect service ":50051" within 1s Nov 21 14:34:06 crc kubenswrapper[4897]: > Nov 21 14:34:06 crc kubenswrapper[4897]: I1121 14:34:06.239647 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fv2n4" Nov 21 14:34:06 crc kubenswrapper[4897]: I1121 14:34:06.381557 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7b274ff-8198-43ca-a3bb-b551c97762dc-utilities\") pod \"e7b274ff-8198-43ca-a3bb-b551c97762dc\" (UID: \"e7b274ff-8198-43ca-a3bb-b551c97762dc\") " Nov 21 14:34:06 crc kubenswrapper[4897]: I1121 14:34:06.381642 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7b274ff-8198-43ca-a3bb-b551c97762dc-catalog-content\") pod \"e7b274ff-8198-43ca-a3bb-b551c97762dc\" (UID: \"e7b274ff-8198-43ca-a3bb-b551c97762dc\") " Nov 21 14:34:06 crc kubenswrapper[4897]: I1121 14:34:06.381725 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-phct8\" (UniqueName: \"kubernetes.io/projected/e7b274ff-8198-43ca-a3bb-b551c97762dc-kube-api-access-phct8\") pod \"e7b274ff-8198-43ca-a3bb-b551c97762dc\" (UID: \"e7b274ff-8198-43ca-a3bb-b551c97762dc\") " Nov 21 14:34:06 crc kubenswrapper[4897]: I1121 14:34:06.382380 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e7b274ff-8198-43ca-a3bb-b551c97762dc-utilities" (OuterVolumeSpecName: "utilities") pod "e7b274ff-8198-43ca-a3bb-b551c97762dc" (UID: "e7b274ff-8198-43ca-a3bb-b551c97762dc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:34:06 crc kubenswrapper[4897]: I1121 14:34:06.390656 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7b274ff-8198-43ca-a3bb-b551c97762dc-kube-api-access-phct8" (OuterVolumeSpecName: "kube-api-access-phct8") pod "e7b274ff-8198-43ca-a3bb-b551c97762dc" (UID: "e7b274ff-8198-43ca-a3bb-b551c97762dc"). InnerVolumeSpecName "kube-api-access-phct8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:34:06 crc kubenswrapper[4897]: I1121 14:34:06.456100 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e7b274ff-8198-43ca-a3bb-b551c97762dc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e7b274ff-8198-43ca-a3bb-b551c97762dc" (UID: "e7b274ff-8198-43ca-a3bb-b551c97762dc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:34:06 crc kubenswrapper[4897]: I1121 14:34:06.484593 4897 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7b274ff-8198-43ca-a3bb-b551c97762dc-utilities\") on node \"crc\" DevicePath \"\"" Nov 21 14:34:06 crc kubenswrapper[4897]: I1121 14:34:06.484627 4897 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7b274ff-8198-43ca-a3bb-b551c97762dc-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 21 14:34:06 crc kubenswrapper[4897]: I1121 14:34:06.484641 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-phct8\" (UniqueName: \"kubernetes.io/projected/e7b274ff-8198-43ca-a3bb-b551c97762dc-kube-api-access-phct8\") on node \"crc\" DevicePath \"\"" Nov 21 14:34:06 crc kubenswrapper[4897]: I1121 14:34:06.592440 4897 generic.go:334] "Generic (PLEG): container finished" podID="e7b274ff-8198-43ca-a3bb-b551c97762dc" containerID="2516c86ebae5b237dc067031873de6ca1e6174e5c88ce3c1b1dfe8b5d6a057f2" exitCode=0 Nov 21 14:34:06 crc kubenswrapper[4897]: I1121 14:34:06.592533 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fv2n4" Nov 21 14:34:06 crc kubenswrapper[4897]: I1121 14:34:06.592554 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fv2n4" event={"ID":"e7b274ff-8198-43ca-a3bb-b551c97762dc","Type":"ContainerDied","Data":"2516c86ebae5b237dc067031873de6ca1e6174e5c88ce3c1b1dfe8b5d6a057f2"} Nov 21 14:34:06 crc kubenswrapper[4897]: I1121 14:34:06.594811 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fv2n4" event={"ID":"e7b274ff-8198-43ca-a3bb-b551c97762dc","Type":"ContainerDied","Data":"ad644f6a2ce4440f96a5a22e67eb9752858380c4b2b1c5d883a0e220659e5765"} Nov 21 14:34:06 crc kubenswrapper[4897]: I1121 14:34:06.594841 4897 scope.go:117] "RemoveContainer" containerID="2516c86ebae5b237dc067031873de6ca1e6174e5c88ce3c1b1dfe8b5d6a057f2" Nov 21 14:34:06 crc kubenswrapper[4897]: I1121 14:34:06.634997 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.634975974 podStartE2EDuration="4.634975974s" podCreationTimestamp="2025-11-21 14:34:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:34:06.61878538 +0000 UTC m=+1523.903378855" watchObservedRunningTime="2025-11-21 14:34:06.634975974 +0000 UTC m=+1523.919569449" Nov 21 14:34:06 crc kubenswrapper[4897]: I1121 14:34:06.655427 4897 scope.go:117] "RemoveContainer" containerID="3dfe33c8352b4bc1ba944517a9f5429db25ada1f9a96f4b7dbb3e9911a14b5a2" Nov 21 14:34:06 crc kubenswrapper[4897]: I1121 14:34:06.668421 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fv2n4"] Nov 21 14:34:06 crc kubenswrapper[4897]: I1121 14:34:06.692702 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-fv2n4"] Nov 21 14:34:06 crc kubenswrapper[4897]: I1121 14:34:06.700649 4897 scope.go:117] "RemoveContainer" containerID="54d2d5ffa197abd65ca3d2125bf24efa2801465e8b983a41d30322bf45000e69" Nov 21 14:34:06 crc kubenswrapper[4897]: I1121 14:34:06.754456 4897 scope.go:117] "RemoveContainer" containerID="2516c86ebae5b237dc067031873de6ca1e6174e5c88ce3c1b1dfe8b5d6a057f2" Nov 21 14:34:06 crc kubenswrapper[4897]: E1121 14:34:06.755271 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2516c86ebae5b237dc067031873de6ca1e6174e5c88ce3c1b1dfe8b5d6a057f2\": container with ID starting with 2516c86ebae5b237dc067031873de6ca1e6174e5c88ce3c1b1dfe8b5d6a057f2 not found: ID does not exist" containerID="2516c86ebae5b237dc067031873de6ca1e6174e5c88ce3c1b1dfe8b5d6a057f2" Nov 21 14:34:06 crc kubenswrapper[4897]: I1121 14:34:06.756864 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2516c86ebae5b237dc067031873de6ca1e6174e5c88ce3c1b1dfe8b5d6a057f2"} err="failed to get container status \"2516c86ebae5b237dc067031873de6ca1e6174e5c88ce3c1b1dfe8b5d6a057f2\": rpc error: code = NotFound desc = could not find container \"2516c86ebae5b237dc067031873de6ca1e6174e5c88ce3c1b1dfe8b5d6a057f2\": container with ID starting with 2516c86ebae5b237dc067031873de6ca1e6174e5c88ce3c1b1dfe8b5d6a057f2 not found: ID does not exist" Nov 21 14:34:06 crc kubenswrapper[4897]: I1121 14:34:06.757561 4897 scope.go:117] "RemoveContainer" containerID="3dfe33c8352b4bc1ba944517a9f5429db25ada1f9a96f4b7dbb3e9911a14b5a2" Nov 21 14:34:06 crc kubenswrapper[4897]: E1121 14:34:06.757880 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3dfe33c8352b4bc1ba944517a9f5429db25ada1f9a96f4b7dbb3e9911a14b5a2\": container with ID starting with 3dfe33c8352b4bc1ba944517a9f5429db25ada1f9a96f4b7dbb3e9911a14b5a2 not found: ID does not exist" containerID="3dfe33c8352b4bc1ba944517a9f5429db25ada1f9a96f4b7dbb3e9911a14b5a2" Nov 21 14:34:06 crc kubenswrapper[4897]: I1121 14:34:06.757913 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3dfe33c8352b4bc1ba944517a9f5429db25ada1f9a96f4b7dbb3e9911a14b5a2"} err="failed to get container status \"3dfe33c8352b4bc1ba944517a9f5429db25ada1f9a96f4b7dbb3e9911a14b5a2\": rpc error: code = NotFound desc = could not find container \"3dfe33c8352b4bc1ba944517a9f5429db25ada1f9a96f4b7dbb3e9911a14b5a2\": container with ID starting with 3dfe33c8352b4bc1ba944517a9f5429db25ada1f9a96f4b7dbb3e9911a14b5a2 not found: ID does not exist" Nov 21 14:34:06 crc kubenswrapper[4897]: I1121 14:34:06.757933 4897 scope.go:117] "RemoveContainer" containerID="54d2d5ffa197abd65ca3d2125bf24efa2801465e8b983a41d30322bf45000e69" Nov 21 14:34:06 crc kubenswrapper[4897]: E1121 14:34:06.758288 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"54d2d5ffa197abd65ca3d2125bf24efa2801465e8b983a41d30322bf45000e69\": container with ID starting with 54d2d5ffa197abd65ca3d2125bf24efa2801465e8b983a41d30322bf45000e69 not found: ID does not exist" containerID="54d2d5ffa197abd65ca3d2125bf24efa2801465e8b983a41d30322bf45000e69" Nov 21 14:34:06 crc kubenswrapper[4897]: I1121 14:34:06.758314 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"54d2d5ffa197abd65ca3d2125bf24efa2801465e8b983a41d30322bf45000e69"} err="failed to get container status \"54d2d5ffa197abd65ca3d2125bf24efa2801465e8b983a41d30322bf45000e69\": rpc error: code = NotFound desc = could not find container \"54d2d5ffa197abd65ca3d2125bf24efa2801465e8b983a41d30322bf45000e69\": container with ID starting with 54d2d5ffa197abd65ca3d2125bf24efa2801465e8b983a41d30322bf45000e69 not found: ID does not exist" Nov 21 14:34:08 crc kubenswrapper[4897]: I1121 14:34:08.101412 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7b274ff-8198-43ca-a3bb-b551c97762dc" path="/var/lib/kubelet/pods/e7b274ff-8198-43ca-a3bb-b551c97762dc/volumes" Nov 21 14:34:09 crc kubenswrapper[4897]: I1121 14:34:09.630075 4897 generic.go:334] "Generic (PLEG): container finished" podID="9b4dd223-2765-4cd7-92c6-293fadac9520" containerID="45dea6b4b4b005a5c41b3ad4e7113041cad0ffea101d5022396b3fbc4e7c2bf8" exitCode=0 Nov 21 14:34:09 crc kubenswrapper[4897]: I1121 14:34:09.630117 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9b4dd223-2765-4cd7-92c6-293fadac9520","Type":"ContainerDied","Data":"45dea6b4b4b005a5c41b3ad4e7113041cad0ffea101d5022396b3fbc4e7c2bf8"} Nov 21 14:34:09 crc kubenswrapper[4897]: I1121 14:34:09.884017 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 21 14:34:09 crc kubenswrapper[4897]: I1121 14:34:09.884066 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 21 14:34:09 crc kubenswrapper[4897]: I1121 14:34:09.934265 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 21 14:34:09 crc kubenswrapper[4897]: I1121 14:34:09.937395 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 21 14:34:10 crc kubenswrapper[4897]: I1121 14:34:10.640791 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 21 14:34:10 crc kubenswrapper[4897]: I1121 14:34:10.641116 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 21 14:34:12 crc kubenswrapper[4897]: I1121 14:34:12.103216 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-7446d77dd8-rg2bn" Nov 21 14:34:13 crc kubenswrapper[4897]: I1121 14:34:13.309326 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 21 14:34:13 crc kubenswrapper[4897]: I1121 14:34:13.309682 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 21 14:34:13 crc kubenswrapper[4897]: I1121 14:34:13.356597 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 21 14:34:13 crc kubenswrapper[4897]: I1121 14:34:13.364587 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 21 14:34:13 crc kubenswrapper[4897]: I1121 14:34:13.676186 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-cv5pw" event={"ID":"c6e90788-4f97-45c0-aacc-cfa848ac0f2f","Type":"ContainerStarted","Data":"03e6cd518a468591cf4ed2a936e891b6a325fffb7bd905bc50948a197be7a6cb"} Nov 21 14:34:13 crc kubenswrapper[4897]: I1121 14:34:13.676530 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 21 14:34:13 crc kubenswrapper[4897]: I1121 14:34:13.676789 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 21 14:34:14 crc kubenswrapper[4897]: I1121 14:34:14.491104 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 21 14:34:14 crc kubenswrapper[4897]: I1121 14:34:14.491552 4897 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 21 14:34:14 crc kubenswrapper[4897]: I1121 14:34:14.496028 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 21 14:34:14 crc kubenswrapper[4897]: I1121 14:34:14.523798 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-cv5pw" podStartSLOduration=3.152230873 podStartE2EDuration="12.523774698s" podCreationTimestamp="2025-11-21 14:34:02 +0000 UTC" firstStartedPulling="2025-11-21 14:34:03.169100678 +0000 UTC m=+1520.453694153" lastFinishedPulling="2025-11-21 14:34:12.540644503 +0000 UTC m=+1529.825237978" observedRunningTime="2025-11-21 14:34:13.699633761 +0000 UTC m=+1530.984227236" watchObservedRunningTime="2025-11-21 14:34:14.523774698 +0000 UTC m=+1531.808368173" Nov 21 14:34:16 crc kubenswrapper[4897]: I1121 14:34:16.082666 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 21 14:34:16 crc kubenswrapper[4897]: I1121 14:34:16.083039 4897 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 21 14:34:16 crc kubenswrapper[4897]: I1121 14:34:16.144628 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 21 14:34:16 crc kubenswrapper[4897]: I1121 14:34:16.179052 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-dtwql" podUID="6d3cad20-fb17-4726-b8fa-9fd5303b5eea" containerName="registry-server" probeResult="failure" output=< Nov 21 14:34:16 crc kubenswrapper[4897]: timeout: failed to connect service ":50051" within 1s Nov 21 14:34:16 crc kubenswrapper[4897]: > Nov 21 14:34:22 crc kubenswrapper[4897]: I1121 14:34:22.101687 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-f46d49697-qwn8n" Nov 21 14:34:22 crc kubenswrapper[4897]: I1121 14:34:22.175917 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-7446d77dd8-rg2bn"] Nov 21 14:34:22 crc kubenswrapper[4897]: I1121 14:34:22.176634 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-7446d77dd8-rg2bn" podUID="bac76d5f-def0-4fe8-babf-2c1a171f04dc" containerName="neutron-api" containerID="cri-o://e8721fc63494ab662dbf3cfe15e933c7d2469a6932ecbc6c02e513fec6121a44" gracePeriod=30 Nov 21 14:34:22 crc kubenswrapper[4897]: I1121 14:34:22.177246 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-7446d77dd8-rg2bn" podUID="bac76d5f-def0-4fe8-babf-2c1a171f04dc" containerName="neutron-httpd" containerID="cri-o://8d266e169a0a0fae0622b99a0bdaf334d3e2e83f1f5b17a4df91a215e16c0d8a" gracePeriod=30 Nov 21 14:34:22 crc kubenswrapper[4897]: I1121 14:34:22.819530 4897 generic.go:334] "Generic (PLEG): container finished" podID="bac76d5f-def0-4fe8-babf-2c1a171f04dc" containerID="8d266e169a0a0fae0622b99a0bdaf334d3e2e83f1f5b17a4df91a215e16c0d8a" exitCode=0 Nov 21 14:34:22 crc kubenswrapper[4897]: I1121 14:34:22.819598 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7446d77dd8-rg2bn" event={"ID":"bac76d5f-def0-4fe8-babf-2c1a171f04dc","Type":"ContainerDied","Data":"8d266e169a0a0fae0622b99a0bdaf334d3e2e83f1f5b17a4df91a215e16c0d8a"} Nov 21 14:34:25 crc kubenswrapper[4897]: I1121 14:34:25.165670 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="9b4dd223-2765-4cd7-92c6-293fadac9520" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Nov 21 14:34:26 crc kubenswrapper[4897]: I1121 14:34:26.167685 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-dtwql" podUID="6d3cad20-fb17-4726-b8fa-9fd5303b5eea" containerName="registry-server" probeResult="failure" output=< Nov 21 14:34:26 crc kubenswrapper[4897]: timeout: failed to connect service ":50051" within 1s Nov 21 14:34:26 crc kubenswrapper[4897]: > Nov 21 14:34:27 crc kubenswrapper[4897]: I1121 14:34:27.874322 4897 generic.go:334] "Generic (PLEG): container finished" podID="bac76d5f-def0-4fe8-babf-2c1a171f04dc" containerID="e8721fc63494ab662dbf3cfe15e933c7d2469a6932ecbc6c02e513fec6121a44" exitCode=0 Nov 21 14:34:27 crc kubenswrapper[4897]: I1121 14:34:27.874420 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7446d77dd8-rg2bn" event={"ID":"bac76d5f-def0-4fe8-babf-2c1a171f04dc","Type":"ContainerDied","Data":"e8721fc63494ab662dbf3cfe15e933c7d2469a6932ecbc6c02e513fec6121a44"} Nov 21 14:34:28 crc kubenswrapper[4897]: I1121 14:34:28.320858 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7446d77dd8-rg2bn" Nov 21 14:34:28 crc kubenswrapper[4897]: I1121 14:34:28.402757 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9ttr9\" (UniqueName: \"kubernetes.io/projected/bac76d5f-def0-4fe8-babf-2c1a171f04dc-kube-api-access-9ttr9\") pod \"bac76d5f-def0-4fe8-babf-2c1a171f04dc\" (UID: \"bac76d5f-def0-4fe8-babf-2c1a171f04dc\") " Nov 21 14:34:28 crc kubenswrapper[4897]: I1121 14:34:28.403119 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/bac76d5f-def0-4fe8-babf-2c1a171f04dc-config\") pod \"bac76d5f-def0-4fe8-babf-2c1a171f04dc\" (UID: \"bac76d5f-def0-4fe8-babf-2c1a171f04dc\") " Nov 21 14:34:28 crc kubenswrapper[4897]: I1121 14:34:28.403260 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bac76d5f-def0-4fe8-babf-2c1a171f04dc-combined-ca-bundle\") pod \"bac76d5f-def0-4fe8-babf-2c1a171f04dc\" (UID: \"bac76d5f-def0-4fe8-babf-2c1a171f04dc\") " Nov 21 14:34:28 crc kubenswrapper[4897]: I1121 14:34:28.403295 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/bac76d5f-def0-4fe8-babf-2c1a171f04dc-httpd-config\") pod \"bac76d5f-def0-4fe8-babf-2c1a171f04dc\" (UID: \"bac76d5f-def0-4fe8-babf-2c1a171f04dc\") " Nov 21 14:34:28 crc kubenswrapper[4897]: I1121 14:34:28.403388 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/bac76d5f-def0-4fe8-babf-2c1a171f04dc-ovndb-tls-certs\") pod \"bac76d5f-def0-4fe8-babf-2c1a171f04dc\" (UID: \"bac76d5f-def0-4fe8-babf-2c1a171f04dc\") " Nov 21 14:34:28 crc kubenswrapper[4897]: I1121 14:34:28.410141 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bac76d5f-def0-4fe8-babf-2c1a171f04dc-kube-api-access-9ttr9" (OuterVolumeSpecName: "kube-api-access-9ttr9") pod "bac76d5f-def0-4fe8-babf-2c1a171f04dc" (UID: "bac76d5f-def0-4fe8-babf-2c1a171f04dc"). InnerVolumeSpecName "kube-api-access-9ttr9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:34:28 crc kubenswrapper[4897]: I1121 14:34:28.411640 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bac76d5f-def0-4fe8-babf-2c1a171f04dc-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "bac76d5f-def0-4fe8-babf-2c1a171f04dc" (UID: "bac76d5f-def0-4fe8-babf-2c1a171f04dc"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:34:28 crc kubenswrapper[4897]: E1121 14:34:28.439988 4897 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9b4dd223_2765_4cd7_92c6_293fadac9520.slice/crio-2c786aefd013c99dc5c1d324fa90173506b4fc5af2323d0c2bdbc16bc6ad25d1.scope\": RecentStats: unable to find data in memory cache]" Nov 21 14:34:28 crc kubenswrapper[4897]: I1121 14:34:28.481539 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bac76d5f-def0-4fe8-babf-2c1a171f04dc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bac76d5f-def0-4fe8-babf-2c1a171f04dc" (UID: "bac76d5f-def0-4fe8-babf-2c1a171f04dc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:34:28 crc kubenswrapper[4897]: I1121 14:34:28.501088 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bac76d5f-def0-4fe8-babf-2c1a171f04dc-config" (OuterVolumeSpecName: "config") pod "bac76d5f-def0-4fe8-babf-2c1a171f04dc" (UID: "bac76d5f-def0-4fe8-babf-2c1a171f04dc"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:34:28 crc kubenswrapper[4897]: I1121 14:34:28.506235 4897 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bac76d5f-def0-4fe8-babf-2c1a171f04dc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 21 14:34:28 crc kubenswrapper[4897]: I1121 14:34:28.506383 4897 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/bac76d5f-def0-4fe8-babf-2c1a171f04dc-httpd-config\") on node \"crc\" DevicePath \"\"" Nov 21 14:34:28 crc kubenswrapper[4897]: I1121 14:34:28.506468 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9ttr9\" (UniqueName: \"kubernetes.io/projected/bac76d5f-def0-4fe8-babf-2c1a171f04dc-kube-api-access-9ttr9\") on node \"crc\" DevicePath \"\"" Nov 21 14:34:28 crc kubenswrapper[4897]: I1121 14:34:28.506597 4897 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/bac76d5f-def0-4fe8-babf-2c1a171f04dc-config\") on node \"crc\" DevicePath \"\"" Nov 21 14:34:28 crc kubenswrapper[4897]: I1121 14:34:28.537201 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bac76d5f-def0-4fe8-babf-2c1a171f04dc-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "bac76d5f-def0-4fe8-babf-2c1a171f04dc" (UID: "bac76d5f-def0-4fe8-babf-2c1a171f04dc"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:34:28 crc kubenswrapper[4897]: I1121 14:34:28.608608 4897 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/bac76d5f-def0-4fe8-babf-2c1a171f04dc-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 21 14:34:28 crc kubenswrapper[4897]: I1121 14:34:28.713382 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 21 14:34:28 crc kubenswrapper[4897]: I1121 14:34:28.811358 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9b4dd223-2765-4cd7-92c6-293fadac9520-config-data\") pod \"9b4dd223-2765-4cd7-92c6-293fadac9520\" (UID: \"9b4dd223-2765-4cd7-92c6-293fadac9520\") " Nov 21 14:34:28 crc kubenswrapper[4897]: I1121 14:34:28.811459 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9b4dd223-2765-4cd7-92c6-293fadac9520-scripts\") pod \"9b4dd223-2765-4cd7-92c6-293fadac9520\" (UID: \"9b4dd223-2765-4cd7-92c6-293fadac9520\") " Nov 21 14:34:28 crc kubenswrapper[4897]: I1121 14:34:28.811629 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9b4dd223-2765-4cd7-92c6-293fadac9520-sg-core-conf-yaml\") pod \"9b4dd223-2765-4cd7-92c6-293fadac9520\" (UID: \"9b4dd223-2765-4cd7-92c6-293fadac9520\") " Nov 21 14:34:28 crc kubenswrapper[4897]: I1121 14:34:28.811663 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9b4dd223-2765-4cd7-92c6-293fadac9520-log-httpd\") pod \"9b4dd223-2765-4cd7-92c6-293fadac9520\" (UID: \"9b4dd223-2765-4cd7-92c6-293fadac9520\") " Nov 21 14:34:28 crc kubenswrapper[4897]: I1121 14:34:28.811871 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wfj4h\" (UniqueName: \"kubernetes.io/projected/9b4dd223-2765-4cd7-92c6-293fadac9520-kube-api-access-wfj4h\") pod \"9b4dd223-2765-4cd7-92c6-293fadac9520\" (UID: \"9b4dd223-2765-4cd7-92c6-293fadac9520\") " Nov 21 14:34:28 crc kubenswrapper[4897]: I1121 14:34:28.811951 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b4dd223-2765-4cd7-92c6-293fadac9520-combined-ca-bundle\") pod \"9b4dd223-2765-4cd7-92c6-293fadac9520\" (UID: \"9b4dd223-2765-4cd7-92c6-293fadac9520\") " Nov 21 14:34:28 crc kubenswrapper[4897]: I1121 14:34:28.812002 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9b4dd223-2765-4cd7-92c6-293fadac9520-run-httpd\") pod \"9b4dd223-2765-4cd7-92c6-293fadac9520\" (UID: \"9b4dd223-2765-4cd7-92c6-293fadac9520\") " Nov 21 14:34:28 crc kubenswrapper[4897]: I1121 14:34:28.813064 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9b4dd223-2765-4cd7-92c6-293fadac9520-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "9b4dd223-2765-4cd7-92c6-293fadac9520" (UID: "9b4dd223-2765-4cd7-92c6-293fadac9520"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:34:28 crc kubenswrapper[4897]: I1121 14:34:28.813155 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9b4dd223-2765-4cd7-92c6-293fadac9520-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "9b4dd223-2765-4cd7-92c6-293fadac9520" (UID: "9b4dd223-2765-4cd7-92c6-293fadac9520"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:34:28 crc kubenswrapper[4897]: I1121 14:34:28.815943 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9b4dd223-2765-4cd7-92c6-293fadac9520-kube-api-access-wfj4h" (OuterVolumeSpecName: "kube-api-access-wfj4h") pod "9b4dd223-2765-4cd7-92c6-293fadac9520" (UID: "9b4dd223-2765-4cd7-92c6-293fadac9520"). InnerVolumeSpecName "kube-api-access-wfj4h". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:34:28 crc kubenswrapper[4897]: I1121 14:34:28.816563 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b4dd223-2765-4cd7-92c6-293fadac9520-scripts" (OuterVolumeSpecName: "scripts") pod "9b4dd223-2765-4cd7-92c6-293fadac9520" (UID: "9b4dd223-2765-4cd7-92c6-293fadac9520"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:34:28 crc kubenswrapper[4897]: I1121 14:34:28.866485 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b4dd223-2765-4cd7-92c6-293fadac9520-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "9b4dd223-2765-4cd7-92c6-293fadac9520" (UID: "9b4dd223-2765-4cd7-92c6-293fadac9520"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:34:28 crc kubenswrapper[4897]: I1121 14:34:28.893703 4897 generic.go:334] "Generic (PLEG): container finished" podID="9b4dd223-2765-4cd7-92c6-293fadac9520" containerID="2c786aefd013c99dc5c1d324fa90173506b4fc5af2323d0c2bdbc16bc6ad25d1" exitCode=137 Nov 21 14:34:28 crc kubenswrapper[4897]: I1121 14:34:28.893803 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9b4dd223-2765-4cd7-92c6-293fadac9520","Type":"ContainerDied","Data":"2c786aefd013c99dc5c1d324fa90173506b4fc5af2323d0c2bdbc16bc6ad25d1"} Nov 21 14:34:28 crc kubenswrapper[4897]: I1121 14:34:28.893833 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9b4dd223-2765-4cd7-92c6-293fadac9520","Type":"ContainerDied","Data":"93b9b9e870a5485e47f3d6d422bf6253387084387ce1848166fc66402fb94d10"} Nov 21 14:34:28 crc kubenswrapper[4897]: I1121 14:34:28.893851 4897 scope.go:117] "RemoveContainer" containerID="2c786aefd013c99dc5c1d324fa90173506b4fc5af2323d0c2bdbc16bc6ad25d1" Nov 21 14:34:28 crc kubenswrapper[4897]: I1121 14:34:28.893893 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 21 14:34:28 crc kubenswrapper[4897]: I1121 14:34:28.899297 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7446d77dd8-rg2bn" event={"ID":"bac76d5f-def0-4fe8-babf-2c1a171f04dc","Type":"ContainerDied","Data":"d8b09365925618b0cb70df4372567c5ebac5e2690ccbb21641ffeefc36d8e164"} Nov 21 14:34:28 crc kubenswrapper[4897]: I1121 14:34:28.899577 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7446d77dd8-rg2bn" Nov 21 14:34:28 crc kubenswrapper[4897]: I1121 14:34:28.901869 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b4dd223-2765-4cd7-92c6-293fadac9520-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9b4dd223-2765-4cd7-92c6-293fadac9520" (UID: "9b4dd223-2765-4cd7-92c6-293fadac9520"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:34:28 crc kubenswrapper[4897]: I1121 14:34:28.914070 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wfj4h\" (UniqueName: \"kubernetes.io/projected/9b4dd223-2765-4cd7-92c6-293fadac9520-kube-api-access-wfj4h\") on node \"crc\" DevicePath \"\"" Nov 21 14:34:28 crc kubenswrapper[4897]: I1121 14:34:28.914095 4897 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b4dd223-2765-4cd7-92c6-293fadac9520-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 21 14:34:28 crc kubenswrapper[4897]: I1121 14:34:28.914103 4897 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9b4dd223-2765-4cd7-92c6-293fadac9520-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 21 14:34:28 crc kubenswrapper[4897]: I1121 14:34:28.914114 4897 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9b4dd223-2765-4cd7-92c6-293fadac9520-scripts\") on node \"crc\" DevicePath \"\"" Nov 21 14:34:28 crc kubenswrapper[4897]: I1121 14:34:28.914153 4897 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9b4dd223-2765-4cd7-92c6-293fadac9520-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 21 14:34:28 crc kubenswrapper[4897]: I1121 14:34:28.914163 4897 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9b4dd223-2765-4cd7-92c6-293fadac9520-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 21 14:34:28 crc kubenswrapper[4897]: I1121 14:34:28.926821 4897 scope.go:117] "RemoveContainer" containerID="84989f9793378edd4304f2cd8abc4845242c3a1d4738690d303837d51a1a273a" Nov 21 14:34:28 crc kubenswrapper[4897]: I1121 14:34:28.939233 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-7446d77dd8-rg2bn"] Nov 21 14:34:28 crc kubenswrapper[4897]: I1121 14:34:28.948050 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-7446d77dd8-rg2bn"] Nov 21 14:34:28 crc kubenswrapper[4897]: I1121 14:34:28.955043 4897 scope.go:117] "RemoveContainer" containerID="361fc6578a5f944b99422cc41c214a4fce5a6bf5a72887df4dfbaf9250665d89" Nov 21 14:34:28 crc kubenswrapper[4897]: I1121 14:34:28.960827 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b4dd223-2765-4cd7-92c6-293fadac9520-config-data" (OuterVolumeSpecName: "config-data") pod "9b4dd223-2765-4cd7-92c6-293fadac9520" (UID: "9b4dd223-2765-4cd7-92c6-293fadac9520"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:34:28 crc kubenswrapper[4897]: I1121 14:34:28.974292 4897 scope.go:117] "RemoveContainer" containerID="45dea6b4b4b005a5c41b3ad4e7113041cad0ffea101d5022396b3fbc4e7c2bf8" Nov 21 14:34:28 crc kubenswrapper[4897]: I1121 14:34:28.997811 4897 scope.go:117] "RemoveContainer" containerID="2c786aefd013c99dc5c1d324fa90173506b4fc5af2323d0c2bdbc16bc6ad25d1" Nov 21 14:34:28 crc kubenswrapper[4897]: E1121 14:34:28.998740 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2c786aefd013c99dc5c1d324fa90173506b4fc5af2323d0c2bdbc16bc6ad25d1\": container with ID starting with 2c786aefd013c99dc5c1d324fa90173506b4fc5af2323d0c2bdbc16bc6ad25d1 not found: ID does not exist" containerID="2c786aefd013c99dc5c1d324fa90173506b4fc5af2323d0c2bdbc16bc6ad25d1" Nov 21 14:34:28 crc kubenswrapper[4897]: I1121 14:34:28.998792 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2c786aefd013c99dc5c1d324fa90173506b4fc5af2323d0c2bdbc16bc6ad25d1"} err="failed to get container status \"2c786aefd013c99dc5c1d324fa90173506b4fc5af2323d0c2bdbc16bc6ad25d1\": rpc error: code = NotFound desc = could not find container \"2c786aefd013c99dc5c1d324fa90173506b4fc5af2323d0c2bdbc16bc6ad25d1\": container with ID starting with 2c786aefd013c99dc5c1d324fa90173506b4fc5af2323d0c2bdbc16bc6ad25d1 not found: ID does not exist" Nov 21 14:34:28 crc kubenswrapper[4897]: I1121 14:34:28.998827 4897 scope.go:117] "RemoveContainer" containerID="84989f9793378edd4304f2cd8abc4845242c3a1d4738690d303837d51a1a273a" Nov 21 14:34:28 crc kubenswrapper[4897]: E1121 14:34:28.999422 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"84989f9793378edd4304f2cd8abc4845242c3a1d4738690d303837d51a1a273a\": container with ID starting with 84989f9793378edd4304f2cd8abc4845242c3a1d4738690d303837d51a1a273a not found: ID does not exist" containerID="84989f9793378edd4304f2cd8abc4845242c3a1d4738690d303837d51a1a273a" Nov 21 14:34:28 crc kubenswrapper[4897]: I1121 14:34:28.999457 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"84989f9793378edd4304f2cd8abc4845242c3a1d4738690d303837d51a1a273a"} err="failed to get container status \"84989f9793378edd4304f2cd8abc4845242c3a1d4738690d303837d51a1a273a\": rpc error: code = NotFound desc = could not find container \"84989f9793378edd4304f2cd8abc4845242c3a1d4738690d303837d51a1a273a\": container with ID starting with 84989f9793378edd4304f2cd8abc4845242c3a1d4738690d303837d51a1a273a not found: ID does not exist" Nov 21 14:34:28 crc kubenswrapper[4897]: I1121 14:34:28.999476 4897 scope.go:117] "RemoveContainer" containerID="361fc6578a5f944b99422cc41c214a4fce5a6bf5a72887df4dfbaf9250665d89" Nov 21 14:34:28 crc kubenswrapper[4897]: E1121 14:34:28.999784 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"361fc6578a5f944b99422cc41c214a4fce5a6bf5a72887df4dfbaf9250665d89\": container with ID starting with 361fc6578a5f944b99422cc41c214a4fce5a6bf5a72887df4dfbaf9250665d89 not found: ID does not exist" containerID="361fc6578a5f944b99422cc41c214a4fce5a6bf5a72887df4dfbaf9250665d89" Nov 21 14:34:28 crc kubenswrapper[4897]: I1121 14:34:28.999809 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"361fc6578a5f944b99422cc41c214a4fce5a6bf5a72887df4dfbaf9250665d89"} err="failed to get container status \"361fc6578a5f944b99422cc41c214a4fce5a6bf5a72887df4dfbaf9250665d89\": rpc error: code = NotFound desc = could not find container \"361fc6578a5f944b99422cc41c214a4fce5a6bf5a72887df4dfbaf9250665d89\": container with ID starting with 361fc6578a5f944b99422cc41c214a4fce5a6bf5a72887df4dfbaf9250665d89 not found: ID does not exist" Nov 21 14:34:28 crc kubenswrapper[4897]: I1121 14:34:28.999822 4897 scope.go:117] "RemoveContainer" containerID="45dea6b4b4b005a5c41b3ad4e7113041cad0ffea101d5022396b3fbc4e7c2bf8" Nov 21 14:34:29 crc kubenswrapper[4897]: E1121 14:34:29.000120 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"45dea6b4b4b005a5c41b3ad4e7113041cad0ffea101d5022396b3fbc4e7c2bf8\": container with ID starting with 45dea6b4b4b005a5c41b3ad4e7113041cad0ffea101d5022396b3fbc4e7c2bf8 not found: ID does not exist" containerID="45dea6b4b4b005a5c41b3ad4e7113041cad0ffea101d5022396b3fbc4e7c2bf8" Nov 21 14:34:29 crc kubenswrapper[4897]: I1121 14:34:29.000156 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"45dea6b4b4b005a5c41b3ad4e7113041cad0ffea101d5022396b3fbc4e7c2bf8"} err="failed to get container status \"45dea6b4b4b005a5c41b3ad4e7113041cad0ffea101d5022396b3fbc4e7c2bf8\": rpc error: code = NotFound desc = could not find container \"45dea6b4b4b005a5c41b3ad4e7113041cad0ffea101d5022396b3fbc4e7c2bf8\": container with ID starting with 45dea6b4b4b005a5c41b3ad4e7113041cad0ffea101d5022396b3fbc4e7c2bf8 not found: ID does not exist" Nov 21 14:34:29 crc kubenswrapper[4897]: I1121 14:34:29.000176 4897 scope.go:117] "RemoveContainer" containerID="8d266e169a0a0fae0622b99a0bdaf334d3e2e83f1f5b17a4df91a215e16c0d8a" Nov 21 14:34:29 crc kubenswrapper[4897]: I1121 14:34:29.016801 4897 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9b4dd223-2765-4cd7-92c6-293fadac9520-config-data\") on node \"crc\" DevicePath \"\"" Nov 21 14:34:29 crc kubenswrapper[4897]: I1121 14:34:29.032457 4897 scope.go:117] "RemoveContainer" containerID="e8721fc63494ab662dbf3cfe15e933c7d2469a6932ecbc6c02e513fec6121a44" Nov 21 14:34:29 crc kubenswrapper[4897]: I1121 14:34:29.343881 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 21 14:34:29 crc kubenswrapper[4897]: I1121 14:34:29.360451 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 21 14:34:29 crc kubenswrapper[4897]: I1121 14:34:29.374760 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 21 14:34:29 crc kubenswrapper[4897]: E1121 14:34:29.375490 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7b274ff-8198-43ca-a3bb-b551c97762dc" containerName="extract-content" Nov 21 14:34:29 crc kubenswrapper[4897]: I1121 14:34:29.375545 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7b274ff-8198-43ca-a3bb-b551c97762dc" containerName="extract-content" Nov 21 14:34:29 crc kubenswrapper[4897]: E1121 14:34:29.375582 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7b274ff-8198-43ca-a3bb-b551c97762dc" containerName="extract-utilities" Nov 21 14:34:29 crc kubenswrapper[4897]: I1121 14:34:29.375595 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7b274ff-8198-43ca-a3bb-b551c97762dc" containerName="extract-utilities" Nov 21 14:34:29 crc kubenswrapper[4897]: E1121 14:34:29.375638 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b4dd223-2765-4cd7-92c6-293fadac9520" containerName="ceilometer-central-agent" Nov 21 14:34:29 crc kubenswrapper[4897]: I1121 14:34:29.375651 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b4dd223-2765-4cd7-92c6-293fadac9520" containerName="ceilometer-central-agent" Nov 21 14:34:29 crc kubenswrapper[4897]: E1121 14:34:29.375678 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b4dd223-2765-4cd7-92c6-293fadac9520" containerName="ceilometer-notification-agent" Nov 21 14:34:29 crc kubenswrapper[4897]: I1121 14:34:29.375690 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b4dd223-2765-4cd7-92c6-293fadac9520" containerName="ceilometer-notification-agent" Nov 21 14:34:29 crc kubenswrapper[4897]: E1121 14:34:29.375720 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b4dd223-2765-4cd7-92c6-293fadac9520" containerName="sg-core" Nov 21 14:34:29 crc kubenswrapper[4897]: I1121 14:34:29.375732 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b4dd223-2765-4cd7-92c6-293fadac9520" containerName="sg-core" Nov 21 14:34:29 crc kubenswrapper[4897]: E1121 14:34:29.375755 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7b274ff-8198-43ca-a3bb-b551c97762dc" containerName="registry-server" Nov 21 14:34:29 crc kubenswrapper[4897]: I1121 14:34:29.375767 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7b274ff-8198-43ca-a3bb-b551c97762dc" containerName="registry-server" Nov 21 14:34:29 crc kubenswrapper[4897]: E1121 14:34:29.375799 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bac76d5f-def0-4fe8-babf-2c1a171f04dc" containerName="neutron-api" Nov 21 14:34:29 crc kubenswrapper[4897]: I1121 14:34:29.375811 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="bac76d5f-def0-4fe8-babf-2c1a171f04dc" containerName="neutron-api" Nov 21 14:34:29 crc kubenswrapper[4897]: E1121 14:34:29.375827 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bac76d5f-def0-4fe8-babf-2c1a171f04dc" containerName="neutron-httpd" Nov 21 14:34:29 crc kubenswrapper[4897]: I1121 14:34:29.375839 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="bac76d5f-def0-4fe8-babf-2c1a171f04dc" containerName="neutron-httpd" Nov 21 14:34:29 crc kubenswrapper[4897]: E1121 14:34:29.375859 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b4dd223-2765-4cd7-92c6-293fadac9520" containerName="proxy-httpd" Nov 21 14:34:29 crc kubenswrapper[4897]: I1121 14:34:29.375870 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b4dd223-2765-4cd7-92c6-293fadac9520" containerName="proxy-httpd" Nov 21 14:34:29 crc kubenswrapper[4897]: I1121 14:34:29.376218 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b4dd223-2765-4cd7-92c6-293fadac9520" containerName="proxy-httpd" Nov 21 14:34:29 crc kubenswrapper[4897]: I1121 14:34:29.376268 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b4dd223-2765-4cd7-92c6-293fadac9520" containerName="ceilometer-notification-agent" Nov 21 14:34:29 crc kubenswrapper[4897]: I1121 14:34:29.376299 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b4dd223-2765-4cd7-92c6-293fadac9520" containerName="sg-core" Nov 21 14:34:29 crc kubenswrapper[4897]: I1121 14:34:29.376326 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="e7b274ff-8198-43ca-a3bb-b551c97762dc" containerName="registry-server" Nov 21 14:34:29 crc kubenswrapper[4897]: I1121 14:34:29.376350 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="bac76d5f-def0-4fe8-babf-2c1a171f04dc" containerName="neutron-httpd" Nov 21 14:34:29 crc kubenswrapper[4897]: I1121 14:34:29.376367 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b4dd223-2765-4cd7-92c6-293fadac9520" containerName="ceilometer-central-agent" Nov 21 14:34:29 crc kubenswrapper[4897]: I1121 14:34:29.376385 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="bac76d5f-def0-4fe8-babf-2c1a171f04dc" containerName="neutron-api" Nov 21 14:34:29 crc kubenswrapper[4897]: I1121 14:34:29.380187 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 21 14:34:29 crc kubenswrapper[4897]: I1121 14:34:29.383581 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 21 14:34:29 crc kubenswrapper[4897]: I1121 14:34:29.384454 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 21 14:34:29 crc kubenswrapper[4897]: I1121 14:34:29.392483 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 21 14:34:29 crc kubenswrapper[4897]: I1121 14:34:29.425129 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a594d7fb-9b6e-4dee-a3bf-391e2fb9db79-scripts\") pod \"ceilometer-0\" (UID: \"a594d7fb-9b6e-4dee-a3bf-391e2fb9db79\") " pod="openstack/ceilometer-0" Nov 21 14:34:29 crc kubenswrapper[4897]: I1121 14:34:29.425179 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a594d7fb-9b6e-4dee-a3bf-391e2fb9db79-log-httpd\") pod \"ceilometer-0\" (UID: \"a594d7fb-9b6e-4dee-a3bf-391e2fb9db79\") " pod="openstack/ceilometer-0" Nov 21 14:34:29 crc kubenswrapper[4897]: I1121 14:34:29.425216 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a594d7fb-9b6e-4dee-a3bf-391e2fb9db79-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a594d7fb-9b6e-4dee-a3bf-391e2fb9db79\") " pod="openstack/ceilometer-0" Nov 21 14:34:29 crc kubenswrapper[4897]: I1121 14:34:29.425239 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a594d7fb-9b6e-4dee-a3bf-391e2fb9db79-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a594d7fb-9b6e-4dee-a3bf-391e2fb9db79\") " pod="openstack/ceilometer-0" Nov 21 14:34:29 crc kubenswrapper[4897]: I1121 14:34:29.425266 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a594d7fb-9b6e-4dee-a3bf-391e2fb9db79-config-data\") pod \"ceilometer-0\" (UID: \"a594d7fb-9b6e-4dee-a3bf-391e2fb9db79\") " pod="openstack/ceilometer-0" Nov 21 14:34:29 crc kubenswrapper[4897]: I1121 14:34:29.425695 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9hg7j\" (UniqueName: \"kubernetes.io/projected/a594d7fb-9b6e-4dee-a3bf-391e2fb9db79-kube-api-access-9hg7j\") pod \"ceilometer-0\" (UID: \"a594d7fb-9b6e-4dee-a3bf-391e2fb9db79\") " pod="openstack/ceilometer-0" Nov 21 14:34:29 crc kubenswrapper[4897]: I1121 14:34:29.425840 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a594d7fb-9b6e-4dee-a3bf-391e2fb9db79-run-httpd\") pod \"ceilometer-0\" (UID: \"a594d7fb-9b6e-4dee-a3bf-391e2fb9db79\") " pod="openstack/ceilometer-0" Nov 21 14:34:29 crc kubenswrapper[4897]: I1121 14:34:29.527685 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a594d7fb-9b6e-4dee-a3bf-391e2fb9db79-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a594d7fb-9b6e-4dee-a3bf-391e2fb9db79\") " pod="openstack/ceilometer-0" Nov 21 14:34:29 crc kubenswrapper[4897]: I1121 14:34:29.527739 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a594d7fb-9b6e-4dee-a3bf-391e2fb9db79-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a594d7fb-9b6e-4dee-a3bf-391e2fb9db79\") " pod="openstack/ceilometer-0" Nov 21 14:34:29 crc kubenswrapper[4897]: I1121 14:34:29.527765 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a594d7fb-9b6e-4dee-a3bf-391e2fb9db79-config-data\") pod \"ceilometer-0\" (UID: \"a594d7fb-9b6e-4dee-a3bf-391e2fb9db79\") " pod="openstack/ceilometer-0" Nov 21 14:34:29 crc kubenswrapper[4897]: I1121 14:34:29.527875 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9hg7j\" (UniqueName: \"kubernetes.io/projected/a594d7fb-9b6e-4dee-a3bf-391e2fb9db79-kube-api-access-9hg7j\") pod \"ceilometer-0\" (UID: \"a594d7fb-9b6e-4dee-a3bf-391e2fb9db79\") " pod="openstack/ceilometer-0" Nov 21 14:34:29 crc kubenswrapper[4897]: I1121 14:34:29.527921 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a594d7fb-9b6e-4dee-a3bf-391e2fb9db79-run-httpd\") pod \"ceilometer-0\" (UID: \"a594d7fb-9b6e-4dee-a3bf-391e2fb9db79\") " pod="openstack/ceilometer-0" Nov 21 14:34:29 crc kubenswrapper[4897]: I1121 14:34:29.527962 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a594d7fb-9b6e-4dee-a3bf-391e2fb9db79-scripts\") pod \"ceilometer-0\" (UID: \"a594d7fb-9b6e-4dee-a3bf-391e2fb9db79\") " pod="openstack/ceilometer-0" Nov 21 14:34:29 crc kubenswrapper[4897]: I1121 14:34:29.527987 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a594d7fb-9b6e-4dee-a3bf-391e2fb9db79-log-httpd\") pod \"ceilometer-0\" (UID: \"a594d7fb-9b6e-4dee-a3bf-391e2fb9db79\") " pod="openstack/ceilometer-0" Nov 21 14:34:29 crc kubenswrapper[4897]: I1121 14:34:29.528415 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a594d7fb-9b6e-4dee-a3bf-391e2fb9db79-log-httpd\") pod \"ceilometer-0\" (UID: \"a594d7fb-9b6e-4dee-a3bf-391e2fb9db79\") " pod="openstack/ceilometer-0" Nov 21 14:34:29 crc kubenswrapper[4897]: I1121 14:34:29.528713 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a594d7fb-9b6e-4dee-a3bf-391e2fb9db79-run-httpd\") pod \"ceilometer-0\" (UID: \"a594d7fb-9b6e-4dee-a3bf-391e2fb9db79\") " pod="openstack/ceilometer-0" Nov 21 14:34:29 crc kubenswrapper[4897]: I1121 14:34:29.532558 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a594d7fb-9b6e-4dee-a3bf-391e2fb9db79-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a594d7fb-9b6e-4dee-a3bf-391e2fb9db79\") " pod="openstack/ceilometer-0" Nov 21 14:34:29 crc kubenswrapper[4897]: I1121 14:34:29.532863 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a594d7fb-9b6e-4dee-a3bf-391e2fb9db79-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a594d7fb-9b6e-4dee-a3bf-391e2fb9db79\") " pod="openstack/ceilometer-0" Nov 21 14:34:29 crc kubenswrapper[4897]: I1121 14:34:29.533621 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a594d7fb-9b6e-4dee-a3bf-391e2fb9db79-scripts\") pod \"ceilometer-0\" (UID: \"a594d7fb-9b6e-4dee-a3bf-391e2fb9db79\") " pod="openstack/ceilometer-0" Nov 21 14:34:29 crc kubenswrapper[4897]: I1121 14:34:29.535494 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a594d7fb-9b6e-4dee-a3bf-391e2fb9db79-config-data\") pod \"ceilometer-0\" (UID: \"a594d7fb-9b6e-4dee-a3bf-391e2fb9db79\") " pod="openstack/ceilometer-0" Nov 21 14:34:29 crc kubenswrapper[4897]: I1121 14:34:29.549349 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9hg7j\" (UniqueName: \"kubernetes.io/projected/a594d7fb-9b6e-4dee-a3bf-391e2fb9db79-kube-api-access-9hg7j\") pod \"ceilometer-0\" (UID: \"a594d7fb-9b6e-4dee-a3bf-391e2fb9db79\") " pod="openstack/ceilometer-0" Nov 21 14:34:29 crc kubenswrapper[4897]: I1121 14:34:29.699926 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 21 14:34:30 crc kubenswrapper[4897]: I1121 14:34:30.104424 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9b4dd223-2765-4cd7-92c6-293fadac9520" path="/var/lib/kubelet/pods/9b4dd223-2765-4cd7-92c6-293fadac9520/volumes" Nov 21 14:34:30 crc kubenswrapper[4897]: I1121 14:34:30.105457 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bac76d5f-def0-4fe8-babf-2c1a171f04dc" path="/var/lib/kubelet/pods/bac76d5f-def0-4fe8-babf-2c1a171f04dc/volumes" Nov 21 14:34:30 crc kubenswrapper[4897]: I1121 14:34:30.130687 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-create-jjdjs"] Nov 21 14:34:30 crc kubenswrapper[4897]: I1121 14:34:30.132658 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-jjdjs" Nov 21 14:34:30 crc kubenswrapper[4897]: I1121 14:34:30.143559 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0f93-account-create-54bcj"] Nov 21 14:34:30 crc kubenswrapper[4897]: I1121 14:34:30.144972 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0f93-account-create-54bcj" Nov 21 14:34:30 crc kubenswrapper[4897]: I1121 14:34:30.147135 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-db-secret" Nov 21 14:34:30 crc kubenswrapper[4897]: I1121 14:34:30.158094 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-jjdjs"] Nov 21 14:34:30 crc kubenswrapper[4897]: I1121 14:34:30.168148 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0f93-account-create-54bcj"] Nov 21 14:34:30 crc kubenswrapper[4897]: I1121 14:34:30.205830 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 21 14:34:30 crc kubenswrapper[4897]: W1121 14:34:30.208474 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda594d7fb_9b6e_4dee_a3bf_391e2fb9db79.slice/crio-2de62a1585b6cabd107dd73d8c96b32664ff679b31d9a581ea7a0d1c17da8d77 WatchSource:0}: Error finding container 2de62a1585b6cabd107dd73d8c96b32664ff679b31d9a581ea7a0d1c17da8d77: Status 404 returned error can't find the container with id 2de62a1585b6cabd107dd73d8c96b32664ff679b31d9a581ea7a0d1c17da8d77 Nov 21 14:34:30 crc kubenswrapper[4897]: I1121 14:34:30.252001 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/407be862-21b7-4f8d-9e62-81e6e8e40ac5-operator-scripts\") pod \"aodh-db-create-jjdjs\" (UID: \"407be862-21b7-4f8d-9e62-81e6e8e40ac5\") " pod="openstack/aodh-db-create-jjdjs" Nov 21 14:34:30 crc kubenswrapper[4897]: I1121 14:34:30.252262 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sp2p9\" (UniqueName: \"kubernetes.io/projected/407be862-21b7-4f8d-9e62-81e6e8e40ac5-kube-api-access-sp2p9\") pod \"aodh-db-create-jjdjs\" (UID: \"407be862-21b7-4f8d-9e62-81e6e8e40ac5\") " pod="openstack/aodh-db-create-jjdjs" Nov 21 14:34:30 crc kubenswrapper[4897]: I1121 14:34:30.252434 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7xlzk\" (UniqueName: \"kubernetes.io/projected/f2251225-86fd-4e3e-87fe-d4ac1db2ee8e-kube-api-access-7xlzk\") pod \"aodh-0f93-account-create-54bcj\" (UID: \"f2251225-86fd-4e3e-87fe-d4ac1db2ee8e\") " pod="openstack/aodh-0f93-account-create-54bcj" Nov 21 14:34:30 crc kubenswrapper[4897]: I1121 14:34:30.252608 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f2251225-86fd-4e3e-87fe-d4ac1db2ee8e-operator-scripts\") pod \"aodh-0f93-account-create-54bcj\" (UID: \"f2251225-86fd-4e3e-87fe-d4ac1db2ee8e\") " pod="openstack/aodh-0f93-account-create-54bcj" Nov 21 14:34:30 crc kubenswrapper[4897]: I1121 14:34:30.354056 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7xlzk\" (UniqueName: \"kubernetes.io/projected/f2251225-86fd-4e3e-87fe-d4ac1db2ee8e-kube-api-access-7xlzk\") pod \"aodh-0f93-account-create-54bcj\" (UID: \"f2251225-86fd-4e3e-87fe-d4ac1db2ee8e\") " pod="openstack/aodh-0f93-account-create-54bcj" Nov 21 14:34:30 crc kubenswrapper[4897]: I1121 14:34:30.354181 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f2251225-86fd-4e3e-87fe-d4ac1db2ee8e-operator-scripts\") pod \"aodh-0f93-account-create-54bcj\" (UID: \"f2251225-86fd-4e3e-87fe-d4ac1db2ee8e\") " pod="openstack/aodh-0f93-account-create-54bcj" Nov 21 14:34:30 crc kubenswrapper[4897]: I1121 14:34:30.354247 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/407be862-21b7-4f8d-9e62-81e6e8e40ac5-operator-scripts\") pod \"aodh-db-create-jjdjs\" (UID: \"407be862-21b7-4f8d-9e62-81e6e8e40ac5\") " pod="openstack/aodh-db-create-jjdjs" Nov 21 14:34:30 crc kubenswrapper[4897]: I1121 14:34:30.354299 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sp2p9\" (UniqueName: \"kubernetes.io/projected/407be862-21b7-4f8d-9e62-81e6e8e40ac5-kube-api-access-sp2p9\") pod \"aodh-db-create-jjdjs\" (UID: \"407be862-21b7-4f8d-9e62-81e6e8e40ac5\") " pod="openstack/aodh-db-create-jjdjs" Nov 21 14:34:30 crc kubenswrapper[4897]: I1121 14:34:30.355491 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f2251225-86fd-4e3e-87fe-d4ac1db2ee8e-operator-scripts\") pod \"aodh-0f93-account-create-54bcj\" (UID: \"f2251225-86fd-4e3e-87fe-d4ac1db2ee8e\") " pod="openstack/aodh-0f93-account-create-54bcj" Nov 21 14:34:30 crc kubenswrapper[4897]: I1121 14:34:30.355730 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/407be862-21b7-4f8d-9e62-81e6e8e40ac5-operator-scripts\") pod \"aodh-db-create-jjdjs\" (UID: \"407be862-21b7-4f8d-9e62-81e6e8e40ac5\") " pod="openstack/aodh-db-create-jjdjs" Nov 21 14:34:30 crc kubenswrapper[4897]: I1121 14:34:30.375256 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sp2p9\" (UniqueName: \"kubernetes.io/projected/407be862-21b7-4f8d-9e62-81e6e8e40ac5-kube-api-access-sp2p9\") pod \"aodh-db-create-jjdjs\" (UID: \"407be862-21b7-4f8d-9e62-81e6e8e40ac5\") " pod="openstack/aodh-db-create-jjdjs" Nov 21 14:34:30 crc kubenswrapper[4897]: I1121 14:34:30.375321 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7xlzk\" (UniqueName: \"kubernetes.io/projected/f2251225-86fd-4e3e-87fe-d4ac1db2ee8e-kube-api-access-7xlzk\") pod \"aodh-0f93-account-create-54bcj\" (UID: \"f2251225-86fd-4e3e-87fe-d4ac1db2ee8e\") " pod="openstack/aodh-0f93-account-create-54bcj" Nov 21 14:34:30 crc kubenswrapper[4897]: I1121 14:34:30.459293 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-jjdjs" Nov 21 14:34:30 crc kubenswrapper[4897]: I1121 14:34:30.467237 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0f93-account-create-54bcj" Nov 21 14:34:30 crc kubenswrapper[4897]: I1121 14:34:30.925195 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a594d7fb-9b6e-4dee-a3bf-391e2fb9db79","Type":"ContainerStarted","Data":"2de62a1585b6cabd107dd73d8c96b32664ff679b31d9a581ea7a0d1c17da8d77"} Nov 21 14:34:30 crc kubenswrapper[4897]: I1121 14:34:30.972466 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0f93-account-create-54bcj"] Nov 21 14:34:30 crc kubenswrapper[4897]: W1121 14:34:30.976787 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod407be862_21b7_4f8d_9e62_81e6e8e40ac5.slice/crio-65c6cf993d5d02830e2b838ca9b987be7b47dbeffdbf7c23425d8a05d24c60f0 WatchSource:0}: Error finding container 65c6cf993d5d02830e2b838ca9b987be7b47dbeffdbf7c23425d8a05d24c60f0: Status 404 returned error can't find the container with id 65c6cf993d5d02830e2b838ca9b987be7b47dbeffdbf7c23425d8a05d24c60f0 Nov 21 14:34:30 crc kubenswrapper[4897]: I1121 14:34:30.985477 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-jjdjs"] Nov 21 14:34:31 crc kubenswrapper[4897]: I1121 14:34:31.941112 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-jjdjs" event={"ID":"407be862-21b7-4f8d-9e62-81e6e8e40ac5","Type":"ContainerStarted","Data":"3723470e7b3e96c5e7a84947a48a7a6b9cf8050e5366b390cb2a6236ee20e58c"} Nov 21 14:34:31 crc kubenswrapper[4897]: I1121 14:34:31.941992 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-jjdjs" event={"ID":"407be862-21b7-4f8d-9e62-81e6e8e40ac5","Type":"ContainerStarted","Data":"65c6cf993d5d02830e2b838ca9b987be7b47dbeffdbf7c23425d8a05d24c60f0"} Nov 21 14:34:31 crc kubenswrapper[4897]: I1121 14:34:31.944875 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0f93-account-create-54bcj" event={"ID":"f2251225-86fd-4e3e-87fe-d4ac1db2ee8e","Type":"ContainerStarted","Data":"37527d0435969a408457ca6e56d44e25c335f7b2875e57166c3d10f6c123e1c7"} Nov 21 14:34:31 crc kubenswrapper[4897]: I1121 14:34:31.945046 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0f93-account-create-54bcj" event={"ID":"f2251225-86fd-4e3e-87fe-d4ac1db2ee8e","Type":"ContainerStarted","Data":"6bba7beb9c371251cb89b0344f6222e79d1da9b97f6a8735c50b07aacd471ad3"} Nov 21 14:34:31 crc kubenswrapper[4897]: I1121 14:34:31.961516 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-db-create-jjdjs" podStartSLOduration=1.9614961640000002 podStartE2EDuration="1.961496164s" podCreationTimestamp="2025-11-21 14:34:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:34:31.958858343 +0000 UTC m=+1549.243451828" watchObservedRunningTime="2025-11-21 14:34:31.961496164 +0000 UTC m=+1549.246089639" Nov 21 14:34:31 crc kubenswrapper[4897]: I1121 14:34:31.976076 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0f93-account-create-54bcj" podStartSLOduration=1.976059064 podStartE2EDuration="1.976059064s" podCreationTimestamp="2025-11-21 14:34:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:34:31.970776533 +0000 UTC m=+1549.255370008" watchObservedRunningTime="2025-11-21 14:34:31.976059064 +0000 UTC m=+1549.260652539" Nov 21 14:34:32 crc kubenswrapper[4897]: I1121 14:34:32.956546 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a594d7fb-9b6e-4dee-a3bf-391e2fb9db79","Type":"ContainerStarted","Data":"18c04b36a73b5151b088829484025696e39c00240447d79e11f023c996e82105"} Nov 21 14:34:33 crc kubenswrapper[4897]: I1121 14:34:33.977217 4897 generic.go:334] "Generic (PLEG): container finished" podID="407be862-21b7-4f8d-9e62-81e6e8e40ac5" containerID="3723470e7b3e96c5e7a84947a48a7a6b9cf8050e5366b390cb2a6236ee20e58c" exitCode=0 Nov 21 14:34:33 crc kubenswrapper[4897]: I1121 14:34:33.977329 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-jjdjs" event={"ID":"407be862-21b7-4f8d-9e62-81e6e8e40ac5","Type":"ContainerDied","Data":"3723470e7b3e96c5e7a84947a48a7a6b9cf8050e5366b390cb2a6236ee20e58c"} Nov 21 14:34:33 crc kubenswrapper[4897]: I1121 14:34:33.980069 4897 generic.go:334] "Generic (PLEG): container finished" podID="f2251225-86fd-4e3e-87fe-d4ac1db2ee8e" containerID="37527d0435969a408457ca6e56d44e25c335f7b2875e57166c3d10f6c123e1c7" exitCode=0 Nov 21 14:34:33 crc kubenswrapper[4897]: I1121 14:34:33.980127 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0f93-account-create-54bcj" event={"ID":"f2251225-86fd-4e3e-87fe-d4ac1db2ee8e","Type":"ContainerDied","Data":"37527d0435969a408457ca6e56d44e25c335f7b2875e57166c3d10f6c123e1c7"} Nov 21 14:34:34 crc kubenswrapper[4897]: I1121 14:34:34.370747 4897 patch_prober.go:28] interesting pod/machine-config-daemon-krv5b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 21 14:34:34 crc kubenswrapper[4897]: I1121 14:34:34.371105 4897 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 21 14:34:35 crc kubenswrapper[4897]: I1121 14:34:35.015589 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a594d7fb-9b6e-4dee-a3bf-391e2fb9db79","Type":"ContainerStarted","Data":"96a3c1e1449c8235b41c8198d4056d1b5f09d1fe21ac9b4de815013719e3bf16"} Nov 21 14:34:35 crc kubenswrapper[4897]: I1121 14:34:35.493179 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-jjdjs" Nov 21 14:34:35 crc kubenswrapper[4897]: I1121 14:34:35.499992 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0f93-account-create-54bcj" Nov 21 14:34:35 crc kubenswrapper[4897]: I1121 14:34:35.606968 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sp2p9\" (UniqueName: \"kubernetes.io/projected/407be862-21b7-4f8d-9e62-81e6e8e40ac5-kube-api-access-sp2p9\") pod \"407be862-21b7-4f8d-9e62-81e6e8e40ac5\" (UID: \"407be862-21b7-4f8d-9e62-81e6e8e40ac5\") " Nov 21 14:34:35 crc kubenswrapper[4897]: I1121 14:34:35.607389 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/407be862-21b7-4f8d-9e62-81e6e8e40ac5-operator-scripts\") pod \"407be862-21b7-4f8d-9e62-81e6e8e40ac5\" (UID: \"407be862-21b7-4f8d-9e62-81e6e8e40ac5\") " Nov 21 14:34:35 crc kubenswrapper[4897]: I1121 14:34:35.607459 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7xlzk\" (UniqueName: \"kubernetes.io/projected/f2251225-86fd-4e3e-87fe-d4ac1db2ee8e-kube-api-access-7xlzk\") pod \"f2251225-86fd-4e3e-87fe-d4ac1db2ee8e\" (UID: \"f2251225-86fd-4e3e-87fe-d4ac1db2ee8e\") " Nov 21 14:34:35 crc kubenswrapper[4897]: I1121 14:34:35.607601 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f2251225-86fd-4e3e-87fe-d4ac1db2ee8e-operator-scripts\") pod \"f2251225-86fd-4e3e-87fe-d4ac1db2ee8e\" (UID: \"f2251225-86fd-4e3e-87fe-d4ac1db2ee8e\") " Nov 21 14:34:35 crc kubenswrapper[4897]: I1121 14:34:35.608048 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/407be862-21b7-4f8d-9e62-81e6e8e40ac5-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "407be862-21b7-4f8d-9e62-81e6e8e40ac5" (UID: "407be862-21b7-4f8d-9e62-81e6e8e40ac5"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:34:35 crc kubenswrapper[4897]: I1121 14:34:35.608339 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f2251225-86fd-4e3e-87fe-d4ac1db2ee8e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f2251225-86fd-4e3e-87fe-d4ac1db2ee8e" (UID: "f2251225-86fd-4e3e-87fe-d4ac1db2ee8e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:34:35 crc kubenswrapper[4897]: I1121 14:34:35.615032 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/407be862-21b7-4f8d-9e62-81e6e8e40ac5-kube-api-access-sp2p9" (OuterVolumeSpecName: "kube-api-access-sp2p9") pod "407be862-21b7-4f8d-9e62-81e6e8e40ac5" (UID: "407be862-21b7-4f8d-9e62-81e6e8e40ac5"). InnerVolumeSpecName "kube-api-access-sp2p9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:34:35 crc kubenswrapper[4897]: I1121 14:34:35.620243 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f2251225-86fd-4e3e-87fe-d4ac1db2ee8e-kube-api-access-7xlzk" (OuterVolumeSpecName: "kube-api-access-7xlzk") pod "f2251225-86fd-4e3e-87fe-d4ac1db2ee8e" (UID: "f2251225-86fd-4e3e-87fe-d4ac1db2ee8e"). InnerVolumeSpecName "kube-api-access-7xlzk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:34:35 crc kubenswrapper[4897]: I1121 14:34:35.710343 4897 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f2251225-86fd-4e3e-87fe-d4ac1db2ee8e-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 21 14:34:35 crc kubenswrapper[4897]: I1121 14:34:35.710627 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sp2p9\" (UniqueName: \"kubernetes.io/projected/407be862-21b7-4f8d-9e62-81e6e8e40ac5-kube-api-access-sp2p9\") on node \"crc\" DevicePath \"\"" Nov 21 14:34:35 crc kubenswrapper[4897]: I1121 14:34:35.710844 4897 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/407be862-21b7-4f8d-9e62-81e6e8e40ac5-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 21 14:34:35 crc kubenswrapper[4897]: I1121 14:34:35.710928 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7xlzk\" (UniqueName: \"kubernetes.io/projected/f2251225-86fd-4e3e-87fe-d4ac1db2ee8e-kube-api-access-7xlzk\") on node \"crc\" DevicePath \"\"" Nov 21 14:34:36 crc kubenswrapper[4897]: I1121 14:34:36.031011 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-jjdjs" event={"ID":"407be862-21b7-4f8d-9e62-81e6e8e40ac5","Type":"ContainerDied","Data":"65c6cf993d5d02830e2b838ca9b987be7b47dbeffdbf7c23425d8a05d24c60f0"} Nov 21 14:34:36 crc kubenswrapper[4897]: I1121 14:34:36.031054 4897 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="65c6cf993d5d02830e2b838ca9b987be7b47dbeffdbf7c23425d8a05d24c60f0" Nov 21 14:34:36 crc kubenswrapper[4897]: I1121 14:34:36.031025 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-jjdjs" Nov 21 14:34:36 crc kubenswrapper[4897]: I1121 14:34:36.035094 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0f93-account-create-54bcj" event={"ID":"f2251225-86fd-4e3e-87fe-d4ac1db2ee8e","Type":"ContainerDied","Data":"6bba7beb9c371251cb89b0344f6222e79d1da9b97f6a8735c50b07aacd471ad3"} Nov 21 14:34:36 crc kubenswrapper[4897]: I1121 14:34:36.035127 4897 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6bba7beb9c371251cb89b0344f6222e79d1da9b97f6a8735c50b07aacd471ad3" Nov 21 14:34:36 crc kubenswrapper[4897]: I1121 14:34:36.035178 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0f93-account-create-54bcj" Nov 21 14:34:36 crc kubenswrapper[4897]: I1121 14:34:36.166197 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-dtwql" podUID="6d3cad20-fb17-4726-b8fa-9fd5303b5eea" containerName="registry-server" probeResult="failure" output=< Nov 21 14:34:36 crc kubenswrapper[4897]: timeout: failed to connect service ":50051" within 1s Nov 21 14:34:36 crc kubenswrapper[4897]: > Nov 21 14:34:38 crc kubenswrapper[4897]: I1121 14:34:38.067362 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a594d7fb-9b6e-4dee-a3bf-391e2fb9db79","Type":"ContainerStarted","Data":"901c18cc885e322e944f4be7f228752ce978bc1d221bc929b698dde78b6aac34"} Nov 21 14:34:39 crc kubenswrapper[4897]: I1121 14:34:39.081836 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a594d7fb-9b6e-4dee-a3bf-391e2fb9db79","Type":"ContainerStarted","Data":"c122f9b6dbb62121086467252bceab944be70a98a9aa396fe5b157c686d35c15"} Nov 21 14:34:39 crc kubenswrapper[4897]: I1121 14:34:39.083180 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 21 14:34:39 crc kubenswrapper[4897]: I1121 14:34:39.109407 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.774513174 podStartE2EDuration="10.109384643s" podCreationTimestamp="2025-11-21 14:34:29 +0000 UTC" firstStartedPulling="2025-11-21 14:34:30.211091892 +0000 UTC m=+1547.495685367" lastFinishedPulling="2025-11-21 14:34:38.545963351 +0000 UTC m=+1555.830556836" observedRunningTime="2025-11-21 14:34:39.104322647 +0000 UTC m=+1556.388916132" watchObservedRunningTime="2025-11-21 14:34:39.109384643 +0000 UTC m=+1556.393978118" Nov 21 14:34:40 crc kubenswrapper[4897]: I1121 14:34:40.421053 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-sync-ctn4r"] Nov 21 14:34:40 crc kubenswrapper[4897]: E1121 14:34:40.422001 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="407be862-21b7-4f8d-9e62-81e6e8e40ac5" containerName="mariadb-database-create" Nov 21 14:34:40 crc kubenswrapper[4897]: I1121 14:34:40.422018 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="407be862-21b7-4f8d-9e62-81e6e8e40ac5" containerName="mariadb-database-create" Nov 21 14:34:40 crc kubenswrapper[4897]: E1121 14:34:40.422040 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2251225-86fd-4e3e-87fe-d4ac1db2ee8e" containerName="mariadb-account-create" Nov 21 14:34:40 crc kubenswrapper[4897]: I1121 14:34:40.422048 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2251225-86fd-4e3e-87fe-d4ac1db2ee8e" containerName="mariadb-account-create" Nov 21 14:34:40 crc kubenswrapper[4897]: I1121 14:34:40.422280 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="f2251225-86fd-4e3e-87fe-d4ac1db2ee8e" containerName="mariadb-account-create" Nov 21 14:34:40 crc kubenswrapper[4897]: I1121 14:34:40.422304 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="407be862-21b7-4f8d-9e62-81e6e8e40ac5" containerName="mariadb-database-create" Nov 21 14:34:40 crc kubenswrapper[4897]: I1121 14:34:40.423256 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-ctn4r" Nov 21 14:34:40 crc kubenswrapper[4897]: I1121 14:34:40.425595 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c729e1a-4290-4cbf-b38b-d235f229f166-config-data\") pod \"aodh-db-sync-ctn4r\" (UID: \"4c729e1a-4290-4cbf-b38b-d235f229f166\") " pod="openstack/aodh-db-sync-ctn4r" Nov 21 14:34:40 crc kubenswrapper[4897]: I1121 14:34:40.425793 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c729e1a-4290-4cbf-b38b-d235f229f166-scripts\") pod \"aodh-db-sync-ctn4r\" (UID: \"4c729e1a-4290-4cbf-b38b-d235f229f166\") " pod="openstack/aodh-db-sync-ctn4r" Nov 21 14:34:40 crc kubenswrapper[4897]: I1121 14:34:40.425896 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v2gm7\" (UniqueName: \"kubernetes.io/projected/4c729e1a-4290-4cbf-b38b-d235f229f166-kube-api-access-v2gm7\") pod \"aodh-db-sync-ctn4r\" (UID: \"4c729e1a-4290-4cbf-b38b-d235f229f166\") " pod="openstack/aodh-db-sync-ctn4r" Nov 21 14:34:40 crc kubenswrapper[4897]: I1121 14:34:40.425937 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c729e1a-4290-4cbf-b38b-d235f229f166-combined-ca-bundle\") pod \"aodh-db-sync-ctn4r\" (UID: \"4c729e1a-4290-4cbf-b38b-d235f229f166\") " pod="openstack/aodh-db-sync-ctn4r" Nov 21 14:34:40 crc kubenswrapper[4897]: I1121 14:34:40.432239 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-m57c8" Nov 21 14:34:40 crc kubenswrapper[4897]: I1121 14:34:40.432780 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Nov 21 14:34:40 crc kubenswrapper[4897]: I1121 14:34:40.432981 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Nov 21 14:34:40 crc kubenswrapper[4897]: I1121 14:34:40.443732 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 21 14:34:40 crc kubenswrapper[4897]: I1121 14:34:40.454200 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-ctn4r"] Nov 21 14:34:40 crc kubenswrapper[4897]: I1121 14:34:40.528764 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c729e1a-4290-4cbf-b38b-d235f229f166-scripts\") pod \"aodh-db-sync-ctn4r\" (UID: \"4c729e1a-4290-4cbf-b38b-d235f229f166\") " pod="openstack/aodh-db-sync-ctn4r" Nov 21 14:34:40 crc kubenswrapper[4897]: I1121 14:34:40.528905 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v2gm7\" (UniqueName: \"kubernetes.io/projected/4c729e1a-4290-4cbf-b38b-d235f229f166-kube-api-access-v2gm7\") pod \"aodh-db-sync-ctn4r\" (UID: \"4c729e1a-4290-4cbf-b38b-d235f229f166\") " pod="openstack/aodh-db-sync-ctn4r" Nov 21 14:34:40 crc kubenswrapper[4897]: I1121 14:34:40.528944 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c729e1a-4290-4cbf-b38b-d235f229f166-combined-ca-bundle\") pod \"aodh-db-sync-ctn4r\" (UID: \"4c729e1a-4290-4cbf-b38b-d235f229f166\") " pod="openstack/aodh-db-sync-ctn4r" Nov 21 14:34:40 crc kubenswrapper[4897]: I1121 14:34:40.528997 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c729e1a-4290-4cbf-b38b-d235f229f166-config-data\") pod \"aodh-db-sync-ctn4r\" (UID: \"4c729e1a-4290-4cbf-b38b-d235f229f166\") " pod="openstack/aodh-db-sync-ctn4r" Nov 21 14:34:40 crc kubenswrapper[4897]: I1121 14:34:40.541824 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c729e1a-4290-4cbf-b38b-d235f229f166-combined-ca-bundle\") pod \"aodh-db-sync-ctn4r\" (UID: \"4c729e1a-4290-4cbf-b38b-d235f229f166\") " pod="openstack/aodh-db-sync-ctn4r" Nov 21 14:34:40 crc kubenswrapper[4897]: I1121 14:34:40.542448 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c729e1a-4290-4cbf-b38b-d235f229f166-config-data\") pod \"aodh-db-sync-ctn4r\" (UID: \"4c729e1a-4290-4cbf-b38b-d235f229f166\") " pod="openstack/aodh-db-sync-ctn4r" Nov 21 14:34:40 crc kubenswrapper[4897]: I1121 14:34:40.549397 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c729e1a-4290-4cbf-b38b-d235f229f166-scripts\") pod \"aodh-db-sync-ctn4r\" (UID: \"4c729e1a-4290-4cbf-b38b-d235f229f166\") " pod="openstack/aodh-db-sync-ctn4r" Nov 21 14:34:40 crc kubenswrapper[4897]: I1121 14:34:40.550772 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v2gm7\" (UniqueName: \"kubernetes.io/projected/4c729e1a-4290-4cbf-b38b-d235f229f166-kube-api-access-v2gm7\") pod \"aodh-db-sync-ctn4r\" (UID: \"4c729e1a-4290-4cbf-b38b-d235f229f166\") " pod="openstack/aodh-db-sync-ctn4r" Nov 21 14:34:40 crc kubenswrapper[4897]: I1121 14:34:40.759226 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-ctn4r" Nov 21 14:34:41 crc kubenswrapper[4897]: I1121 14:34:41.266083 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-ctn4r"] Nov 21 14:34:41 crc kubenswrapper[4897]: W1121 14:34:41.277247 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4c729e1a_4290_4cbf_b38b_d235f229f166.slice/crio-bd623e222a868e0e08b2ef59fb61705198b7f3169efc08e587c0bd93b14429f0 WatchSource:0}: Error finding container bd623e222a868e0e08b2ef59fb61705198b7f3169efc08e587c0bd93b14429f0: Status 404 returned error can't find the container with id bd623e222a868e0e08b2ef59fb61705198b7f3169efc08e587c0bd93b14429f0 Nov 21 14:34:42 crc kubenswrapper[4897]: I1121 14:34:42.127141 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-ctn4r" event={"ID":"4c729e1a-4290-4cbf-b38b-d235f229f166","Type":"ContainerStarted","Data":"bd623e222a868e0e08b2ef59fb61705198b7f3169efc08e587c0bd93b14429f0"} Nov 21 14:34:46 crc kubenswrapper[4897]: I1121 14:34:46.161337 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-dtwql" podUID="6d3cad20-fb17-4726-b8fa-9fd5303b5eea" containerName="registry-server" probeResult="failure" output=< Nov 21 14:34:46 crc kubenswrapper[4897]: timeout: failed to connect service ":50051" within 1s Nov 21 14:34:46 crc kubenswrapper[4897]: > Nov 21 14:34:50 crc kubenswrapper[4897]: I1121 14:34:50.303103 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 21 14:34:51 crc kubenswrapper[4897]: I1121 14:34:51.283095 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-ctn4r" event={"ID":"4c729e1a-4290-4cbf-b38b-d235f229f166","Type":"ContainerStarted","Data":"848d3bf026757c3eac458bc2ae8f7730ed6a47c48b0f7cf0f0de11da3cf9f147"} Nov 21 14:34:51 crc kubenswrapper[4897]: I1121 14:34:51.299726 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-db-sync-ctn4r" podStartSLOduration=2.282673731 podStartE2EDuration="11.299709587s" podCreationTimestamp="2025-11-21 14:34:40 +0000 UTC" firstStartedPulling="2025-11-21 14:34:41.282835602 +0000 UTC m=+1558.567429077" lastFinishedPulling="2025-11-21 14:34:50.299871458 +0000 UTC m=+1567.584464933" observedRunningTime="2025-11-21 14:34:51.296851211 +0000 UTC m=+1568.581444696" watchObservedRunningTime="2025-11-21 14:34:51.299709587 +0000 UTC m=+1568.584303062" Nov 21 14:34:56 crc kubenswrapper[4897]: I1121 14:34:56.178513 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-dtwql" podUID="6d3cad20-fb17-4726-b8fa-9fd5303b5eea" containerName="registry-server" probeResult="failure" output=< Nov 21 14:34:56 crc kubenswrapper[4897]: timeout: failed to connect service ":50051" within 1s Nov 21 14:34:56 crc kubenswrapper[4897]: > Nov 21 14:34:58 crc kubenswrapper[4897]: I1121 14:34:58.365588 4897 generic.go:334] "Generic (PLEG): container finished" podID="4c729e1a-4290-4cbf-b38b-d235f229f166" containerID="848d3bf026757c3eac458bc2ae8f7730ed6a47c48b0f7cf0f0de11da3cf9f147" exitCode=0 Nov 21 14:34:58 crc kubenswrapper[4897]: I1121 14:34:58.365675 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-ctn4r" event={"ID":"4c729e1a-4290-4cbf-b38b-d235f229f166","Type":"ContainerDied","Data":"848d3bf026757c3eac458bc2ae8f7730ed6a47c48b0f7cf0f0de11da3cf9f147"} Nov 21 14:34:59 crc kubenswrapper[4897]: I1121 14:34:59.725145 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 21 14:34:59 crc kubenswrapper[4897]: I1121 14:34:59.825114 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-ctn4r" Nov 21 14:34:59 crc kubenswrapper[4897]: I1121 14:34:59.935634 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c729e1a-4290-4cbf-b38b-d235f229f166-scripts\") pod \"4c729e1a-4290-4cbf-b38b-d235f229f166\" (UID: \"4c729e1a-4290-4cbf-b38b-d235f229f166\") " Nov 21 14:34:59 crc kubenswrapper[4897]: I1121 14:34:59.935740 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c729e1a-4290-4cbf-b38b-d235f229f166-combined-ca-bundle\") pod \"4c729e1a-4290-4cbf-b38b-d235f229f166\" (UID: \"4c729e1a-4290-4cbf-b38b-d235f229f166\") " Nov 21 14:34:59 crc kubenswrapper[4897]: I1121 14:34:59.935819 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c729e1a-4290-4cbf-b38b-d235f229f166-config-data\") pod \"4c729e1a-4290-4cbf-b38b-d235f229f166\" (UID: \"4c729e1a-4290-4cbf-b38b-d235f229f166\") " Nov 21 14:34:59 crc kubenswrapper[4897]: I1121 14:34:59.935956 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v2gm7\" (UniqueName: \"kubernetes.io/projected/4c729e1a-4290-4cbf-b38b-d235f229f166-kube-api-access-v2gm7\") pod \"4c729e1a-4290-4cbf-b38b-d235f229f166\" (UID: \"4c729e1a-4290-4cbf-b38b-d235f229f166\") " Nov 21 14:34:59 crc kubenswrapper[4897]: I1121 14:34:59.949380 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c729e1a-4290-4cbf-b38b-d235f229f166-kube-api-access-v2gm7" (OuterVolumeSpecName: "kube-api-access-v2gm7") pod "4c729e1a-4290-4cbf-b38b-d235f229f166" (UID: "4c729e1a-4290-4cbf-b38b-d235f229f166"). InnerVolumeSpecName "kube-api-access-v2gm7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:34:59 crc kubenswrapper[4897]: I1121 14:34:59.950191 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c729e1a-4290-4cbf-b38b-d235f229f166-scripts" (OuterVolumeSpecName: "scripts") pod "4c729e1a-4290-4cbf-b38b-d235f229f166" (UID: "4c729e1a-4290-4cbf-b38b-d235f229f166"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:34:59 crc kubenswrapper[4897]: I1121 14:34:59.974863 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c729e1a-4290-4cbf-b38b-d235f229f166-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4c729e1a-4290-4cbf-b38b-d235f229f166" (UID: "4c729e1a-4290-4cbf-b38b-d235f229f166"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:34:59 crc kubenswrapper[4897]: I1121 14:34:59.988715 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c729e1a-4290-4cbf-b38b-d235f229f166-config-data" (OuterVolumeSpecName: "config-data") pod "4c729e1a-4290-4cbf-b38b-d235f229f166" (UID: "4c729e1a-4290-4cbf-b38b-d235f229f166"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:35:00 crc kubenswrapper[4897]: I1121 14:35:00.038475 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v2gm7\" (UniqueName: \"kubernetes.io/projected/4c729e1a-4290-4cbf-b38b-d235f229f166-kube-api-access-v2gm7\") on node \"crc\" DevicePath \"\"" Nov 21 14:35:00 crc kubenswrapper[4897]: I1121 14:35:00.038535 4897 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c729e1a-4290-4cbf-b38b-d235f229f166-scripts\") on node \"crc\" DevicePath \"\"" Nov 21 14:35:00 crc kubenswrapper[4897]: I1121 14:35:00.038551 4897 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c729e1a-4290-4cbf-b38b-d235f229f166-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 21 14:35:00 crc kubenswrapper[4897]: I1121 14:35:00.038561 4897 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c729e1a-4290-4cbf-b38b-d235f229f166-config-data\") on node \"crc\" DevicePath \"\"" Nov 21 14:35:00 crc kubenswrapper[4897]: I1121 14:35:00.398394 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-ctn4r" event={"ID":"4c729e1a-4290-4cbf-b38b-d235f229f166","Type":"ContainerDied","Data":"bd623e222a868e0e08b2ef59fb61705198b7f3169efc08e587c0bd93b14429f0"} Nov 21 14:35:00 crc kubenswrapper[4897]: I1121 14:35:00.398500 4897 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bd623e222a868e0e08b2ef59fb61705198b7f3169efc08e587c0bd93b14429f0" Nov 21 14:35:00 crc kubenswrapper[4897]: I1121 14:35:00.399042 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-ctn4r" Nov 21 14:35:04 crc kubenswrapper[4897]: I1121 14:35:04.371331 4897 patch_prober.go:28] interesting pod/machine-config-daemon-krv5b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 21 14:35:04 crc kubenswrapper[4897]: I1121 14:35:04.372004 4897 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 21 14:35:04 crc kubenswrapper[4897]: I1121 14:35:04.372054 4897 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" Nov 21 14:35:04 crc kubenswrapper[4897]: I1121 14:35:04.373119 4897 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c68f1b003d12b4a2964169a85a04e3f0272d86c891f52e6bebd52f3781520d1a"} pod="openshift-machine-config-operator/machine-config-daemon-krv5b" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 21 14:35:04 crc kubenswrapper[4897]: I1121 14:35:04.373178 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" containerName="machine-config-daemon" containerID="cri-o://c68f1b003d12b4a2964169a85a04e3f0272d86c891f52e6bebd52f3781520d1a" gracePeriod=600 Nov 21 14:35:05 crc kubenswrapper[4897]: E1121 14:35:05.084679 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 14:35:05 crc kubenswrapper[4897]: I1121 14:35:05.132177 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Nov 21 14:35:05 crc kubenswrapper[4897]: E1121 14:35:05.135885 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c729e1a-4290-4cbf-b38b-d235f229f166" containerName="aodh-db-sync" Nov 21 14:35:05 crc kubenswrapper[4897]: I1121 14:35:05.135911 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c729e1a-4290-4cbf-b38b-d235f229f166" containerName="aodh-db-sync" Nov 21 14:35:05 crc kubenswrapper[4897]: I1121 14:35:05.136146 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c729e1a-4290-4cbf-b38b-d235f229f166" containerName="aodh-db-sync" Nov 21 14:35:05 crc kubenswrapper[4897]: I1121 14:35:05.138304 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Nov 21 14:35:05 crc kubenswrapper[4897]: I1121 14:35:05.143622 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Nov 21 14:35:05 crc kubenswrapper[4897]: I1121 14:35:05.143806 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-m57c8" Nov 21 14:35:05 crc kubenswrapper[4897]: I1121 14:35:05.144028 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Nov 21 14:35:05 crc kubenswrapper[4897]: I1121 14:35:05.167596 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Nov 21 14:35:05 crc kubenswrapper[4897]: I1121 14:35:05.271224 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53e2a389-78b0-463a-82df-dc3d1d870b7b-scripts\") pod \"aodh-0\" (UID: \"53e2a389-78b0-463a-82df-dc3d1d870b7b\") " pod="openstack/aodh-0" Nov 21 14:35:05 crc kubenswrapper[4897]: I1121 14:35:05.271400 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53e2a389-78b0-463a-82df-dc3d1d870b7b-combined-ca-bundle\") pod \"aodh-0\" (UID: \"53e2a389-78b0-463a-82df-dc3d1d870b7b\") " pod="openstack/aodh-0" Nov 21 14:35:05 crc kubenswrapper[4897]: I1121 14:35:05.271528 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hmg45\" (UniqueName: \"kubernetes.io/projected/53e2a389-78b0-463a-82df-dc3d1d870b7b-kube-api-access-hmg45\") pod \"aodh-0\" (UID: \"53e2a389-78b0-463a-82df-dc3d1d870b7b\") " pod="openstack/aodh-0" Nov 21 14:35:05 crc kubenswrapper[4897]: I1121 14:35:05.271641 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53e2a389-78b0-463a-82df-dc3d1d870b7b-config-data\") pod \"aodh-0\" (UID: \"53e2a389-78b0-463a-82df-dc3d1d870b7b\") " pod="openstack/aodh-0" Nov 21 14:35:05 crc kubenswrapper[4897]: I1121 14:35:05.373631 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53e2a389-78b0-463a-82df-dc3d1d870b7b-config-data\") pod \"aodh-0\" (UID: \"53e2a389-78b0-463a-82df-dc3d1d870b7b\") " pod="openstack/aodh-0" Nov 21 14:35:05 crc kubenswrapper[4897]: I1121 14:35:05.373737 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53e2a389-78b0-463a-82df-dc3d1d870b7b-scripts\") pod \"aodh-0\" (UID: \"53e2a389-78b0-463a-82df-dc3d1d870b7b\") " pod="openstack/aodh-0" Nov 21 14:35:05 crc kubenswrapper[4897]: I1121 14:35:05.373860 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53e2a389-78b0-463a-82df-dc3d1d870b7b-combined-ca-bundle\") pod \"aodh-0\" (UID: \"53e2a389-78b0-463a-82df-dc3d1d870b7b\") " pod="openstack/aodh-0" Nov 21 14:35:05 crc kubenswrapper[4897]: I1121 14:35:05.373932 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hmg45\" (UniqueName: \"kubernetes.io/projected/53e2a389-78b0-463a-82df-dc3d1d870b7b-kube-api-access-hmg45\") pod \"aodh-0\" (UID: \"53e2a389-78b0-463a-82df-dc3d1d870b7b\") " pod="openstack/aodh-0" Nov 21 14:35:05 crc kubenswrapper[4897]: I1121 14:35:05.379664 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53e2a389-78b0-463a-82df-dc3d1d870b7b-scripts\") pod \"aodh-0\" (UID: \"53e2a389-78b0-463a-82df-dc3d1d870b7b\") " pod="openstack/aodh-0" Nov 21 14:35:05 crc kubenswrapper[4897]: I1121 14:35:05.380138 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53e2a389-78b0-463a-82df-dc3d1d870b7b-combined-ca-bundle\") pod \"aodh-0\" (UID: \"53e2a389-78b0-463a-82df-dc3d1d870b7b\") " pod="openstack/aodh-0" Nov 21 14:35:05 crc kubenswrapper[4897]: I1121 14:35:05.383409 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53e2a389-78b0-463a-82df-dc3d1d870b7b-config-data\") pod \"aodh-0\" (UID: \"53e2a389-78b0-463a-82df-dc3d1d870b7b\") " pod="openstack/aodh-0" Nov 21 14:35:05 crc kubenswrapper[4897]: I1121 14:35:05.405460 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hmg45\" (UniqueName: \"kubernetes.io/projected/53e2a389-78b0-463a-82df-dc3d1d870b7b-kube-api-access-hmg45\") pod \"aodh-0\" (UID: \"53e2a389-78b0-463a-82df-dc3d1d870b7b\") " pod="openstack/aodh-0" Nov 21 14:35:05 crc kubenswrapper[4897]: I1121 14:35:05.467328 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Nov 21 14:35:05 crc kubenswrapper[4897]: I1121 14:35:05.468847 4897 generic.go:334] "Generic (PLEG): container finished" podID="e7670227-d280-4847-b882-754429f56b0f" containerID="c68f1b003d12b4a2964169a85a04e3f0272d86c891f52e6bebd52f3781520d1a" exitCode=0 Nov 21 14:35:05 crc kubenswrapper[4897]: I1121 14:35:05.468897 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" event={"ID":"e7670227-d280-4847-b882-754429f56b0f","Type":"ContainerDied","Data":"c68f1b003d12b4a2964169a85a04e3f0272d86c891f52e6bebd52f3781520d1a"} Nov 21 14:35:05 crc kubenswrapper[4897]: I1121 14:35:05.468945 4897 scope.go:117] "RemoveContainer" containerID="fbff5a4c19b1cfbe0c7e57eb05f0093aa3829fbabee820a190f916e0ce4fe6b9" Nov 21 14:35:05 crc kubenswrapper[4897]: I1121 14:35:05.469498 4897 scope.go:117] "RemoveContainer" containerID="c68f1b003d12b4a2964169a85a04e3f0272d86c891f52e6bebd52f3781520d1a" Nov 21 14:35:05 crc kubenswrapper[4897]: E1121 14:35:05.469951 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 14:35:05 crc kubenswrapper[4897]: I1121 14:35:05.718309 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 21 14:35:05 crc kubenswrapper[4897]: I1121 14:35:05.718850 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="95ad29e4-b4b6-4cec-b978-23f7ea25edc6" containerName="kube-state-metrics" containerID="cri-o://b82c8eaa87df5cc93b49764cd8875d1ee658727483ebf80e00a24ca9b4bece30" gracePeriod=30 Nov 21 14:35:05 crc kubenswrapper[4897]: I1121 14:35:05.807263 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mysqld-exporter-0"] Nov 21 14:35:05 crc kubenswrapper[4897]: I1121 14:35:05.807805 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/mysqld-exporter-0" podUID="813eae15-fec1-4d3a-bae2-1dc597c528f7" containerName="mysqld-exporter" containerID="cri-o://73c0e396184e52ce4ab802d647e0f9d82f1f11f7f86b7c1d9df80065526d91e9" gracePeriod=30 Nov 21 14:35:06 crc kubenswrapper[4897]: I1121 14:35:06.036716 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Nov 21 14:35:06 crc kubenswrapper[4897]: I1121 14:35:06.054161 4897 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 21 14:35:06 crc kubenswrapper[4897]: I1121 14:35:06.211089 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-dtwql" podUID="6d3cad20-fb17-4726-b8fa-9fd5303b5eea" containerName="registry-server" probeResult="failure" output=< Nov 21 14:35:06 crc kubenswrapper[4897]: timeout: failed to connect service ":50051" within 1s Nov 21 14:35:06 crc kubenswrapper[4897]: > Nov 21 14:35:06 crc kubenswrapper[4897]: I1121 14:35:06.334976 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 21 14:35:06 crc kubenswrapper[4897]: I1121 14:35:06.487107 4897 generic.go:334] "Generic (PLEG): container finished" podID="95ad29e4-b4b6-4cec-b978-23f7ea25edc6" containerID="b82c8eaa87df5cc93b49764cd8875d1ee658727483ebf80e00a24ca9b4bece30" exitCode=2 Nov 21 14:35:06 crc kubenswrapper[4897]: I1121 14:35:06.487165 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 21 14:35:06 crc kubenswrapper[4897]: I1121 14:35:06.487180 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"95ad29e4-b4b6-4cec-b978-23f7ea25edc6","Type":"ContainerDied","Data":"b82c8eaa87df5cc93b49764cd8875d1ee658727483ebf80e00a24ca9b4bece30"} Nov 21 14:35:06 crc kubenswrapper[4897]: I1121 14:35:06.487208 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"95ad29e4-b4b6-4cec-b978-23f7ea25edc6","Type":"ContainerDied","Data":"f02f019457b52e5712c17087ebfe728d1d845fff91f68d9f881356575fb8cb4a"} Nov 21 14:35:06 crc kubenswrapper[4897]: I1121 14:35:06.487224 4897 scope.go:117] "RemoveContainer" containerID="b82c8eaa87df5cc93b49764cd8875d1ee658727483ebf80e00a24ca9b4bece30" Nov 21 14:35:06 crc kubenswrapper[4897]: I1121 14:35:06.490408 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0" Nov 21 14:35:06 crc kubenswrapper[4897]: I1121 14:35:06.491827 4897 generic.go:334] "Generic (PLEG): container finished" podID="813eae15-fec1-4d3a-bae2-1dc597c528f7" containerID="73c0e396184e52ce4ab802d647e0f9d82f1f11f7f86b7c1d9df80065526d91e9" exitCode=2 Nov 21 14:35:06 crc kubenswrapper[4897]: I1121 14:35:06.491895 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0" event={"ID":"813eae15-fec1-4d3a-bae2-1dc597c528f7","Type":"ContainerDied","Data":"73c0e396184e52ce4ab802d647e0f9d82f1f11f7f86b7c1d9df80065526d91e9"} Nov 21 14:35:06 crc kubenswrapper[4897]: I1121 14:35:06.492904 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"53e2a389-78b0-463a-82df-dc3d1d870b7b","Type":"ContainerStarted","Data":"71192c01fc3c53970c06f13f68ea3d7f4f6667b0812854d8bae5a4272c250d4b"} Nov 21 14:35:06 crc kubenswrapper[4897]: I1121 14:35:06.505292 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pnbx9\" (UniqueName: \"kubernetes.io/projected/95ad29e4-b4b6-4cec-b978-23f7ea25edc6-kube-api-access-pnbx9\") pod \"95ad29e4-b4b6-4cec-b978-23f7ea25edc6\" (UID: \"95ad29e4-b4b6-4cec-b978-23f7ea25edc6\") " Nov 21 14:35:06 crc kubenswrapper[4897]: I1121 14:35:06.517259 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95ad29e4-b4b6-4cec-b978-23f7ea25edc6-kube-api-access-pnbx9" (OuterVolumeSpecName: "kube-api-access-pnbx9") pod "95ad29e4-b4b6-4cec-b978-23f7ea25edc6" (UID: "95ad29e4-b4b6-4cec-b978-23f7ea25edc6"). InnerVolumeSpecName "kube-api-access-pnbx9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:35:06 crc kubenswrapper[4897]: I1121 14:35:06.558473 4897 scope.go:117] "RemoveContainer" containerID="b82c8eaa87df5cc93b49764cd8875d1ee658727483ebf80e00a24ca9b4bece30" Nov 21 14:35:06 crc kubenswrapper[4897]: E1121 14:35:06.559189 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b82c8eaa87df5cc93b49764cd8875d1ee658727483ebf80e00a24ca9b4bece30\": container with ID starting with b82c8eaa87df5cc93b49764cd8875d1ee658727483ebf80e00a24ca9b4bece30 not found: ID does not exist" containerID="b82c8eaa87df5cc93b49764cd8875d1ee658727483ebf80e00a24ca9b4bece30" Nov 21 14:35:06 crc kubenswrapper[4897]: I1121 14:35:06.559216 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b82c8eaa87df5cc93b49764cd8875d1ee658727483ebf80e00a24ca9b4bece30"} err="failed to get container status \"b82c8eaa87df5cc93b49764cd8875d1ee658727483ebf80e00a24ca9b4bece30\": rpc error: code = NotFound desc = could not find container \"b82c8eaa87df5cc93b49764cd8875d1ee658727483ebf80e00a24ca9b4bece30\": container with ID starting with b82c8eaa87df5cc93b49764cd8875d1ee658727483ebf80e00a24ca9b4bece30 not found: ID does not exist" Nov 21 14:35:06 crc kubenswrapper[4897]: I1121 14:35:06.559249 4897 scope.go:117] "RemoveContainer" containerID="73c0e396184e52ce4ab802d647e0f9d82f1f11f7f86b7c1d9df80065526d91e9" Nov 21 14:35:06 crc kubenswrapper[4897]: I1121 14:35:06.609152 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/813eae15-fec1-4d3a-bae2-1dc597c528f7-config-data\") pod \"813eae15-fec1-4d3a-bae2-1dc597c528f7\" (UID: \"813eae15-fec1-4d3a-bae2-1dc597c528f7\") " Nov 21 14:35:06 crc kubenswrapper[4897]: I1121 14:35:06.609264 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dvt2r\" (UniqueName: \"kubernetes.io/projected/813eae15-fec1-4d3a-bae2-1dc597c528f7-kube-api-access-dvt2r\") pod \"813eae15-fec1-4d3a-bae2-1dc597c528f7\" (UID: \"813eae15-fec1-4d3a-bae2-1dc597c528f7\") " Nov 21 14:35:06 crc kubenswrapper[4897]: I1121 14:35:06.609871 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/813eae15-fec1-4d3a-bae2-1dc597c528f7-combined-ca-bundle\") pod \"813eae15-fec1-4d3a-bae2-1dc597c528f7\" (UID: \"813eae15-fec1-4d3a-bae2-1dc597c528f7\") " Nov 21 14:35:06 crc kubenswrapper[4897]: I1121 14:35:06.611904 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pnbx9\" (UniqueName: \"kubernetes.io/projected/95ad29e4-b4b6-4cec-b978-23f7ea25edc6-kube-api-access-pnbx9\") on node \"crc\" DevicePath \"\"" Nov 21 14:35:06 crc kubenswrapper[4897]: I1121 14:35:06.617736 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/813eae15-fec1-4d3a-bae2-1dc597c528f7-kube-api-access-dvt2r" (OuterVolumeSpecName: "kube-api-access-dvt2r") pod "813eae15-fec1-4d3a-bae2-1dc597c528f7" (UID: "813eae15-fec1-4d3a-bae2-1dc597c528f7"). InnerVolumeSpecName "kube-api-access-dvt2r". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:35:06 crc kubenswrapper[4897]: I1121 14:35:06.663759 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/813eae15-fec1-4d3a-bae2-1dc597c528f7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "813eae15-fec1-4d3a-bae2-1dc597c528f7" (UID: "813eae15-fec1-4d3a-bae2-1dc597c528f7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:35:06 crc kubenswrapper[4897]: I1121 14:35:06.708696 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/813eae15-fec1-4d3a-bae2-1dc597c528f7-config-data" (OuterVolumeSpecName: "config-data") pod "813eae15-fec1-4d3a-bae2-1dc597c528f7" (UID: "813eae15-fec1-4d3a-bae2-1dc597c528f7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:35:06 crc kubenswrapper[4897]: I1121 14:35:06.719941 4897 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/813eae15-fec1-4d3a-bae2-1dc597c528f7-config-data\") on node \"crc\" DevicePath \"\"" Nov 21 14:35:06 crc kubenswrapper[4897]: I1121 14:35:06.719997 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dvt2r\" (UniqueName: \"kubernetes.io/projected/813eae15-fec1-4d3a-bae2-1dc597c528f7-kube-api-access-dvt2r\") on node \"crc\" DevicePath \"\"" Nov 21 14:35:06 crc kubenswrapper[4897]: I1121 14:35:06.720014 4897 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/813eae15-fec1-4d3a-bae2-1dc597c528f7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 21 14:35:06 crc kubenswrapper[4897]: I1121 14:35:06.903581 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 21 14:35:06 crc kubenswrapper[4897]: I1121 14:35:06.925602 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 21 14:35:06 crc kubenswrapper[4897]: I1121 14:35:06.931580 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Nov 21 14:35:06 crc kubenswrapper[4897]: E1121 14:35:06.932134 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="813eae15-fec1-4d3a-bae2-1dc597c528f7" containerName="mysqld-exporter" Nov 21 14:35:06 crc kubenswrapper[4897]: I1121 14:35:06.932149 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="813eae15-fec1-4d3a-bae2-1dc597c528f7" containerName="mysqld-exporter" Nov 21 14:35:06 crc kubenswrapper[4897]: E1121 14:35:06.932183 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95ad29e4-b4b6-4cec-b978-23f7ea25edc6" containerName="kube-state-metrics" Nov 21 14:35:06 crc kubenswrapper[4897]: I1121 14:35:06.932189 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="95ad29e4-b4b6-4cec-b978-23f7ea25edc6" containerName="kube-state-metrics" Nov 21 14:35:06 crc kubenswrapper[4897]: I1121 14:35:06.932392 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="95ad29e4-b4b6-4cec-b978-23f7ea25edc6" containerName="kube-state-metrics" Nov 21 14:35:06 crc kubenswrapper[4897]: I1121 14:35:06.932414 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="813eae15-fec1-4d3a-bae2-1dc597c528f7" containerName="mysqld-exporter" Nov 21 14:35:06 crc kubenswrapper[4897]: I1121 14:35:06.933210 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 21 14:35:06 crc kubenswrapper[4897]: I1121 14:35:06.938069 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Nov 21 14:35:06 crc kubenswrapper[4897]: I1121 14:35:06.938297 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Nov 21 14:35:06 crc kubenswrapper[4897]: I1121 14:35:06.948091 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 21 14:35:07 crc kubenswrapper[4897]: I1121 14:35:07.035489 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/a112f75a-fe6e-4150-89bb-bcae6e49ef7b-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"a112f75a-fe6e-4150-89bb-bcae6e49ef7b\") " pod="openstack/kube-state-metrics-0" Nov 21 14:35:07 crc kubenswrapper[4897]: I1121 14:35:07.035578 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a112f75a-fe6e-4150-89bb-bcae6e49ef7b-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"a112f75a-fe6e-4150-89bb-bcae6e49ef7b\") " pod="openstack/kube-state-metrics-0" Nov 21 14:35:07 crc kubenswrapper[4897]: I1121 14:35:07.035621 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/a112f75a-fe6e-4150-89bb-bcae6e49ef7b-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"a112f75a-fe6e-4150-89bb-bcae6e49ef7b\") " pod="openstack/kube-state-metrics-0" Nov 21 14:35:07 crc kubenswrapper[4897]: I1121 14:35:07.035680 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bzndn\" (UniqueName: \"kubernetes.io/projected/a112f75a-fe6e-4150-89bb-bcae6e49ef7b-kube-api-access-bzndn\") pod \"kube-state-metrics-0\" (UID: \"a112f75a-fe6e-4150-89bb-bcae6e49ef7b\") " pod="openstack/kube-state-metrics-0" Nov 21 14:35:07 crc kubenswrapper[4897]: I1121 14:35:07.137850 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/a112f75a-fe6e-4150-89bb-bcae6e49ef7b-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"a112f75a-fe6e-4150-89bb-bcae6e49ef7b\") " pod="openstack/kube-state-metrics-0" Nov 21 14:35:07 crc kubenswrapper[4897]: I1121 14:35:07.137920 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a112f75a-fe6e-4150-89bb-bcae6e49ef7b-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"a112f75a-fe6e-4150-89bb-bcae6e49ef7b\") " pod="openstack/kube-state-metrics-0" Nov 21 14:35:07 crc kubenswrapper[4897]: I1121 14:35:07.138043 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/a112f75a-fe6e-4150-89bb-bcae6e49ef7b-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"a112f75a-fe6e-4150-89bb-bcae6e49ef7b\") " pod="openstack/kube-state-metrics-0" Nov 21 14:35:07 crc kubenswrapper[4897]: I1121 14:35:07.138098 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bzndn\" (UniqueName: \"kubernetes.io/projected/a112f75a-fe6e-4150-89bb-bcae6e49ef7b-kube-api-access-bzndn\") pod \"kube-state-metrics-0\" (UID: \"a112f75a-fe6e-4150-89bb-bcae6e49ef7b\") " pod="openstack/kube-state-metrics-0" Nov 21 14:35:07 crc kubenswrapper[4897]: I1121 14:35:07.142703 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/a112f75a-fe6e-4150-89bb-bcae6e49ef7b-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"a112f75a-fe6e-4150-89bb-bcae6e49ef7b\") " pod="openstack/kube-state-metrics-0" Nov 21 14:35:07 crc kubenswrapper[4897]: I1121 14:35:07.143317 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/a112f75a-fe6e-4150-89bb-bcae6e49ef7b-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"a112f75a-fe6e-4150-89bb-bcae6e49ef7b\") " pod="openstack/kube-state-metrics-0" Nov 21 14:35:07 crc kubenswrapper[4897]: I1121 14:35:07.146001 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a112f75a-fe6e-4150-89bb-bcae6e49ef7b-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"a112f75a-fe6e-4150-89bb-bcae6e49ef7b\") " pod="openstack/kube-state-metrics-0" Nov 21 14:35:07 crc kubenswrapper[4897]: I1121 14:35:07.155889 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bzndn\" (UniqueName: \"kubernetes.io/projected/a112f75a-fe6e-4150-89bb-bcae6e49ef7b-kube-api-access-bzndn\") pod \"kube-state-metrics-0\" (UID: \"a112f75a-fe6e-4150-89bb-bcae6e49ef7b\") " pod="openstack/kube-state-metrics-0" Nov 21 14:35:07 crc kubenswrapper[4897]: I1121 14:35:07.297261 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 21 14:35:07 crc kubenswrapper[4897]: I1121 14:35:07.522131 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0" event={"ID":"813eae15-fec1-4d3a-bae2-1dc597c528f7","Type":"ContainerDied","Data":"90f8d530e55f456883c5f70d3642d86ec952d99a11869feb42d70bf1ee5dcfe9"} Nov 21 14:35:07 crc kubenswrapper[4897]: I1121 14:35:07.522162 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0" Nov 21 14:35:07 crc kubenswrapper[4897]: I1121 14:35:07.534801 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"53e2a389-78b0-463a-82df-dc3d1d870b7b","Type":"ContainerStarted","Data":"739afdf043be77117d422b7a6c54fdb232351f5cf29a3121b8e40c85d307efbc"} Nov 21 14:35:07 crc kubenswrapper[4897]: I1121 14:35:07.640863 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mysqld-exporter-0"] Nov 21 14:35:07 crc kubenswrapper[4897]: I1121 14:35:07.653858 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mysqld-exporter-0"] Nov 21 14:35:07 crc kubenswrapper[4897]: I1121 14:35:07.668322 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mysqld-exporter-0"] Nov 21 14:35:07 crc kubenswrapper[4897]: I1121 14:35:07.670390 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0" Nov 21 14:35:07 crc kubenswrapper[4897]: I1121 14:35:07.674119 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-0"] Nov 21 14:35:07 crc kubenswrapper[4897]: I1121 14:35:07.676367 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-mysqld-exporter-svc" Nov 21 14:35:07 crc kubenswrapper[4897]: I1121 14:35:07.676543 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"mysqld-exporter-config-data" Nov 21 14:35:07 crc kubenswrapper[4897]: I1121 14:35:07.765789 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eeb2be94-07dd-4fa9-8b09-9fed45f415ba-combined-ca-bundle\") pod \"mysqld-exporter-0\" (UID: \"eeb2be94-07dd-4fa9-8b09-9fed45f415ba\") " pod="openstack/mysqld-exporter-0" Nov 21 14:35:07 crc kubenswrapper[4897]: I1121 14:35:07.765843 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p4pkj\" (UniqueName: \"kubernetes.io/projected/eeb2be94-07dd-4fa9-8b09-9fed45f415ba-kube-api-access-p4pkj\") pod \"mysqld-exporter-0\" (UID: \"eeb2be94-07dd-4fa9-8b09-9fed45f415ba\") " pod="openstack/mysqld-exporter-0" Nov 21 14:35:07 crc kubenswrapper[4897]: I1121 14:35:07.766110 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eeb2be94-07dd-4fa9-8b09-9fed45f415ba-config-data\") pod \"mysqld-exporter-0\" (UID: \"eeb2be94-07dd-4fa9-8b09-9fed45f415ba\") " pod="openstack/mysqld-exporter-0" Nov 21 14:35:07 crc kubenswrapper[4897]: I1121 14:35:07.766181 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mysqld-exporter-tls-certs\" (UniqueName: \"kubernetes.io/secret/eeb2be94-07dd-4fa9-8b09-9fed45f415ba-mysqld-exporter-tls-certs\") pod \"mysqld-exporter-0\" (UID: \"eeb2be94-07dd-4fa9-8b09-9fed45f415ba\") " pod="openstack/mysqld-exporter-0" Nov 21 14:35:07 crc kubenswrapper[4897]: I1121 14:35:07.799199 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 21 14:35:07 crc kubenswrapper[4897]: I1121 14:35:07.868223 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eeb2be94-07dd-4fa9-8b09-9fed45f415ba-config-data\") pod \"mysqld-exporter-0\" (UID: \"eeb2be94-07dd-4fa9-8b09-9fed45f415ba\") " pod="openstack/mysqld-exporter-0" Nov 21 14:35:07 crc kubenswrapper[4897]: I1121 14:35:07.868331 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mysqld-exporter-tls-certs\" (UniqueName: \"kubernetes.io/secret/eeb2be94-07dd-4fa9-8b09-9fed45f415ba-mysqld-exporter-tls-certs\") pod \"mysqld-exporter-0\" (UID: \"eeb2be94-07dd-4fa9-8b09-9fed45f415ba\") " pod="openstack/mysqld-exporter-0" Nov 21 14:35:07 crc kubenswrapper[4897]: I1121 14:35:07.868442 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eeb2be94-07dd-4fa9-8b09-9fed45f415ba-combined-ca-bundle\") pod \"mysqld-exporter-0\" (UID: \"eeb2be94-07dd-4fa9-8b09-9fed45f415ba\") " pod="openstack/mysqld-exporter-0" Nov 21 14:35:07 crc kubenswrapper[4897]: I1121 14:35:07.868470 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p4pkj\" (UniqueName: \"kubernetes.io/projected/eeb2be94-07dd-4fa9-8b09-9fed45f415ba-kube-api-access-p4pkj\") pod \"mysqld-exporter-0\" (UID: \"eeb2be94-07dd-4fa9-8b09-9fed45f415ba\") " pod="openstack/mysqld-exporter-0" Nov 21 14:35:07 crc kubenswrapper[4897]: I1121 14:35:07.876607 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mysqld-exporter-tls-certs\" (UniqueName: \"kubernetes.io/secret/eeb2be94-07dd-4fa9-8b09-9fed45f415ba-mysqld-exporter-tls-certs\") pod \"mysqld-exporter-0\" (UID: \"eeb2be94-07dd-4fa9-8b09-9fed45f415ba\") " pod="openstack/mysqld-exporter-0" Nov 21 14:35:07 crc kubenswrapper[4897]: I1121 14:35:07.891329 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eeb2be94-07dd-4fa9-8b09-9fed45f415ba-config-data\") pod \"mysqld-exporter-0\" (UID: \"eeb2be94-07dd-4fa9-8b09-9fed45f415ba\") " pod="openstack/mysqld-exporter-0" Nov 21 14:35:07 crc kubenswrapper[4897]: I1121 14:35:07.892997 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p4pkj\" (UniqueName: \"kubernetes.io/projected/eeb2be94-07dd-4fa9-8b09-9fed45f415ba-kube-api-access-p4pkj\") pod \"mysqld-exporter-0\" (UID: \"eeb2be94-07dd-4fa9-8b09-9fed45f415ba\") " pod="openstack/mysqld-exporter-0" Nov 21 14:35:07 crc kubenswrapper[4897]: I1121 14:35:07.895349 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eeb2be94-07dd-4fa9-8b09-9fed45f415ba-combined-ca-bundle\") pod \"mysqld-exporter-0\" (UID: \"eeb2be94-07dd-4fa9-8b09-9fed45f415ba\") " pod="openstack/mysqld-exporter-0" Nov 21 14:35:08 crc kubenswrapper[4897]: I1121 14:35:08.005728 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0" Nov 21 14:35:08 crc kubenswrapper[4897]: I1121 14:35:08.101957 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="813eae15-fec1-4d3a-bae2-1dc597c528f7" path="/var/lib/kubelet/pods/813eae15-fec1-4d3a-bae2-1dc597c528f7/volumes" Nov 21 14:35:08 crc kubenswrapper[4897]: I1121 14:35:08.202002 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="95ad29e4-b4b6-4cec-b978-23f7ea25edc6" path="/var/lib/kubelet/pods/95ad29e4-b4b6-4cec-b978-23f7ea25edc6/volumes" Nov 21 14:35:08 crc kubenswrapper[4897]: W1121 14:35:08.551704 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeeb2be94_07dd_4fa9_8b09_9fed45f415ba.slice/crio-edd4d61b5f94342a19d0960506a05821391a515c6ad8c8bc1ca2fcecd4552e01 WatchSource:0}: Error finding container edd4d61b5f94342a19d0960506a05821391a515c6ad8c8bc1ca2fcecd4552e01: Status 404 returned error can't find the container with id edd4d61b5f94342a19d0960506a05821391a515c6ad8c8bc1ca2fcecd4552e01 Nov 21 14:35:08 crc kubenswrapper[4897]: I1121 14:35:08.555786 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"a112f75a-fe6e-4150-89bb-bcae6e49ef7b","Type":"ContainerStarted","Data":"d6d262eec00f33d1c143bd8fe900b60e7258e5c3e40fddf6cc30d8a4bca6790b"} Nov 21 14:35:08 crc kubenswrapper[4897]: I1121 14:35:08.556879 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-0"] Nov 21 14:35:08 crc kubenswrapper[4897]: I1121 14:35:08.789272 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Nov 21 14:35:08 crc kubenswrapper[4897]: I1121 14:35:08.917678 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 21 14:35:08 crc kubenswrapper[4897]: I1121 14:35:08.917948 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a594d7fb-9b6e-4dee-a3bf-391e2fb9db79" containerName="ceilometer-central-agent" containerID="cri-o://18c04b36a73b5151b088829484025696e39c00240447d79e11f023c996e82105" gracePeriod=30 Nov 21 14:35:08 crc kubenswrapper[4897]: I1121 14:35:08.918062 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a594d7fb-9b6e-4dee-a3bf-391e2fb9db79" containerName="sg-core" containerID="cri-o://901c18cc885e322e944f4be7f228752ce978bc1d221bc929b698dde78b6aac34" gracePeriod=30 Nov 21 14:35:08 crc kubenswrapper[4897]: I1121 14:35:08.918051 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a594d7fb-9b6e-4dee-a3bf-391e2fb9db79" containerName="proxy-httpd" containerID="cri-o://c122f9b6dbb62121086467252bceab944be70a98a9aa396fe5b157c686d35c15" gracePeriod=30 Nov 21 14:35:08 crc kubenswrapper[4897]: I1121 14:35:08.918105 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a594d7fb-9b6e-4dee-a3bf-391e2fb9db79" containerName="ceilometer-notification-agent" containerID="cri-o://96a3c1e1449c8235b41c8198d4056d1b5f09d1fe21ac9b4de815013719e3bf16" gracePeriod=30 Nov 21 14:35:09 crc kubenswrapper[4897]: I1121 14:35:09.570440 4897 generic.go:334] "Generic (PLEG): container finished" podID="a594d7fb-9b6e-4dee-a3bf-391e2fb9db79" containerID="901c18cc885e322e944f4be7f228752ce978bc1d221bc929b698dde78b6aac34" exitCode=2 Nov 21 14:35:09 crc kubenswrapper[4897]: I1121 14:35:09.570761 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a594d7fb-9b6e-4dee-a3bf-391e2fb9db79","Type":"ContainerDied","Data":"901c18cc885e322e944f4be7f228752ce978bc1d221bc929b698dde78b6aac34"} Nov 21 14:35:09 crc kubenswrapper[4897]: I1121 14:35:09.573676 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0" event={"ID":"eeb2be94-07dd-4fa9-8b09-9fed45f415ba","Type":"ContainerStarted","Data":"edd4d61b5f94342a19d0960506a05821391a515c6ad8c8bc1ca2fcecd4552e01"} Nov 21 14:35:09 crc kubenswrapper[4897]: E1121 14:35:09.662484 4897 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda594d7fb_9b6e_4dee_a3bf_391e2fb9db79.slice/crio-18c04b36a73b5151b088829484025696e39c00240447d79e11f023c996e82105.scope\": RecentStats: unable to find data in memory cache]" Nov 21 14:35:10 crc kubenswrapper[4897]: I1121 14:35:10.587593 4897 generic.go:334] "Generic (PLEG): container finished" podID="a594d7fb-9b6e-4dee-a3bf-391e2fb9db79" containerID="18c04b36a73b5151b088829484025696e39c00240447d79e11f023c996e82105" exitCode=0 Nov 21 14:35:10 crc kubenswrapper[4897]: I1121 14:35:10.587759 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a594d7fb-9b6e-4dee-a3bf-391e2fb9db79","Type":"ContainerDied","Data":"18c04b36a73b5151b088829484025696e39c00240447d79e11f023c996e82105"} Nov 21 14:35:11 crc kubenswrapper[4897]: I1121 14:35:11.615683 4897 generic.go:334] "Generic (PLEG): container finished" podID="a594d7fb-9b6e-4dee-a3bf-391e2fb9db79" containerID="c122f9b6dbb62121086467252bceab944be70a98a9aa396fe5b157c686d35c15" exitCode=0 Nov 21 14:35:11 crc kubenswrapper[4897]: I1121 14:35:11.616143 4897 generic.go:334] "Generic (PLEG): container finished" podID="a594d7fb-9b6e-4dee-a3bf-391e2fb9db79" containerID="96a3c1e1449c8235b41c8198d4056d1b5f09d1fe21ac9b4de815013719e3bf16" exitCode=0 Nov 21 14:35:11 crc kubenswrapper[4897]: I1121 14:35:11.615750 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a594d7fb-9b6e-4dee-a3bf-391e2fb9db79","Type":"ContainerDied","Data":"c122f9b6dbb62121086467252bceab944be70a98a9aa396fe5b157c686d35c15"} Nov 21 14:35:11 crc kubenswrapper[4897]: I1121 14:35:11.616185 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a594d7fb-9b6e-4dee-a3bf-391e2fb9db79","Type":"ContainerDied","Data":"96a3c1e1449c8235b41c8198d4056d1b5f09d1fe21ac9b4de815013719e3bf16"} Nov 21 14:35:12 crc kubenswrapper[4897]: I1121 14:35:12.364898 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 21 14:35:12 crc kubenswrapper[4897]: I1121 14:35:12.488607 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9hg7j\" (UniqueName: \"kubernetes.io/projected/a594d7fb-9b6e-4dee-a3bf-391e2fb9db79-kube-api-access-9hg7j\") pod \"a594d7fb-9b6e-4dee-a3bf-391e2fb9db79\" (UID: \"a594d7fb-9b6e-4dee-a3bf-391e2fb9db79\") " Nov 21 14:35:12 crc kubenswrapper[4897]: I1121 14:35:12.488717 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a594d7fb-9b6e-4dee-a3bf-391e2fb9db79-combined-ca-bundle\") pod \"a594d7fb-9b6e-4dee-a3bf-391e2fb9db79\" (UID: \"a594d7fb-9b6e-4dee-a3bf-391e2fb9db79\") " Nov 21 14:35:12 crc kubenswrapper[4897]: I1121 14:35:12.488815 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a594d7fb-9b6e-4dee-a3bf-391e2fb9db79-log-httpd\") pod \"a594d7fb-9b6e-4dee-a3bf-391e2fb9db79\" (UID: \"a594d7fb-9b6e-4dee-a3bf-391e2fb9db79\") " Nov 21 14:35:12 crc kubenswrapper[4897]: I1121 14:35:12.488868 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a594d7fb-9b6e-4dee-a3bf-391e2fb9db79-sg-core-conf-yaml\") pod \"a594d7fb-9b6e-4dee-a3bf-391e2fb9db79\" (UID: \"a594d7fb-9b6e-4dee-a3bf-391e2fb9db79\") " Nov 21 14:35:12 crc kubenswrapper[4897]: I1121 14:35:12.489127 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a594d7fb-9b6e-4dee-a3bf-391e2fb9db79-scripts\") pod \"a594d7fb-9b6e-4dee-a3bf-391e2fb9db79\" (UID: \"a594d7fb-9b6e-4dee-a3bf-391e2fb9db79\") " Nov 21 14:35:12 crc kubenswrapper[4897]: I1121 14:35:12.489158 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a594d7fb-9b6e-4dee-a3bf-391e2fb9db79-run-httpd\") pod \"a594d7fb-9b6e-4dee-a3bf-391e2fb9db79\" (UID: \"a594d7fb-9b6e-4dee-a3bf-391e2fb9db79\") " Nov 21 14:35:12 crc kubenswrapper[4897]: I1121 14:35:12.489304 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a594d7fb-9b6e-4dee-a3bf-391e2fb9db79-config-data\") pod \"a594d7fb-9b6e-4dee-a3bf-391e2fb9db79\" (UID: \"a594d7fb-9b6e-4dee-a3bf-391e2fb9db79\") " Nov 21 14:35:12 crc kubenswrapper[4897]: I1121 14:35:12.489727 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a594d7fb-9b6e-4dee-a3bf-391e2fb9db79-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "a594d7fb-9b6e-4dee-a3bf-391e2fb9db79" (UID: "a594d7fb-9b6e-4dee-a3bf-391e2fb9db79"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:35:12 crc kubenswrapper[4897]: I1121 14:35:12.489868 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a594d7fb-9b6e-4dee-a3bf-391e2fb9db79-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "a594d7fb-9b6e-4dee-a3bf-391e2fb9db79" (UID: "a594d7fb-9b6e-4dee-a3bf-391e2fb9db79"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:35:12 crc kubenswrapper[4897]: I1121 14:35:12.491288 4897 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a594d7fb-9b6e-4dee-a3bf-391e2fb9db79-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 21 14:35:12 crc kubenswrapper[4897]: I1121 14:35:12.491315 4897 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a594d7fb-9b6e-4dee-a3bf-391e2fb9db79-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 21 14:35:12 crc kubenswrapper[4897]: I1121 14:35:12.496281 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a594d7fb-9b6e-4dee-a3bf-391e2fb9db79-kube-api-access-9hg7j" (OuterVolumeSpecName: "kube-api-access-9hg7j") pod "a594d7fb-9b6e-4dee-a3bf-391e2fb9db79" (UID: "a594d7fb-9b6e-4dee-a3bf-391e2fb9db79"). InnerVolumeSpecName "kube-api-access-9hg7j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:35:12 crc kubenswrapper[4897]: I1121 14:35:12.497079 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a594d7fb-9b6e-4dee-a3bf-391e2fb9db79-scripts" (OuterVolumeSpecName: "scripts") pod "a594d7fb-9b6e-4dee-a3bf-391e2fb9db79" (UID: "a594d7fb-9b6e-4dee-a3bf-391e2fb9db79"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:35:12 crc kubenswrapper[4897]: I1121 14:35:12.537187 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a594d7fb-9b6e-4dee-a3bf-391e2fb9db79-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "a594d7fb-9b6e-4dee-a3bf-391e2fb9db79" (UID: "a594d7fb-9b6e-4dee-a3bf-391e2fb9db79"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:35:12 crc kubenswrapper[4897]: I1121 14:35:12.588147 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a594d7fb-9b6e-4dee-a3bf-391e2fb9db79-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a594d7fb-9b6e-4dee-a3bf-391e2fb9db79" (UID: "a594d7fb-9b6e-4dee-a3bf-391e2fb9db79"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:35:12 crc kubenswrapper[4897]: I1121 14:35:12.593871 4897 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a594d7fb-9b6e-4dee-a3bf-391e2fb9db79-scripts\") on node \"crc\" DevicePath \"\"" Nov 21 14:35:12 crc kubenswrapper[4897]: I1121 14:35:12.593909 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9hg7j\" (UniqueName: \"kubernetes.io/projected/a594d7fb-9b6e-4dee-a3bf-391e2fb9db79-kube-api-access-9hg7j\") on node \"crc\" DevicePath \"\"" Nov 21 14:35:12 crc kubenswrapper[4897]: I1121 14:35:12.593920 4897 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a594d7fb-9b6e-4dee-a3bf-391e2fb9db79-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 21 14:35:12 crc kubenswrapper[4897]: I1121 14:35:12.593930 4897 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a594d7fb-9b6e-4dee-a3bf-391e2fb9db79-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 21 14:35:12 crc kubenswrapper[4897]: I1121 14:35:12.633700 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a594d7fb-9b6e-4dee-a3bf-391e2fb9db79","Type":"ContainerDied","Data":"2de62a1585b6cabd107dd73d8c96b32664ff679b31d9a581ea7a0d1c17da8d77"} Nov 21 14:35:12 crc kubenswrapper[4897]: I1121 14:35:12.634189 4897 scope.go:117] "RemoveContainer" containerID="c122f9b6dbb62121086467252bceab944be70a98a9aa396fe5b157c686d35c15" Nov 21 14:35:12 crc kubenswrapper[4897]: I1121 14:35:12.633844 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 21 14:35:12 crc kubenswrapper[4897]: I1121 14:35:12.678328 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a594d7fb-9b6e-4dee-a3bf-391e2fb9db79-config-data" (OuterVolumeSpecName: "config-data") pod "a594d7fb-9b6e-4dee-a3bf-391e2fb9db79" (UID: "a594d7fb-9b6e-4dee-a3bf-391e2fb9db79"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:35:12 crc kubenswrapper[4897]: I1121 14:35:12.696637 4897 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a594d7fb-9b6e-4dee-a3bf-391e2fb9db79-config-data\") on node \"crc\" DevicePath \"\"" Nov 21 14:35:12 crc kubenswrapper[4897]: I1121 14:35:12.729789 4897 scope.go:117] "RemoveContainer" containerID="901c18cc885e322e944f4be7f228752ce978bc1d221bc929b698dde78b6aac34" Nov 21 14:35:12 crc kubenswrapper[4897]: I1121 14:35:12.764711 4897 scope.go:117] "RemoveContainer" containerID="96a3c1e1449c8235b41c8198d4056d1b5f09d1fe21ac9b4de815013719e3bf16" Nov 21 14:35:12 crc kubenswrapper[4897]: I1121 14:35:12.793901 4897 scope.go:117] "RemoveContainer" containerID="18c04b36a73b5151b088829484025696e39c00240447d79e11f023c996e82105" Nov 21 14:35:13 crc kubenswrapper[4897]: I1121 14:35:13.004156 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 21 14:35:13 crc kubenswrapper[4897]: I1121 14:35:13.036302 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 21 14:35:13 crc kubenswrapper[4897]: I1121 14:35:13.074558 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 21 14:35:13 crc kubenswrapper[4897]: E1121 14:35:13.075132 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a594d7fb-9b6e-4dee-a3bf-391e2fb9db79" containerName="ceilometer-notification-agent" Nov 21 14:35:13 crc kubenswrapper[4897]: I1121 14:35:13.075157 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="a594d7fb-9b6e-4dee-a3bf-391e2fb9db79" containerName="ceilometer-notification-agent" Nov 21 14:35:13 crc kubenswrapper[4897]: E1121 14:35:13.075190 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a594d7fb-9b6e-4dee-a3bf-391e2fb9db79" containerName="sg-core" Nov 21 14:35:13 crc kubenswrapper[4897]: I1121 14:35:13.075199 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="a594d7fb-9b6e-4dee-a3bf-391e2fb9db79" containerName="sg-core" Nov 21 14:35:13 crc kubenswrapper[4897]: E1121 14:35:13.075213 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a594d7fb-9b6e-4dee-a3bf-391e2fb9db79" containerName="ceilometer-central-agent" Nov 21 14:35:13 crc kubenswrapper[4897]: I1121 14:35:13.075221 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="a594d7fb-9b6e-4dee-a3bf-391e2fb9db79" containerName="ceilometer-central-agent" Nov 21 14:35:13 crc kubenswrapper[4897]: E1121 14:35:13.075237 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a594d7fb-9b6e-4dee-a3bf-391e2fb9db79" containerName="proxy-httpd" Nov 21 14:35:13 crc kubenswrapper[4897]: I1121 14:35:13.075244 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="a594d7fb-9b6e-4dee-a3bf-391e2fb9db79" containerName="proxy-httpd" Nov 21 14:35:13 crc kubenswrapper[4897]: I1121 14:35:13.075575 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="a594d7fb-9b6e-4dee-a3bf-391e2fb9db79" containerName="ceilometer-notification-agent" Nov 21 14:35:13 crc kubenswrapper[4897]: I1121 14:35:13.075607 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="a594d7fb-9b6e-4dee-a3bf-391e2fb9db79" containerName="sg-core" Nov 21 14:35:13 crc kubenswrapper[4897]: I1121 14:35:13.075624 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="a594d7fb-9b6e-4dee-a3bf-391e2fb9db79" containerName="ceilometer-central-agent" Nov 21 14:35:13 crc kubenswrapper[4897]: I1121 14:35:13.075640 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="a594d7fb-9b6e-4dee-a3bf-391e2fb9db79" containerName="proxy-httpd" Nov 21 14:35:13 crc kubenswrapper[4897]: I1121 14:35:13.077869 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 21 14:35:13 crc kubenswrapper[4897]: I1121 14:35:13.079903 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 21 14:35:13 crc kubenswrapper[4897]: I1121 14:35:13.081022 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 21 14:35:13 crc kubenswrapper[4897]: I1121 14:35:13.081643 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 21 14:35:13 crc kubenswrapper[4897]: I1121 14:35:13.091312 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 21 14:35:13 crc kubenswrapper[4897]: I1121 14:35:13.226044 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ac00807a-ef51-4204-af3f-efddb3dc8af4-log-httpd\") pod \"ceilometer-0\" (UID: \"ac00807a-ef51-4204-af3f-efddb3dc8af4\") " pod="openstack/ceilometer-0" Nov 21 14:35:13 crc kubenswrapper[4897]: I1121 14:35:13.226103 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ac00807a-ef51-4204-af3f-efddb3dc8af4-run-httpd\") pod \"ceilometer-0\" (UID: \"ac00807a-ef51-4204-af3f-efddb3dc8af4\") " pod="openstack/ceilometer-0" Nov 21 14:35:13 crc kubenswrapper[4897]: I1121 14:35:13.226159 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac00807a-ef51-4204-af3f-efddb3dc8af4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ac00807a-ef51-4204-af3f-efddb3dc8af4\") " pod="openstack/ceilometer-0" Nov 21 14:35:13 crc kubenswrapper[4897]: I1121 14:35:13.226194 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac00807a-ef51-4204-af3f-efddb3dc8af4-config-data\") pod \"ceilometer-0\" (UID: \"ac00807a-ef51-4204-af3f-efddb3dc8af4\") " pod="openstack/ceilometer-0" Nov 21 14:35:13 crc kubenswrapper[4897]: I1121 14:35:13.226223 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ac00807a-ef51-4204-af3f-efddb3dc8af4-scripts\") pod \"ceilometer-0\" (UID: \"ac00807a-ef51-4204-af3f-efddb3dc8af4\") " pod="openstack/ceilometer-0" Nov 21 14:35:13 crc kubenswrapper[4897]: I1121 14:35:13.226239 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ac00807a-ef51-4204-af3f-efddb3dc8af4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ac00807a-ef51-4204-af3f-efddb3dc8af4\") " pod="openstack/ceilometer-0" Nov 21 14:35:13 crc kubenswrapper[4897]: I1121 14:35:13.226256 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-285dv\" (UniqueName: \"kubernetes.io/projected/ac00807a-ef51-4204-af3f-efddb3dc8af4-kube-api-access-285dv\") pod \"ceilometer-0\" (UID: \"ac00807a-ef51-4204-af3f-efddb3dc8af4\") " pod="openstack/ceilometer-0" Nov 21 14:35:13 crc kubenswrapper[4897]: I1121 14:35:13.226352 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ac00807a-ef51-4204-af3f-efddb3dc8af4-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"ac00807a-ef51-4204-af3f-efddb3dc8af4\") " pod="openstack/ceilometer-0" Nov 21 14:35:13 crc kubenswrapper[4897]: I1121 14:35:13.328252 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ac00807a-ef51-4204-af3f-efddb3dc8af4-log-httpd\") pod \"ceilometer-0\" (UID: \"ac00807a-ef51-4204-af3f-efddb3dc8af4\") " pod="openstack/ceilometer-0" Nov 21 14:35:13 crc kubenswrapper[4897]: I1121 14:35:13.328329 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ac00807a-ef51-4204-af3f-efddb3dc8af4-run-httpd\") pod \"ceilometer-0\" (UID: \"ac00807a-ef51-4204-af3f-efddb3dc8af4\") " pod="openstack/ceilometer-0" Nov 21 14:35:13 crc kubenswrapper[4897]: I1121 14:35:13.328407 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac00807a-ef51-4204-af3f-efddb3dc8af4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ac00807a-ef51-4204-af3f-efddb3dc8af4\") " pod="openstack/ceilometer-0" Nov 21 14:35:13 crc kubenswrapper[4897]: I1121 14:35:13.328441 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac00807a-ef51-4204-af3f-efddb3dc8af4-config-data\") pod \"ceilometer-0\" (UID: \"ac00807a-ef51-4204-af3f-efddb3dc8af4\") " pod="openstack/ceilometer-0" Nov 21 14:35:13 crc kubenswrapper[4897]: I1121 14:35:13.328487 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ac00807a-ef51-4204-af3f-efddb3dc8af4-scripts\") pod \"ceilometer-0\" (UID: \"ac00807a-ef51-4204-af3f-efddb3dc8af4\") " pod="openstack/ceilometer-0" Nov 21 14:35:13 crc kubenswrapper[4897]: I1121 14:35:13.328544 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ac00807a-ef51-4204-af3f-efddb3dc8af4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ac00807a-ef51-4204-af3f-efddb3dc8af4\") " pod="openstack/ceilometer-0" Nov 21 14:35:13 crc kubenswrapper[4897]: I1121 14:35:13.328586 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-285dv\" (UniqueName: \"kubernetes.io/projected/ac00807a-ef51-4204-af3f-efddb3dc8af4-kube-api-access-285dv\") pod \"ceilometer-0\" (UID: \"ac00807a-ef51-4204-af3f-efddb3dc8af4\") " pod="openstack/ceilometer-0" Nov 21 14:35:13 crc kubenswrapper[4897]: I1121 14:35:13.328721 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ac00807a-ef51-4204-af3f-efddb3dc8af4-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"ac00807a-ef51-4204-af3f-efddb3dc8af4\") " pod="openstack/ceilometer-0" Nov 21 14:35:13 crc kubenswrapper[4897]: I1121 14:35:13.328758 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ac00807a-ef51-4204-af3f-efddb3dc8af4-log-httpd\") pod \"ceilometer-0\" (UID: \"ac00807a-ef51-4204-af3f-efddb3dc8af4\") " pod="openstack/ceilometer-0" Nov 21 14:35:13 crc kubenswrapper[4897]: I1121 14:35:13.328811 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ac00807a-ef51-4204-af3f-efddb3dc8af4-run-httpd\") pod \"ceilometer-0\" (UID: \"ac00807a-ef51-4204-af3f-efddb3dc8af4\") " pod="openstack/ceilometer-0" Nov 21 14:35:13 crc kubenswrapper[4897]: I1121 14:35:13.332837 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ac00807a-ef51-4204-af3f-efddb3dc8af4-scripts\") pod \"ceilometer-0\" (UID: \"ac00807a-ef51-4204-af3f-efddb3dc8af4\") " pod="openstack/ceilometer-0" Nov 21 14:35:13 crc kubenswrapper[4897]: I1121 14:35:13.332935 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac00807a-ef51-4204-af3f-efddb3dc8af4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ac00807a-ef51-4204-af3f-efddb3dc8af4\") " pod="openstack/ceilometer-0" Nov 21 14:35:13 crc kubenswrapper[4897]: I1121 14:35:13.333216 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ac00807a-ef51-4204-af3f-efddb3dc8af4-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"ac00807a-ef51-4204-af3f-efddb3dc8af4\") " pod="openstack/ceilometer-0" Nov 21 14:35:13 crc kubenswrapper[4897]: I1121 14:35:13.335070 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ac00807a-ef51-4204-af3f-efddb3dc8af4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ac00807a-ef51-4204-af3f-efddb3dc8af4\") " pod="openstack/ceilometer-0" Nov 21 14:35:13 crc kubenswrapper[4897]: I1121 14:35:13.336683 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac00807a-ef51-4204-af3f-efddb3dc8af4-config-data\") pod \"ceilometer-0\" (UID: \"ac00807a-ef51-4204-af3f-efddb3dc8af4\") " pod="openstack/ceilometer-0" Nov 21 14:35:13 crc kubenswrapper[4897]: I1121 14:35:13.346712 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-285dv\" (UniqueName: \"kubernetes.io/projected/ac00807a-ef51-4204-af3f-efddb3dc8af4-kube-api-access-285dv\") pod \"ceilometer-0\" (UID: \"ac00807a-ef51-4204-af3f-efddb3dc8af4\") " pod="openstack/ceilometer-0" Nov 21 14:35:13 crc kubenswrapper[4897]: I1121 14:35:13.405180 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 21 14:35:13 crc kubenswrapper[4897]: I1121 14:35:13.908661 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 21 14:35:14 crc kubenswrapper[4897]: I1121 14:35:14.107860 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a594d7fb-9b6e-4dee-a3bf-391e2fb9db79" path="/var/lib/kubelet/pods/a594d7fb-9b6e-4dee-a3bf-391e2fb9db79/volumes" Nov 21 14:35:14 crc kubenswrapper[4897]: I1121 14:35:14.663100 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ac00807a-ef51-4204-af3f-efddb3dc8af4","Type":"ContainerStarted","Data":"d3d55b20f98ec93fcdba4bae00e1b09606030a61abe1568c168bdee50946675d"} Nov 21 14:35:14 crc kubenswrapper[4897]: I1121 14:35:14.669338 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0" event={"ID":"eeb2be94-07dd-4fa9-8b09-9fed45f415ba","Type":"ContainerStarted","Data":"8ffe8ba21bba18c838d3ec7a264389bb2308210626355c38e722554fe7b8d037"} Nov 21 14:35:14 crc kubenswrapper[4897]: I1121 14:35:14.697910 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mysqld-exporter-0" podStartSLOduration=2.140562788 podStartE2EDuration="7.697887554s" podCreationTimestamp="2025-11-21 14:35:07 +0000 UTC" firstStartedPulling="2025-11-21 14:35:08.555750141 +0000 UTC m=+1585.840343606" lastFinishedPulling="2025-11-21 14:35:14.113074897 +0000 UTC m=+1591.397668372" observedRunningTime="2025-11-21 14:35:14.688786159 +0000 UTC m=+1591.973379674" watchObservedRunningTime="2025-11-21 14:35:14.697887554 +0000 UTC m=+1591.982481059" Nov 21 14:35:15 crc kubenswrapper[4897]: I1121 14:35:15.437938 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 21 14:35:15 crc kubenswrapper[4897]: I1121 14:35:15.679253 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"53e2a389-78b0-463a-82df-dc3d1d870b7b","Type":"ContainerStarted","Data":"51ef72dcaa68128a85131783b7460b847ef18d69b8d518d355a8e2fc52cea669"} Nov 21 14:35:15 crc kubenswrapper[4897]: I1121 14:35:15.680647 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"a112f75a-fe6e-4150-89bb-bcae6e49ef7b","Type":"ContainerStarted","Data":"82307cf0784974b51435e1d551c6e125d6d806bde236e6e1b255be05fa1bbf91"} Nov 21 14:35:15 crc kubenswrapper[4897]: I1121 14:35:15.681649 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Nov 21 14:35:15 crc kubenswrapper[4897]: I1121 14:35:15.683689 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ac00807a-ef51-4204-af3f-efddb3dc8af4","Type":"ContainerStarted","Data":"eb938a4500d5755dc62df995062319c2c3965a09b16cde73e9ac8a686ade6a5f"} Nov 21 14:35:15 crc kubenswrapper[4897]: I1121 14:35:15.703313 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=3.248830483 podStartE2EDuration="9.703294962s" podCreationTimestamp="2025-11-21 14:35:06 +0000 UTC" firstStartedPulling="2025-11-21 14:35:07.804603833 +0000 UTC m=+1585.089197308" lastFinishedPulling="2025-11-21 14:35:14.259068312 +0000 UTC m=+1591.543661787" observedRunningTime="2025-11-21 14:35:15.7006217 +0000 UTC m=+1592.985215195" watchObservedRunningTime="2025-11-21 14:35:15.703294962 +0000 UTC m=+1592.987888437" Nov 21 14:35:16 crc kubenswrapper[4897]: I1121 14:35:16.158301 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-dtwql" podUID="6d3cad20-fb17-4726-b8fa-9fd5303b5eea" containerName="registry-server" probeResult="failure" output=< Nov 21 14:35:16 crc kubenswrapper[4897]: timeout: failed to connect service ":50051" within 1s Nov 21 14:35:16 crc kubenswrapper[4897]: > Nov 21 14:35:16 crc kubenswrapper[4897]: I1121 14:35:16.696559 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ac00807a-ef51-4204-af3f-efddb3dc8af4","Type":"ContainerStarted","Data":"a475231d66f3729ab90ffc303ef53bc90030a074d7538fb1da4803ea215ce967"} Nov 21 14:35:17 crc kubenswrapper[4897]: I1121 14:35:17.711810 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"53e2a389-78b0-463a-82df-dc3d1d870b7b","Type":"ContainerStarted","Data":"d2f404c2708b34449a7d699e98fd40e367eaae057219a0725d1c6c79b5893c1e"} Nov 21 14:35:17 crc kubenswrapper[4897]: I1121 14:35:17.716111 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ac00807a-ef51-4204-af3f-efddb3dc8af4","Type":"ContainerStarted","Data":"2e4b3e2467588683943bf462488555daf357e16195e37ba10243bf69ae183f54"} Nov 21 14:35:18 crc kubenswrapper[4897]: I1121 14:35:18.089408 4897 scope.go:117] "RemoveContainer" containerID="c68f1b003d12b4a2964169a85a04e3f0272d86c891f52e6bebd52f3781520d1a" Nov 21 14:35:18 crc kubenswrapper[4897]: E1121 14:35:18.089769 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 14:35:19 crc kubenswrapper[4897]: I1121 14:35:19.745887 4897 generic.go:334] "Generic (PLEG): container finished" podID="c6e90788-4f97-45c0-aacc-cfa848ac0f2f" containerID="03e6cd518a468591cf4ed2a936e891b6a325fffb7bd905bc50948a197be7a6cb" exitCode=0 Nov 21 14:35:19 crc kubenswrapper[4897]: I1121 14:35:19.746295 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-cv5pw" event={"ID":"c6e90788-4f97-45c0-aacc-cfa848ac0f2f","Type":"ContainerDied","Data":"03e6cd518a468591cf4ed2a936e891b6a325fffb7bd905bc50948a197be7a6cb"} Nov 21 14:35:19 crc kubenswrapper[4897]: I1121 14:35:19.754284 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"53e2a389-78b0-463a-82df-dc3d1d870b7b","Type":"ContainerStarted","Data":"b82fb5422034aef083739986bf603d009ad47087d5635868123e0241717761b3"} Nov 21 14:35:19 crc kubenswrapper[4897]: I1121 14:35:19.754632 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="53e2a389-78b0-463a-82df-dc3d1d870b7b" containerName="aodh-api" containerID="cri-o://739afdf043be77117d422b7a6c54fdb232351f5cf29a3121b8e40c85d307efbc" gracePeriod=30 Nov 21 14:35:19 crc kubenswrapper[4897]: I1121 14:35:19.754820 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="53e2a389-78b0-463a-82df-dc3d1d870b7b" containerName="aodh-notifier" containerID="cri-o://d2f404c2708b34449a7d699e98fd40e367eaae057219a0725d1c6c79b5893c1e" gracePeriod=30 Nov 21 14:35:19 crc kubenswrapper[4897]: I1121 14:35:19.754948 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="53e2a389-78b0-463a-82df-dc3d1d870b7b" containerName="aodh-evaluator" containerID="cri-o://51ef72dcaa68128a85131783b7460b847ef18d69b8d518d355a8e2fc52cea669" gracePeriod=30 Nov 21 14:35:19 crc kubenswrapper[4897]: I1121 14:35:19.755054 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="53e2a389-78b0-463a-82df-dc3d1d870b7b" containerName="aodh-listener" containerID="cri-o://b82fb5422034aef083739986bf603d009ad47087d5635868123e0241717761b3" gracePeriod=30 Nov 21 14:35:19 crc kubenswrapper[4897]: I1121 14:35:19.776989 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ac00807a-ef51-4204-af3f-efddb3dc8af4","Type":"ContainerStarted","Data":"d9d5c15f7a13479759204052fc598454eb624a9eb7e0e665e01e2425ab601055"} Nov 21 14:35:19 crc kubenswrapper[4897]: I1121 14:35:19.777154 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ac00807a-ef51-4204-af3f-efddb3dc8af4" containerName="ceilometer-central-agent" containerID="cri-o://eb938a4500d5755dc62df995062319c2c3965a09b16cde73e9ac8a686ade6a5f" gracePeriod=30 Nov 21 14:35:19 crc kubenswrapper[4897]: I1121 14:35:19.777221 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ac00807a-ef51-4204-af3f-efddb3dc8af4" containerName="sg-core" containerID="cri-o://2e4b3e2467588683943bf462488555daf357e16195e37ba10243bf69ae183f54" gracePeriod=30 Nov 21 14:35:19 crc kubenswrapper[4897]: I1121 14:35:19.777280 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ac00807a-ef51-4204-af3f-efddb3dc8af4" containerName="ceilometer-notification-agent" containerID="cri-o://a475231d66f3729ab90ffc303ef53bc90030a074d7538fb1da4803ea215ce967" gracePeriod=30 Nov 21 14:35:19 crc kubenswrapper[4897]: I1121 14:35:19.777302 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ac00807a-ef51-4204-af3f-efddb3dc8af4" containerName="proxy-httpd" containerID="cri-o://d9d5c15f7a13479759204052fc598454eb624a9eb7e0e665e01e2425ab601055" gracePeriod=30 Nov 21 14:35:19 crc kubenswrapper[4897]: I1121 14:35:19.777174 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 21 14:35:19 crc kubenswrapper[4897]: I1121 14:35:19.795072 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=1.829215601 podStartE2EDuration="14.795048406s" podCreationTimestamp="2025-11-21 14:35:05 +0000 UTC" firstStartedPulling="2025-11-21 14:35:06.053839512 +0000 UTC m=+1583.338432987" lastFinishedPulling="2025-11-21 14:35:19.019672317 +0000 UTC m=+1596.304265792" observedRunningTime="2025-11-21 14:35:19.790797572 +0000 UTC m=+1597.075391057" watchObservedRunningTime="2025-11-21 14:35:19.795048406 +0000 UTC m=+1597.079641901" Nov 21 14:35:19 crc kubenswrapper[4897]: I1121 14:35:19.832070 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.871622484 podStartE2EDuration="7.832051109s" podCreationTimestamp="2025-11-21 14:35:12 +0000 UTC" firstStartedPulling="2025-11-21 14:35:14.0725539 +0000 UTC m=+1591.357147375" lastFinishedPulling="2025-11-21 14:35:19.032982525 +0000 UTC m=+1596.317576000" observedRunningTime="2025-11-21 14:35:19.817152429 +0000 UTC m=+1597.101745914" watchObservedRunningTime="2025-11-21 14:35:19.832051109 +0000 UTC m=+1597.116644584" Nov 21 14:35:20 crc kubenswrapper[4897]: I1121 14:35:20.791064 4897 generic.go:334] "Generic (PLEG): container finished" podID="53e2a389-78b0-463a-82df-dc3d1d870b7b" containerID="d2f404c2708b34449a7d699e98fd40e367eaae057219a0725d1c6c79b5893c1e" exitCode=0 Nov 21 14:35:20 crc kubenswrapper[4897]: I1121 14:35:20.791322 4897 generic.go:334] "Generic (PLEG): container finished" podID="53e2a389-78b0-463a-82df-dc3d1d870b7b" containerID="51ef72dcaa68128a85131783b7460b847ef18d69b8d518d355a8e2fc52cea669" exitCode=0 Nov 21 14:35:20 crc kubenswrapper[4897]: I1121 14:35:20.791331 4897 generic.go:334] "Generic (PLEG): container finished" podID="53e2a389-78b0-463a-82df-dc3d1d870b7b" containerID="739afdf043be77117d422b7a6c54fdb232351f5cf29a3121b8e40c85d307efbc" exitCode=0 Nov 21 14:35:20 crc kubenswrapper[4897]: I1121 14:35:20.791116 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"53e2a389-78b0-463a-82df-dc3d1d870b7b","Type":"ContainerDied","Data":"d2f404c2708b34449a7d699e98fd40e367eaae057219a0725d1c6c79b5893c1e"} Nov 21 14:35:20 crc kubenswrapper[4897]: I1121 14:35:20.791447 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"53e2a389-78b0-463a-82df-dc3d1d870b7b","Type":"ContainerDied","Data":"51ef72dcaa68128a85131783b7460b847ef18d69b8d518d355a8e2fc52cea669"} Nov 21 14:35:20 crc kubenswrapper[4897]: I1121 14:35:20.791464 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"53e2a389-78b0-463a-82df-dc3d1d870b7b","Type":"ContainerDied","Data":"739afdf043be77117d422b7a6c54fdb232351f5cf29a3121b8e40c85d307efbc"} Nov 21 14:35:20 crc kubenswrapper[4897]: I1121 14:35:20.794885 4897 generic.go:334] "Generic (PLEG): container finished" podID="ac00807a-ef51-4204-af3f-efddb3dc8af4" containerID="d9d5c15f7a13479759204052fc598454eb624a9eb7e0e665e01e2425ab601055" exitCode=0 Nov 21 14:35:20 crc kubenswrapper[4897]: I1121 14:35:20.794918 4897 generic.go:334] "Generic (PLEG): container finished" podID="ac00807a-ef51-4204-af3f-efddb3dc8af4" containerID="2e4b3e2467588683943bf462488555daf357e16195e37ba10243bf69ae183f54" exitCode=2 Nov 21 14:35:20 crc kubenswrapper[4897]: I1121 14:35:20.794928 4897 generic.go:334] "Generic (PLEG): container finished" podID="ac00807a-ef51-4204-af3f-efddb3dc8af4" containerID="a475231d66f3729ab90ffc303ef53bc90030a074d7538fb1da4803ea215ce967" exitCode=0 Nov 21 14:35:20 crc kubenswrapper[4897]: I1121 14:35:20.794990 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ac00807a-ef51-4204-af3f-efddb3dc8af4","Type":"ContainerDied","Data":"d9d5c15f7a13479759204052fc598454eb624a9eb7e0e665e01e2425ab601055"} Nov 21 14:35:20 crc kubenswrapper[4897]: I1121 14:35:20.795031 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ac00807a-ef51-4204-af3f-efddb3dc8af4","Type":"ContainerDied","Data":"2e4b3e2467588683943bf462488555daf357e16195e37ba10243bf69ae183f54"} Nov 21 14:35:20 crc kubenswrapper[4897]: I1121 14:35:20.795045 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ac00807a-ef51-4204-af3f-efddb3dc8af4","Type":"ContainerDied","Data":"a475231d66f3729ab90ffc303ef53bc90030a074d7538fb1da4803ea215ce967"} Nov 21 14:35:21 crc kubenswrapper[4897]: I1121 14:35:21.199836 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-cv5pw" Nov 21 14:35:21 crc kubenswrapper[4897]: I1121 14:35:21.334512 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tqxjv\" (UniqueName: \"kubernetes.io/projected/c6e90788-4f97-45c0-aacc-cfa848ac0f2f-kube-api-access-tqxjv\") pod \"c6e90788-4f97-45c0-aacc-cfa848ac0f2f\" (UID: \"c6e90788-4f97-45c0-aacc-cfa848ac0f2f\") " Nov 21 14:35:21 crc kubenswrapper[4897]: I1121 14:35:21.334690 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6e90788-4f97-45c0-aacc-cfa848ac0f2f-config-data\") pod \"c6e90788-4f97-45c0-aacc-cfa848ac0f2f\" (UID: \"c6e90788-4f97-45c0-aacc-cfa848ac0f2f\") " Nov 21 14:35:21 crc kubenswrapper[4897]: I1121 14:35:21.334727 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c6e90788-4f97-45c0-aacc-cfa848ac0f2f-scripts\") pod \"c6e90788-4f97-45c0-aacc-cfa848ac0f2f\" (UID: \"c6e90788-4f97-45c0-aacc-cfa848ac0f2f\") " Nov 21 14:35:21 crc kubenswrapper[4897]: I1121 14:35:21.334888 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6e90788-4f97-45c0-aacc-cfa848ac0f2f-combined-ca-bundle\") pod \"c6e90788-4f97-45c0-aacc-cfa848ac0f2f\" (UID: \"c6e90788-4f97-45c0-aacc-cfa848ac0f2f\") " Nov 21 14:35:21 crc kubenswrapper[4897]: I1121 14:35:21.468169 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c6e90788-4f97-45c0-aacc-cfa848ac0f2f-kube-api-access-tqxjv" (OuterVolumeSpecName: "kube-api-access-tqxjv") pod "c6e90788-4f97-45c0-aacc-cfa848ac0f2f" (UID: "c6e90788-4f97-45c0-aacc-cfa848ac0f2f"). InnerVolumeSpecName "kube-api-access-tqxjv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:35:21 crc kubenswrapper[4897]: I1121 14:35:21.469412 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6e90788-4f97-45c0-aacc-cfa848ac0f2f-scripts" (OuterVolumeSpecName: "scripts") pod "c6e90788-4f97-45c0-aacc-cfa848ac0f2f" (UID: "c6e90788-4f97-45c0-aacc-cfa848ac0f2f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:35:21 crc kubenswrapper[4897]: I1121 14:35:21.540161 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tqxjv\" (UniqueName: \"kubernetes.io/projected/c6e90788-4f97-45c0-aacc-cfa848ac0f2f-kube-api-access-tqxjv\") on node \"crc\" DevicePath \"\"" Nov 21 14:35:21 crc kubenswrapper[4897]: I1121 14:35:21.540454 4897 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c6e90788-4f97-45c0-aacc-cfa848ac0f2f-scripts\") on node \"crc\" DevicePath \"\"" Nov 21 14:35:21 crc kubenswrapper[4897]: I1121 14:35:21.545449 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6e90788-4f97-45c0-aacc-cfa848ac0f2f-config-data" (OuterVolumeSpecName: "config-data") pod "c6e90788-4f97-45c0-aacc-cfa848ac0f2f" (UID: "c6e90788-4f97-45c0-aacc-cfa848ac0f2f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:35:21 crc kubenswrapper[4897]: I1121 14:35:21.556213 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6e90788-4f97-45c0-aacc-cfa848ac0f2f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c6e90788-4f97-45c0-aacc-cfa848ac0f2f" (UID: "c6e90788-4f97-45c0-aacc-cfa848ac0f2f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:35:21 crc kubenswrapper[4897]: I1121 14:35:21.642417 4897 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6e90788-4f97-45c0-aacc-cfa848ac0f2f-config-data\") on node \"crc\" DevicePath \"\"" Nov 21 14:35:21 crc kubenswrapper[4897]: I1121 14:35:21.642795 4897 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6e90788-4f97-45c0-aacc-cfa848ac0f2f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 21 14:35:21 crc kubenswrapper[4897]: I1121 14:35:21.808212 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-cv5pw" event={"ID":"c6e90788-4f97-45c0-aacc-cfa848ac0f2f","Type":"ContainerDied","Data":"08b721dc40c171056df1d3e5579a86b1a03c8a8b48881d1df0c2c5cd1e3bba32"} Nov 21 14:35:21 crc kubenswrapper[4897]: I1121 14:35:21.808258 4897 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="08b721dc40c171056df1d3e5579a86b1a03c8a8b48881d1df0c2c5cd1e3bba32" Nov 21 14:35:21 crc kubenswrapper[4897]: I1121 14:35:21.808315 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-cv5pw" Nov 21 14:35:21 crc kubenswrapper[4897]: I1121 14:35:21.872565 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 21 14:35:21 crc kubenswrapper[4897]: E1121 14:35:21.873317 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6e90788-4f97-45c0-aacc-cfa848ac0f2f" containerName="nova-cell0-conductor-db-sync" Nov 21 14:35:21 crc kubenswrapper[4897]: I1121 14:35:21.873448 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6e90788-4f97-45c0-aacc-cfa848ac0f2f" containerName="nova-cell0-conductor-db-sync" Nov 21 14:35:21 crc kubenswrapper[4897]: I1121 14:35:21.873795 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6e90788-4f97-45c0-aacc-cfa848ac0f2f" containerName="nova-cell0-conductor-db-sync" Nov 21 14:35:21 crc kubenswrapper[4897]: I1121 14:35:21.874739 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 21 14:35:21 crc kubenswrapper[4897]: I1121 14:35:21.877651 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Nov 21 14:35:21 crc kubenswrapper[4897]: I1121 14:35:21.877859 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-26nl9" Nov 21 14:35:21 crc kubenswrapper[4897]: I1121 14:35:21.893368 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 21 14:35:21 crc kubenswrapper[4897]: I1121 14:35:21.951051 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dr77s\" (UniqueName: \"kubernetes.io/projected/889200a7-8f94-4186-ae27-4a5a9c408146-kube-api-access-dr77s\") pod \"nova-cell0-conductor-0\" (UID: \"889200a7-8f94-4186-ae27-4a5a9c408146\") " pod="openstack/nova-cell0-conductor-0" Nov 21 14:35:21 crc kubenswrapper[4897]: I1121 14:35:21.951143 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/889200a7-8f94-4186-ae27-4a5a9c408146-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"889200a7-8f94-4186-ae27-4a5a9c408146\") " pod="openstack/nova-cell0-conductor-0" Nov 21 14:35:21 crc kubenswrapper[4897]: I1121 14:35:21.951215 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/889200a7-8f94-4186-ae27-4a5a9c408146-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"889200a7-8f94-4186-ae27-4a5a9c408146\") " pod="openstack/nova-cell0-conductor-0" Nov 21 14:35:22 crc kubenswrapper[4897]: I1121 14:35:22.053806 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/889200a7-8f94-4186-ae27-4a5a9c408146-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"889200a7-8f94-4186-ae27-4a5a9c408146\") " pod="openstack/nova-cell0-conductor-0" Nov 21 14:35:22 crc kubenswrapper[4897]: I1121 14:35:22.053981 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/889200a7-8f94-4186-ae27-4a5a9c408146-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"889200a7-8f94-4186-ae27-4a5a9c408146\") " pod="openstack/nova-cell0-conductor-0" Nov 21 14:35:22 crc kubenswrapper[4897]: I1121 14:35:22.054235 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dr77s\" (UniqueName: \"kubernetes.io/projected/889200a7-8f94-4186-ae27-4a5a9c408146-kube-api-access-dr77s\") pod \"nova-cell0-conductor-0\" (UID: \"889200a7-8f94-4186-ae27-4a5a9c408146\") " pod="openstack/nova-cell0-conductor-0" Nov 21 14:35:22 crc kubenswrapper[4897]: I1121 14:35:22.058295 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/889200a7-8f94-4186-ae27-4a5a9c408146-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"889200a7-8f94-4186-ae27-4a5a9c408146\") " pod="openstack/nova-cell0-conductor-0" Nov 21 14:35:22 crc kubenswrapper[4897]: I1121 14:35:22.058323 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/889200a7-8f94-4186-ae27-4a5a9c408146-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"889200a7-8f94-4186-ae27-4a5a9c408146\") " pod="openstack/nova-cell0-conductor-0" Nov 21 14:35:22 crc kubenswrapper[4897]: I1121 14:35:22.070956 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dr77s\" (UniqueName: \"kubernetes.io/projected/889200a7-8f94-4186-ae27-4a5a9c408146-kube-api-access-dr77s\") pod \"nova-cell0-conductor-0\" (UID: \"889200a7-8f94-4186-ae27-4a5a9c408146\") " pod="openstack/nova-cell0-conductor-0" Nov 21 14:35:22 crc kubenswrapper[4897]: I1121 14:35:22.202158 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 21 14:35:23 crc kubenswrapper[4897]: I1121 14:35:23.265781 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 21 14:35:23 crc kubenswrapper[4897]: I1121 14:35:23.530249 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 21 14:35:23 crc kubenswrapper[4897]: I1121 14:35:23.695799 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ac00807a-ef51-4204-af3f-efddb3dc8af4-log-httpd\") pod \"ac00807a-ef51-4204-af3f-efddb3dc8af4\" (UID: \"ac00807a-ef51-4204-af3f-efddb3dc8af4\") " Nov 21 14:35:23 crc kubenswrapper[4897]: I1121 14:35:23.695903 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ac00807a-ef51-4204-af3f-efddb3dc8af4-ceilometer-tls-certs\") pod \"ac00807a-ef51-4204-af3f-efddb3dc8af4\" (UID: \"ac00807a-ef51-4204-af3f-efddb3dc8af4\") " Nov 21 14:35:23 crc kubenswrapper[4897]: I1121 14:35:23.696010 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac00807a-ef51-4204-af3f-efddb3dc8af4-config-data\") pod \"ac00807a-ef51-4204-af3f-efddb3dc8af4\" (UID: \"ac00807a-ef51-4204-af3f-efddb3dc8af4\") " Nov 21 14:35:23 crc kubenswrapper[4897]: I1121 14:35:23.696098 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ac00807a-ef51-4204-af3f-efddb3dc8af4-run-httpd\") pod \"ac00807a-ef51-4204-af3f-efddb3dc8af4\" (UID: \"ac00807a-ef51-4204-af3f-efddb3dc8af4\") " Nov 21 14:35:23 crc kubenswrapper[4897]: I1121 14:35:23.696151 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ac00807a-ef51-4204-af3f-efddb3dc8af4-scripts\") pod \"ac00807a-ef51-4204-af3f-efddb3dc8af4\" (UID: \"ac00807a-ef51-4204-af3f-efddb3dc8af4\") " Nov 21 14:35:23 crc kubenswrapper[4897]: I1121 14:35:23.696559 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ac00807a-ef51-4204-af3f-efddb3dc8af4-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "ac00807a-ef51-4204-af3f-efddb3dc8af4" (UID: "ac00807a-ef51-4204-af3f-efddb3dc8af4"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:35:23 crc kubenswrapper[4897]: I1121 14:35:23.696609 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ac00807a-ef51-4204-af3f-efddb3dc8af4-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "ac00807a-ef51-4204-af3f-efddb3dc8af4" (UID: "ac00807a-ef51-4204-af3f-efddb3dc8af4"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:35:23 crc kubenswrapper[4897]: I1121 14:35:23.696269 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac00807a-ef51-4204-af3f-efddb3dc8af4-combined-ca-bundle\") pod \"ac00807a-ef51-4204-af3f-efddb3dc8af4\" (UID: \"ac00807a-ef51-4204-af3f-efddb3dc8af4\") " Nov 21 14:35:23 crc kubenswrapper[4897]: I1121 14:35:23.697705 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ac00807a-ef51-4204-af3f-efddb3dc8af4-sg-core-conf-yaml\") pod \"ac00807a-ef51-4204-af3f-efddb3dc8af4\" (UID: \"ac00807a-ef51-4204-af3f-efddb3dc8af4\") " Nov 21 14:35:23 crc kubenswrapper[4897]: I1121 14:35:23.697769 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-285dv\" (UniqueName: \"kubernetes.io/projected/ac00807a-ef51-4204-af3f-efddb3dc8af4-kube-api-access-285dv\") pod \"ac00807a-ef51-4204-af3f-efddb3dc8af4\" (UID: \"ac00807a-ef51-4204-af3f-efddb3dc8af4\") " Nov 21 14:35:23 crc kubenswrapper[4897]: I1121 14:35:23.701901 4897 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ac00807a-ef51-4204-af3f-efddb3dc8af4-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 21 14:35:23 crc kubenswrapper[4897]: I1121 14:35:23.701942 4897 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ac00807a-ef51-4204-af3f-efddb3dc8af4-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 21 14:35:23 crc kubenswrapper[4897]: I1121 14:35:23.702447 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac00807a-ef51-4204-af3f-efddb3dc8af4-scripts" (OuterVolumeSpecName: "scripts") pod "ac00807a-ef51-4204-af3f-efddb3dc8af4" (UID: "ac00807a-ef51-4204-af3f-efddb3dc8af4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:35:23 crc kubenswrapper[4897]: I1121 14:35:23.704579 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac00807a-ef51-4204-af3f-efddb3dc8af4-kube-api-access-285dv" (OuterVolumeSpecName: "kube-api-access-285dv") pod "ac00807a-ef51-4204-af3f-efddb3dc8af4" (UID: "ac00807a-ef51-4204-af3f-efddb3dc8af4"). InnerVolumeSpecName "kube-api-access-285dv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:35:23 crc kubenswrapper[4897]: I1121 14:35:23.732936 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac00807a-ef51-4204-af3f-efddb3dc8af4-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "ac00807a-ef51-4204-af3f-efddb3dc8af4" (UID: "ac00807a-ef51-4204-af3f-efddb3dc8af4"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:35:23 crc kubenswrapper[4897]: I1121 14:35:23.769202 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac00807a-ef51-4204-af3f-efddb3dc8af4-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "ac00807a-ef51-4204-af3f-efddb3dc8af4" (UID: "ac00807a-ef51-4204-af3f-efddb3dc8af4"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:35:23 crc kubenswrapper[4897]: I1121 14:35:23.796083 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac00807a-ef51-4204-af3f-efddb3dc8af4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ac00807a-ef51-4204-af3f-efddb3dc8af4" (UID: "ac00807a-ef51-4204-af3f-efddb3dc8af4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:35:23 crc kubenswrapper[4897]: I1121 14:35:23.803951 4897 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ac00807a-ef51-4204-af3f-efddb3dc8af4-scripts\") on node \"crc\" DevicePath \"\"" Nov 21 14:35:23 crc kubenswrapper[4897]: I1121 14:35:23.803986 4897 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac00807a-ef51-4204-af3f-efddb3dc8af4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 21 14:35:23 crc kubenswrapper[4897]: I1121 14:35:23.803998 4897 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ac00807a-ef51-4204-af3f-efddb3dc8af4-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 21 14:35:23 crc kubenswrapper[4897]: I1121 14:35:23.804007 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-285dv\" (UniqueName: \"kubernetes.io/projected/ac00807a-ef51-4204-af3f-efddb3dc8af4-kube-api-access-285dv\") on node \"crc\" DevicePath \"\"" Nov 21 14:35:23 crc kubenswrapper[4897]: I1121 14:35:23.804015 4897 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ac00807a-ef51-4204-af3f-efddb3dc8af4-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 21 14:35:23 crc kubenswrapper[4897]: I1121 14:35:23.823153 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac00807a-ef51-4204-af3f-efddb3dc8af4-config-data" (OuterVolumeSpecName: "config-data") pod "ac00807a-ef51-4204-af3f-efddb3dc8af4" (UID: "ac00807a-ef51-4204-af3f-efddb3dc8af4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:35:23 crc kubenswrapper[4897]: I1121 14:35:23.834318 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"889200a7-8f94-4186-ae27-4a5a9c408146","Type":"ContainerStarted","Data":"ad8e3ea1653b3a77ca2d0868caeb1774787c20c8134f666b55002d3a77e0da73"} Nov 21 14:35:23 crc kubenswrapper[4897]: I1121 14:35:23.834366 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"889200a7-8f94-4186-ae27-4a5a9c408146","Type":"ContainerStarted","Data":"16fad5f6b5df07df89044c4f6e15c0b4c109bb57909eb2902ee02f61f97de3d7"} Nov 21 14:35:23 crc kubenswrapper[4897]: I1121 14:35:23.835153 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Nov 21 14:35:23 crc kubenswrapper[4897]: I1121 14:35:23.841690 4897 generic.go:334] "Generic (PLEG): container finished" podID="ac00807a-ef51-4204-af3f-efddb3dc8af4" containerID="eb938a4500d5755dc62df995062319c2c3965a09b16cde73e9ac8a686ade6a5f" exitCode=0 Nov 21 14:35:23 crc kubenswrapper[4897]: I1121 14:35:23.841758 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ac00807a-ef51-4204-af3f-efddb3dc8af4","Type":"ContainerDied","Data":"eb938a4500d5755dc62df995062319c2c3965a09b16cde73e9ac8a686ade6a5f"} Nov 21 14:35:23 crc kubenswrapper[4897]: I1121 14:35:23.841798 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ac00807a-ef51-4204-af3f-efddb3dc8af4","Type":"ContainerDied","Data":"d3d55b20f98ec93fcdba4bae00e1b09606030a61abe1568c168bdee50946675d"} Nov 21 14:35:23 crc kubenswrapper[4897]: I1121 14:35:23.841824 4897 scope.go:117] "RemoveContainer" containerID="d9d5c15f7a13479759204052fc598454eb624a9eb7e0e665e01e2425ab601055" Nov 21 14:35:23 crc kubenswrapper[4897]: I1121 14:35:23.841916 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 21 14:35:23 crc kubenswrapper[4897]: I1121 14:35:23.863440 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.863420153 podStartE2EDuration="2.863420153s" podCreationTimestamp="2025-11-21 14:35:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:35:23.850415864 +0000 UTC m=+1601.135009339" watchObservedRunningTime="2025-11-21 14:35:23.863420153 +0000 UTC m=+1601.148013628" Nov 21 14:35:23 crc kubenswrapper[4897]: I1121 14:35:23.886727 4897 scope.go:117] "RemoveContainer" containerID="2e4b3e2467588683943bf462488555daf357e16195e37ba10243bf69ae183f54" Nov 21 14:35:23 crc kubenswrapper[4897]: I1121 14:35:23.902562 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 21 14:35:23 crc kubenswrapper[4897]: I1121 14:35:23.915256 4897 scope.go:117] "RemoveContainer" containerID="a475231d66f3729ab90ffc303ef53bc90030a074d7538fb1da4803ea215ce967" Nov 21 14:35:23 crc kubenswrapper[4897]: I1121 14:35:23.916213 4897 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac00807a-ef51-4204-af3f-efddb3dc8af4-config-data\") on node \"crc\" DevicePath \"\"" Nov 21 14:35:23 crc kubenswrapper[4897]: I1121 14:35:23.917310 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 21 14:35:23 crc kubenswrapper[4897]: I1121 14:35:23.931431 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 21 14:35:23 crc kubenswrapper[4897]: E1121 14:35:23.932019 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac00807a-ef51-4204-af3f-efddb3dc8af4" containerName="ceilometer-central-agent" Nov 21 14:35:23 crc kubenswrapper[4897]: I1121 14:35:23.932038 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac00807a-ef51-4204-af3f-efddb3dc8af4" containerName="ceilometer-central-agent" Nov 21 14:35:23 crc kubenswrapper[4897]: E1121 14:35:23.932057 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac00807a-ef51-4204-af3f-efddb3dc8af4" containerName="sg-core" Nov 21 14:35:23 crc kubenswrapper[4897]: I1121 14:35:23.932063 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac00807a-ef51-4204-af3f-efddb3dc8af4" containerName="sg-core" Nov 21 14:35:23 crc kubenswrapper[4897]: E1121 14:35:23.932087 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac00807a-ef51-4204-af3f-efddb3dc8af4" containerName="ceilometer-notification-agent" Nov 21 14:35:23 crc kubenswrapper[4897]: I1121 14:35:23.932094 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac00807a-ef51-4204-af3f-efddb3dc8af4" containerName="ceilometer-notification-agent" Nov 21 14:35:23 crc kubenswrapper[4897]: E1121 14:35:23.932119 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac00807a-ef51-4204-af3f-efddb3dc8af4" containerName="proxy-httpd" Nov 21 14:35:23 crc kubenswrapper[4897]: I1121 14:35:23.932124 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac00807a-ef51-4204-af3f-efddb3dc8af4" containerName="proxy-httpd" Nov 21 14:35:23 crc kubenswrapper[4897]: I1121 14:35:23.932355 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac00807a-ef51-4204-af3f-efddb3dc8af4" containerName="ceilometer-central-agent" Nov 21 14:35:23 crc kubenswrapper[4897]: I1121 14:35:23.932376 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac00807a-ef51-4204-af3f-efddb3dc8af4" containerName="sg-core" Nov 21 14:35:23 crc kubenswrapper[4897]: I1121 14:35:23.932383 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac00807a-ef51-4204-af3f-efddb3dc8af4" containerName="proxy-httpd" Nov 21 14:35:23 crc kubenswrapper[4897]: I1121 14:35:23.932406 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac00807a-ef51-4204-af3f-efddb3dc8af4" containerName="ceilometer-notification-agent" Nov 21 14:35:23 crc kubenswrapper[4897]: I1121 14:35:23.934353 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 21 14:35:23 crc kubenswrapper[4897]: I1121 14:35:23.938879 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 21 14:35:23 crc kubenswrapper[4897]: I1121 14:35:23.939137 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 21 14:35:23 crc kubenswrapper[4897]: I1121 14:35:23.939362 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 21 14:35:23 crc kubenswrapper[4897]: I1121 14:35:23.944499 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 21 14:35:23 crc kubenswrapper[4897]: I1121 14:35:23.955065 4897 scope.go:117] "RemoveContainer" containerID="eb938a4500d5755dc62df995062319c2c3965a09b16cde73e9ac8a686ade6a5f" Nov 21 14:35:23 crc kubenswrapper[4897]: I1121 14:35:23.994210 4897 scope.go:117] "RemoveContainer" containerID="d9d5c15f7a13479759204052fc598454eb624a9eb7e0e665e01e2425ab601055" Nov 21 14:35:23 crc kubenswrapper[4897]: E1121 14:35:23.994618 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d9d5c15f7a13479759204052fc598454eb624a9eb7e0e665e01e2425ab601055\": container with ID starting with d9d5c15f7a13479759204052fc598454eb624a9eb7e0e665e01e2425ab601055 not found: ID does not exist" containerID="d9d5c15f7a13479759204052fc598454eb624a9eb7e0e665e01e2425ab601055" Nov 21 14:35:23 crc kubenswrapper[4897]: I1121 14:35:23.994657 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9d5c15f7a13479759204052fc598454eb624a9eb7e0e665e01e2425ab601055"} err="failed to get container status \"d9d5c15f7a13479759204052fc598454eb624a9eb7e0e665e01e2425ab601055\": rpc error: code = NotFound desc = could not find container \"d9d5c15f7a13479759204052fc598454eb624a9eb7e0e665e01e2425ab601055\": container with ID starting with d9d5c15f7a13479759204052fc598454eb624a9eb7e0e665e01e2425ab601055 not found: ID does not exist" Nov 21 14:35:23 crc kubenswrapper[4897]: I1121 14:35:23.994680 4897 scope.go:117] "RemoveContainer" containerID="2e4b3e2467588683943bf462488555daf357e16195e37ba10243bf69ae183f54" Nov 21 14:35:23 crc kubenswrapper[4897]: E1121 14:35:23.995078 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2e4b3e2467588683943bf462488555daf357e16195e37ba10243bf69ae183f54\": container with ID starting with 2e4b3e2467588683943bf462488555daf357e16195e37ba10243bf69ae183f54 not found: ID does not exist" containerID="2e4b3e2467588683943bf462488555daf357e16195e37ba10243bf69ae183f54" Nov 21 14:35:23 crc kubenswrapper[4897]: I1121 14:35:23.995109 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e4b3e2467588683943bf462488555daf357e16195e37ba10243bf69ae183f54"} err="failed to get container status \"2e4b3e2467588683943bf462488555daf357e16195e37ba10243bf69ae183f54\": rpc error: code = NotFound desc = could not find container \"2e4b3e2467588683943bf462488555daf357e16195e37ba10243bf69ae183f54\": container with ID starting with 2e4b3e2467588683943bf462488555daf357e16195e37ba10243bf69ae183f54 not found: ID does not exist" Nov 21 14:35:23 crc kubenswrapper[4897]: I1121 14:35:23.995127 4897 scope.go:117] "RemoveContainer" containerID="a475231d66f3729ab90ffc303ef53bc90030a074d7538fb1da4803ea215ce967" Nov 21 14:35:23 crc kubenswrapper[4897]: E1121 14:35:23.995452 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a475231d66f3729ab90ffc303ef53bc90030a074d7538fb1da4803ea215ce967\": container with ID starting with a475231d66f3729ab90ffc303ef53bc90030a074d7538fb1da4803ea215ce967 not found: ID does not exist" containerID="a475231d66f3729ab90ffc303ef53bc90030a074d7538fb1da4803ea215ce967" Nov 21 14:35:23 crc kubenswrapper[4897]: I1121 14:35:23.995483 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a475231d66f3729ab90ffc303ef53bc90030a074d7538fb1da4803ea215ce967"} err="failed to get container status \"a475231d66f3729ab90ffc303ef53bc90030a074d7538fb1da4803ea215ce967\": rpc error: code = NotFound desc = could not find container \"a475231d66f3729ab90ffc303ef53bc90030a074d7538fb1da4803ea215ce967\": container with ID starting with a475231d66f3729ab90ffc303ef53bc90030a074d7538fb1da4803ea215ce967 not found: ID does not exist" Nov 21 14:35:23 crc kubenswrapper[4897]: I1121 14:35:23.995535 4897 scope.go:117] "RemoveContainer" containerID="eb938a4500d5755dc62df995062319c2c3965a09b16cde73e9ac8a686ade6a5f" Nov 21 14:35:23 crc kubenswrapper[4897]: E1121 14:35:23.995923 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eb938a4500d5755dc62df995062319c2c3965a09b16cde73e9ac8a686ade6a5f\": container with ID starting with eb938a4500d5755dc62df995062319c2c3965a09b16cde73e9ac8a686ade6a5f not found: ID does not exist" containerID="eb938a4500d5755dc62df995062319c2c3965a09b16cde73e9ac8a686ade6a5f" Nov 21 14:35:23 crc kubenswrapper[4897]: I1121 14:35:23.995969 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb938a4500d5755dc62df995062319c2c3965a09b16cde73e9ac8a686ade6a5f"} err="failed to get container status \"eb938a4500d5755dc62df995062319c2c3965a09b16cde73e9ac8a686ade6a5f\": rpc error: code = NotFound desc = could not find container \"eb938a4500d5755dc62df995062319c2c3965a09b16cde73e9ac8a686ade6a5f\": container with ID starting with eb938a4500d5755dc62df995062319c2c3965a09b16cde73e9ac8a686ade6a5f not found: ID does not exist" Nov 21 14:35:24 crc kubenswrapper[4897]: I1121 14:35:24.123099 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/efed1737-5de8-4313-9127-724bbd239f1a-scripts\") pod \"ceilometer-0\" (UID: \"efed1737-5de8-4313-9127-724bbd239f1a\") " pod="openstack/ceilometer-0" Nov 21 14:35:24 crc kubenswrapper[4897]: I1121 14:35:24.123176 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/efed1737-5de8-4313-9127-724bbd239f1a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"efed1737-5de8-4313-9127-724bbd239f1a\") " pod="openstack/ceilometer-0" Nov 21 14:35:24 crc kubenswrapper[4897]: I1121 14:35:24.123250 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/efed1737-5de8-4313-9127-724bbd239f1a-config-data\") pod \"ceilometer-0\" (UID: \"efed1737-5de8-4313-9127-724bbd239f1a\") " pod="openstack/ceilometer-0" Nov 21 14:35:24 crc kubenswrapper[4897]: I1121 14:35:24.123315 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-86hmv\" (UniqueName: \"kubernetes.io/projected/efed1737-5de8-4313-9127-724bbd239f1a-kube-api-access-86hmv\") pod \"ceilometer-0\" (UID: \"efed1737-5de8-4313-9127-724bbd239f1a\") " pod="openstack/ceilometer-0" Nov 21 14:35:24 crc kubenswrapper[4897]: I1121 14:35:24.123535 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/efed1737-5de8-4313-9127-724bbd239f1a-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"efed1737-5de8-4313-9127-724bbd239f1a\") " pod="openstack/ceilometer-0" Nov 21 14:35:24 crc kubenswrapper[4897]: I1121 14:35:24.123795 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/efed1737-5de8-4313-9127-724bbd239f1a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"efed1737-5de8-4313-9127-724bbd239f1a\") " pod="openstack/ceilometer-0" Nov 21 14:35:24 crc kubenswrapper[4897]: I1121 14:35:24.124080 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/efed1737-5de8-4313-9127-724bbd239f1a-log-httpd\") pod \"ceilometer-0\" (UID: \"efed1737-5de8-4313-9127-724bbd239f1a\") " pod="openstack/ceilometer-0" Nov 21 14:35:24 crc kubenswrapper[4897]: I1121 14:35:24.124250 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/efed1737-5de8-4313-9127-724bbd239f1a-run-httpd\") pod \"ceilometer-0\" (UID: \"efed1737-5de8-4313-9127-724bbd239f1a\") " pod="openstack/ceilometer-0" Nov 21 14:35:24 crc kubenswrapper[4897]: I1121 14:35:24.128300 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ac00807a-ef51-4204-af3f-efddb3dc8af4" path="/var/lib/kubelet/pods/ac00807a-ef51-4204-af3f-efddb3dc8af4/volumes" Nov 21 14:35:24 crc kubenswrapper[4897]: I1121 14:35:24.226815 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/efed1737-5de8-4313-9127-724bbd239f1a-log-httpd\") pod \"ceilometer-0\" (UID: \"efed1737-5de8-4313-9127-724bbd239f1a\") " pod="openstack/ceilometer-0" Nov 21 14:35:24 crc kubenswrapper[4897]: I1121 14:35:24.226883 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/efed1737-5de8-4313-9127-724bbd239f1a-run-httpd\") pod \"ceilometer-0\" (UID: \"efed1737-5de8-4313-9127-724bbd239f1a\") " pod="openstack/ceilometer-0" Nov 21 14:35:24 crc kubenswrapper[4897]: I1121 14:35:24.226941 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/efed1737-5de8-4313-9127-724bbd239f1a-scripts\") pod \"ceilometer-0\" (UID: \"efed1737-5de8-4313-9127-724bbd239f1a\") " pod="openstack/ceilometer-0" Nov 21 14:35:24 crc kubenswrapper[4897]: I1121 14:35:24.226965 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/efed1737-5de8-4313-9127-724bbd239f1a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"efed1737-5de8-4313-9127-724bbd239f1a\") " pod="openstack/ceilometer-0" Nov 21 14:35:24 crc kubenswrapper[4897]: I1121 14:35:24.227006 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/efed1737-5de8-4313-9127-724bbd239f1a-config-data\") pod \"ceilometer-0\" (UID: \"efed1737-5de8-4313-9127-724bbd239f1a\") " pod="openstack/ceilometer-0" Nov 21 14:35:24 crc kubenswrapper[4897]: I1121 14:35:24.227042 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-86hmv\" (UniqueName: \"kubernetes.io/projected/efed1737-5de8-4313-9127-724bbd239f1a-kube-api-access-86hmv\") pod \"ceilometer-0\" (UID: \"efed1737-5de8-4313-9127-724bbd239f1a\") " pod="openstack/ceilometer-0" Nov 21 14:35:24 crc kubenswrapper[4897]: I1121 14:35:24.227071 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/efed1737-5de8-4313-9127-724bbd239f1a-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"efed1737-5de8-4313-9127-724bbd239f1a\") " pod="openstack/ceilometer-0" Nov 21 14:35:24 crc kubenswrapper[4897]: I1121 14:35:24.227136 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/efed1737-5de8-4313-9127-724bbd239f1a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"efed1737-5de8-4313-9127-724bbd239f1a\") " pod="openstack/ceilometer-0" Nov 21 14:35:24 crc kubenswrapper[4897]: I1121 14:35:24.227359 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/efed1737-5de8-4313-9127-724bbd239f1a-log-httpd\") pod \"ceilometer-0\" (UID: \"efed1737-5de8-4313-9127-724bbd239f1a\") " pod="openstack/ceilometer-0" Nov 21 14:35:24 crc kubenswrapper[4897]: I1121 14:35:24.228271 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/efed1737-5de8-4313-9127-724bbd239f1a-run-httpd\") pod \"ceilometer-0\" (UID: \"efed1737-5de8-4313-9127-724bbd239f1a\") " pod="openstack/ceilometer-0" Nov 21 14:35:24 crc kubenswrapper[4897]: I1121 14:35:24.235382 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/efed1737-5de8-4313-9127-724bbd239f1a-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"efed1737-5de8-4313-9127-724bbd239f1a\") " pod="openstack/ceilometer-0" Nov 21 14:35:24 crc kubenswrapper[4897]: I1121 14:35:24.235764 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/efed1737-5de8-4313-9127-724bbd239f1a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"efed1737-5de8-4313-9127-724bbd239f1a\") " pod="openstack/ceilometer-0" Nov 21 14:35:24 crc kubenswrapper[4897]: I1121 14:35:24.236953 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/efed1737-5de8-4313-9127-724bbd239f1a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"efed1737-5de8-4313-9127-724bbd239f1a\") " pod="openstack/ceilometer-0" Nov 21 14:35:24 crc kubenswrapper[4897]: I1121 14:35:24.238851 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/efed1737-5de8-4313-9127-724bbd239f1a-config-data\") pod \"ceilometer-0\" (UID: \"efed1737-5de8-4313-9127-724bbd239f1a\") " pod="openstack/ceilometer-0" Nov 21 14:35:24 crc kubenswrapper[4897]: I1121 14:35:24.239626 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/efed1737-5de8-4313-9127-724bbd239f1a-scripts\") pod \"ceilometer-0\" (UID: \"efed1737-5de8-4313-9127-724bbd239f1a\") " pod="openstack/ceilometer-0" Nov 21 14:35:24 crc kubenswrapper[4897]: I1121 14:35:24.245451 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-86hmv\" (UniqueName: \"kubernetes.io/projected/efed1737-5de8-4313-9127-724bbd239f1a-kube-api-access-86hmv\") pod \"ceilometer-0\" (UID: \"efed1737-5de8-4313-9127-724bbd239f1a\") " pod="openstack/ceilometer-0" Nov 21 14:35:24 crc kubenswrapper[4897]: I1121 14:35:24.253131 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 21 14:35:24 crc kubenswrapper[4897]: I1121 14:35:24.759560 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 21 14:35:24 crc kubenswrapper[4897]: W1121 14:35:24.765193 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podefed1737_5de8_4313_9127_724bbd239f1a.slice/crio-159cd7aa94cac5b32c8dbee75041c735a73b20d7d9cac3a941696bf88c467720 WatchSource:0}: Error finding container 159cd7aa94cac5b32c8dbee75041c735a73b20d7d9cac3a941696bf88c467720: Status 404 returned error can't find the container with id 159cd7aa94cac5b32c8dbee75041c735a73b20d7d9cac3a941696bf88c467720 Nov 21 14:35:24 crc kubenswrapper[4897]: I1121 14:35:24.853185 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"efed1737-5de8-4313-9127-724bbd239f1a","Type":"ContainerStarted","Data":"159cd7aa94cac5b32c8dbee75041c735a73b20d7d9cac3a941696bf88c467720"} Nov 21 14:35:26 crc kubenswrapper[4897]: I1121 14:35:26.176548 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-dtwql" podUID="6d3cad20-fb17-4726-b8fa-9fd5303b5eea" containerName="registry-server" probeResult="failure" output=< Nov 21 14:35:26 crc kubenswrapper[4897]: timeout: failed to connect service ":50051" within 1s Nov 21 14:35:26 crc kubenswrapper[4897]: > Nov 21 14:35:26 crc kubenswrapper[4897]: I1121 14:35:26.176634 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-dtwql" Nov 21 14:35:26 crc kubenswrapper[4897]: I1121 14:35:26.177516 4897 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="registry-server" containerStatusID={"Type":"cri-o","ID":"36d1ba740cdac40a73ac8d07a9d49188d3f951e9e6e9797089e9005483687829"} pod="openshift-marketplace/redhat-operators-dtwql" containerMessage="Container registry-server failed startup probe, will be restarted" Nov 21 14:35:26 crc kubenswrapper[4897]: I1121 14:35:26.177545 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-dtwql" podUID="6d3cad20-fb17-4726-b8fa-9fd5303b5eea" containerName="registry-server" containerID="cri-o://36d1ba740cdac40a73ac8d07a9d49188d3f951e9e6e9797089e9005483687829" gracePeriod=30 Nov 21 14:35:27 crc kubenswrapper[4897]: I1121 14:35:27.307997 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Nov 21 14:35:29 crc kubenswrapper[4897]: I1121 14:35:29.915002 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"efed1737-5de8-4313-9127-724bbd239f1a","Type":"ContainerStarted","Data":"5c1ffe3f03b119bb9d7d244acd512729212c30aafbda3158b86b3b645cdc77e5"} Nov 21 14:35:30 crc kubenswrapper[4897]: I1121 14:35:30.928561 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"efed1737-5de8-4313-9127-724bbd239f1a","Type":"ContainerStarted","Data":"37c0701bd55a3a69856270d224fe1f7c29a3befbcc9c45fb1e1596e394a69933"} Nov 21 14:35:31 crc kubenswrapper[4897]: I1121 14:35:31.089055 4897 scope.go:117] "RemoveContainer" containerID="c68f1b003d12b4a2964169a85a04e3f0272d86c891f52e6bebd52f3781520d1a" Nov 21 14:35:31 crc kubenswrapper[4897]: E1121 14:35:31.089527 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 14:35:31 crc kubenswrapper[4897]: I1121 14:35:31.946982 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"efed1737-5de8-4313-9127-724bbd239f1a","Type":"ContainerStarted","Data":"aba81a8a1224fdc2f5618aeb6eb3be0ce285ed4b26b0e1d6ceaf5c836c7536a1"} Nov 21 14:35:32 crc kubenswrapper[4897]: I1121 14:35:32.236746 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Nov 21 14:35:32 crc kubenswrapper[4897]: I1121 14:35:32.729415 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-x8ttp"] Nov 21 14:35:32 crc kubenswrapper[4897]: I1121 14:35:32.731425 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-x8ttp" Nov 21 14:35:32 crc kubenswrapper[4897]: I1121 14:35:32.733845 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Nov 21 14:35:32 crc kubenswrapper[4897]: I1121 14:35:32.734078 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Nov 21 14:35:32 crc kubenswrapper[4897]: I1121 14:35:32.747980 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-x8ttp"] Nov 21 14:35:32 crc kubenswrapper[4897]: I1121 14:35:32.832373 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/518b64a4-d073-462f-ada5-20209e14ff27-scripts\") pod \"nova-cell0-cell-mapping-x8ttp\" (UID: \"518b64a4-d073-462f-ada5-20209e14ff27\") " pod="openstack/nova-cell0-cell-mapping-x8ttp" Nov 21 14:35:32 crc kubenswrapper[4897]: I1121 14:35:32.832605 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/518b64a4-d073-462f-ada5-20209e14ff27-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-x8ttp\" (UID: \"518b64a4-d073-462f-ada5-20209e14ff27\") " pod="openstack/nova-cell0-cell-mapping-x8ttp" Nov 21 14:35:32 crc kubenswrapper[4897]: I1121 14:35:32.832643 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/518b64a4-d073-462f-ada5-20209e14ff27-config-data\") pod \"nova-cell0-cell-mapping-x8ttp\" (UID: \"518b64a4-d073-462f-ada5-20209e14ff27\") " pod="openstack/nova-cell0-cell-mapping-x8ttp" Nov 21 14:35:32 crc kubenswrapper[4897]: I1121 14:35:32.832712 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qx8n8\" (UniqueName: \"kubernetes.io/projected/518b64a4-d073-462f-ada5-20209e14ff27-kube-api-access-qx8n8\") pod \"nova-cell0-cell-mapping-x8ttp\" (UID: \"518b64a4-d073-462f-ada5-20209e14ff27\") " pod="openstack/nova-cell0-cell-mapping-x8ttp" Nov 21 14:35:32 crc kubenswrapper[4897]: I1121 14:35:32.905674 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 21 14:35:32 crc kubenswrapper[4897]: I1121 14:35:32.907836 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 21 14:35:32 crc kubenswrapper[4897]: I1121 14:35:32.918069 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 21 14:35:32 crc kubenswrapper[4897]: I1121 14:35:32.926023 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 21 14:35:32 crc kubenswrapper[4897]: I1121 14:35:32.935003 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/518b64a4-d073-462f-ada5-20209e14ff27-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-x8ttp\" (UID: \"518b64a4-d073-462f-ada5-20209e14ff27\") " pod="openstack/nova-cell0-cell-mapping-x8ttp" Nov 21 14:35:32 crc kubenswrapper[4897]: I1121 14:35:32.935063 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/518b64a4-d073-462f-ada5-20209e14ff27-config-data\") pod \"nova-cell0-cell-mapping-x8ttp\" (UID: \"518b64a4-d073-462f-ada5-20209e14ff27\") " pod="openstack/nova-cell0-cell-mapping-x8ttp" Nov 21 14:35:32 crc kubenswrapper[4897]: I1121 14:35:32.935156 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qx8n8\" (UniqueName: \"kubernetes.io/projected/518b64a4-d073-462f-ada5-20209e14ff27-kube-api-access-qx8n8\") pod \"nova-cell0-cell-mapping-x8ttp\" (UID: \"518b64a4-d073-462f-ada5-20209e14ff27\") " pod="openstack/nova-cell0-cell-mapping-x8ttp" Nov 21 14:35:32 crc kubenswrapper[4897]: I1121 14:35:32.935313 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/518b64a4-d073-462f-ada5-20209e14ff27-scripts\") pod \"nova-cell0-cell-mapping-x8ttp\" (UID: \"518b64a4-d073-462f-ada5-20209e14ff27\") " pod="openstack/nova-cell0-cell-mapping-x8ttp" Nov 21 14:35:32 crc kubenswrapper[4897]: I1121 14:35:32.939916 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/518b64a4-d073-462f-ada5-20209e14ff27-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-x8ttp\" (UID: \"518b64a4-d073-462f-ada5-20209e14ff27\") " pod="openstack/nova-cell0-cell-mapping-x8ttp" Nov 21 14:35:32 crc kubenswrapper[4897]: I1121 14:35:32.954444 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/518b64a4-d073-462f-ada5-20209e14ff27-scripts\") pod \"nova-cell0-cell-mapping-x8ttp\" (UID: \"518b64a4-d073-462f-ada5-20209e14ff27\") " pod="openstack/nova-cell0-cell-mapping-x8ttp" Nov 21 14:35:32 crc kubenswrapper[4897]: I1121 14:35:32.968300 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/518b64a4-d073-462f-ada5-20209e14ff27-config-data\") pod \"nova-cell0-cell-mapping-x8ttp\" (UID: \"518b64a4-d073-462f-ada5-20209e14ff27\") " pod="openstack/nova-cell0-cell-mapping-x8ttp" Nov 21 14:35:32 crc kubenswrapper[4897]: I1121 14:35:32.972483 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"efed1737-5de8-4313-9127-724bbd239f1a","Type":"ContainerStarted","Data":"76bd36bbbb963c0c157f72905e2682ad3a3471d6d16c1c47c51a3e87bf44837e"} Nov 21 14:35:32 crc kubenswrapper[4897]: I1121 14:35:32.975425 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 21 14:35:32 crc kubenswrapper[4897]: I1121 14:35:32.980355 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qx8n8\" (UniqueName: \"kubernetes.io/projected/518b64a4-d073-462f-ada5-20209e14ff27-kube-api-access-qx8n8\") pod \"nova-cell0-cell-mapping-x8ttp\" (UID: \"518b64a4-d073-462f-ada5-20209e14ff27\") " pod="openstack/nova-cell0-cell-mapping-x8ttp" Nov 21 14:35:33 crc kubenswrapper[4897]: I1121 14:35:33.050864 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-x8ttp" Nov 21 14:35:33 crc kubenswrapper[4897]: I1121 14:35:33.057021 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf6d459c-961c-4361-a034-4a7c7d5ce7b8-config-data\") pod \"nova-api-0\" (UID: \"cf6d459c-961c-4361-a034-4a7c7d5ce7b8\") " pod="openstack/nova-api-0" Nov 21 14:35:33 crc kubenswrapper[4897]: I1121 14:35:33.057067 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf6d459c-961c-4361-a034-4a7c7d5ce7b8-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"cf6d459c-961c-4361-a034-4a7c7d5ce7b8\") " pod="openstack/nova-api-0" Nov 21 14:35:33 crc kubenswrapper[4897]: I1121 14:35:33.057194 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cf6d459c-961c-4361-a034-4a7c7d5ce7b8-logs\") pod \"nova-api-0\" (UID: \"cf6d459c-961c-4361-a034-4a7c7d5ce7b8\") " pod="openstack/nova-api-0" Nov 21 14:35:33 crc kubenswrapper[4897]: I1121 14:35:33.057380 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6bsn7\" (UniqueName: \"kubernetes.io/projected/cf6d459c-961c-4361-a034-4a7c7d5ce7b8-kube-api-access-6bsn7\") pod \"nova-api-0\" (UID: \"cf6d459c-961c-4361-a034-4a7c7d5ce7b8\") " pod="openstack/nova-api-0" Nov 21 14:35:33 crc kubenswrapper[4897]: I1121 14:35:33.092165 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.39784543 podStartE2EDuration="10.092140867s" podCreationTimestamp="2025-11-21 14:35:23 +0000 UTC" firstStartedPulling="2025-11-21 14:35:24.768639493 +0000 UTC m=+1602.053232968" lastFinishedPulling="2025-11-21 14:35:32.46293492 +0000 UTC m=+1609.747528405" observedRunningTime="2025-11-21 14:35:33.027957375 +0000 UTC m=+1610.312550850" watchObservedRunningTime="2025-11-21 14:35:33.092140867 +0000 UTC m=+1610.376734332" Nov 21 14:35:33 crc kubenswrapper[4897]: I1121 14:35:33.150622 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 21 14:35:33 crc kubenswrapper[4897]: I1121 14:35:33.152907 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 21 14:35:33 crc kubenswrapper[4897]: I1121 14:35:33.159113 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cf6d459c-961c-4361-a034-4a7c7d5ce7b8-logs\") pod \"nova-api-0\" (UID: \"cf6d459c-961c-4361-a034-4a7c7d5ce7b8\") " pod="openstack/nova-api-0" Nov 21 14:35:33 crc kubenswrapper[4897]: I1121 14:35:33.159265 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6bsn7\" (UniqueName: \"kubernetes.io/projected/cf6d459c-961c-4361-a034-4a7c7d5ce7b8-kube-api-access-6bsn7\") pod \"nova-api-0\" (UID: \"cf6d459c-961c-4361-a034-4a7c7d5ce7b8\") " pod="openstack/nova-api-0" Nov 21 14:35:33 crc kubenswrapper[4897]: I1121 14:35:33.159406 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf6d459c-961c-4361-a034-4a7c7d5ce7b8-config-data\") pod \"nova-api-0\" (UID: \"cf6d459c-961c-4361-a034-4a7c7d5ce7b8\") " pod="openstack/nova-api-0" Nov 21 14:35:33 crc kubenswrapper[4897]: I1121 14:35:33.159423 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf6d459c-961c-4361-a034-4a7c7d5ce7b8-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"cf6d459c-961c-4361-a034-4a7c7d5ce7b8\") " pod="openstack/nova-api-0" Nov 21 14:35:33 crc kubenswrapper[4897]: I1121 14:35:33.169444 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 21 14:35:33 crc kubenswrapper[4897]: I1121 14:35:33.170326 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cf6d459c-961c-4361-a034-4a7c7d5ce7b8-logs\") pod \"nova-api-0\" (UID: \"cf6d459c-961c-4361-a034-4a7c7d5ce7b8\") " pod="openstack/nova-api-0" Nov 21 14:35:33 crc kubenswrapper[4897]: I1121 14:35:33.181698 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf6d459c-961c-4361-a034-4a7c7d5ce7b8-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"cf6d459c-961c-4361-a034-4a7c7d5ce7b8\") " pod="openstack/nova-api-0" Nov 21 14:35:33 crc kubenswrapper[4897]: I1121 14:35:33.184637 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 21 14:35:33 crc kubenswrapper[4897]: I1121 14:35:33.189100 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf6d459c-961c-4361-a034-4a7c7d5ce7b8-config-data\") pod \"nova-api-0\" (UID: \"cf6d459c-961c-4361-a034-4a7c7d5ce7b8\") " pod="openstack/nova-api-0" Nov 21 14:35:33 crc kubenswrapper[4897]: I1121 14:35:33.189946 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6bsn7\" (UniqueName: \"kubernetes.io/projected/cf6d459c-961c-4361-a034-4a7c7d5ce7b8-kube-api-access-6bsn7\") pod \"nova-api-0\" (UID: \"cf6d459c-961c-4361-a034-4a7c7d5ce7b8\") " pod="openstack/nova-api-0" Nov 21 14:35:33 crc kubenswrapper[4897]: I1121 14:35:33.232863 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 21 14:35:33 crc kubenswrapper[4897]: I1121 14:35:33.234642 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 21 14:35:33 crc kubenswrapper[4897]: I1121 14:35:33.232804 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 21 14:35:33 crc kubenswrapper[4897]: I1121 14:35:33.237850 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Nov 21 14:35:33 crc kubenswrapper[4897]: I1121 14:35:33.268485 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ca85306-d5f3-414f-abd6-6f4623335792-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"4ca85306-d5f3-414f-abd6-6f4623335792\") " pod="openstack/nova-scheduler-0" Nov 21 14:35:33 crc kubenswrapper[4897]: I1121 14:35:33.268561 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ca85306-d5f3-414f-abd6-6f4623335792-config-data\") pod \"nova-scheduler-0\" (UID: \"4ca85306-d5f3-414f-abd6-6f4623335792\") " pod="openstack/nova-scheduler-0" Nov 21 14:35:33 crc kubenswrapper[4897]: I1121 14:35:33.268723 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qggzh\" (UniqueName: \"kubernetes.io/projected/4ca85306-d5f3-414f-abd6-6f4623335792-kube-api-access-qggzh\") pod \"nova-scheduler-0\" (UID: \"4ca85306-d5f3-414f-abd6-6f4623335792\") " pod="openstack/nova-scheduler-0" Nov 21 14:35:33 crc kubenswrapper[4897]: I1121 14:35:33.270329 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 21 14:35:33 crc kubenswrapper[4897]: I1121 14:35:33.278277 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 21 14:35:33 crc kubenswrapper[4897]: I1121 14:35:33.296210 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 21 14:35:33 crc kubenswrapper[4897]: I1121 14:35:33.367182 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 21 14:35:33 crc kubenswrapper[4897]: I1121 14:35:33.371773 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6aa3a86a-b083-40ac-a299-5c53fb5710a7-logs\") pod \"nova-metadata-0\" (UID: \"6aa3a86a-b083-40ac-a299-5c53fb5710a7\") " pod="openstack/nova-metadata-0" Nov 21 14:35:33 crc kubenswrapper[4897]: I1121 14:35:33.371851 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6aa3a86a-b083-40ac-a299-5c53fb5710a7-config-data\") pod \"nova-metadata-0\" (UID: \"6aa3a86a-b083-40ac-a299-5c53fb5710a7\") " pod="openstack/nova-metadata-0" Nov 21 14:35:33 crc kubenswrapper[4897]: I1121 14:35:33.371936 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe6f78aa-838c-4bf1-8a19-9726831f2941-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"fe6f78aa-838c-4bf1-8a19-9726831f2941\") " pod="openstack/nova-cell1-novncproxy-0" Nov 21 14:35:33 crc kubenswrapper[4897]: I1121 14:35:33.371967 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ca85306-d5f3-414f-abd6-6f4623335792-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"4ca85306-d5f3-414f-abd6-6f4623335792\") " pod="openstack/nova-scheduler-0" Nov 21 14:35:33 crc kubenswrapper[4897]: I1121 14:35:33.371993 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ca85306-d5f3-414f-abd6-6f4623335792-config-data\") pod \"nova-scheduler-0\" (UID: \"4ca85306-d5f3-414f-abd6-6f4623335792\") " pod="openstack/nova-scheduler-0" Nov 21 14:35:33 crc kubenswrapper[4897]: I1121 14:35:33.372066 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6aa3a86a-b083-40ac-a299-5c53fb5710a7-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6aa3a86a-b083-40ac-a299-5c53fb5710a7\") " pod="openstack/nova-metadata-0" Nov 21 14:35:33 crc kubenswrapper[4897]: I1121 14:35:33.372095 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8dfkf\" (UniqueName: \"kubernetes.io/projected/fe6f78aa-838c-4bf1-8a19-9726831f2941-kube-api-access-8dfkf\") pod \"nova-cell1-novncproxy-0\" (UID: \"fe6f78aa-838c-4bf1-8a19-9726831f2941\") " pod="openstack/nova-cell1-novncproxy-0" Nov 21 14:35:33 crc kubenswrapper[4897]: I1121 14:35:33.372132 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qggzh\" (UniqueName: \"kubernetes.io/projected/4ca85306-d5f3-414f-abd6-6f4623335792-kube-api-access-qggzh\") pod \"nova-scheduler-0\" (UID: \"4ca85306-d5f3-414f-abd6-6f4623335792\") " pod="openstack/nova-scheduler-0" Nov 21 14:35:33 crc kubenswrapper[4897]: I1121 14:35:33.372161 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dt4d5\" (UniqueName: \"kubernetes.io/projected/6aa3a86a-b083-40ac-a299-5c53fb5710a7-kube-api-access-dt4d5\") pod \"nova-metadata-0\" (UID: \"6aa3a86a-b083-40ac-a299-5c53fb5710a7\") " pod="openstack/nova-metadata-0" Nov 21 14:35:33 crc kubenswrapper[4897]: I1121 14:35:33.372203 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe6f78aa-838c-4bf1-8a19-9726831f2941-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"fe6f78aa-838c-4bf1-8a19-9726831f2941\") " pod="openstack/nova-cell1-novncproxy-0" Nov 21 14:35:33 crc kubenswrapper[4897]: I1121 14:35:33.379359 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ca85306-d5f3-414f-abd6-6f4623335792-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"4ca85306-d5f3-414f-abd6-6f4623335792\") " pod="openstack/nova-scheduler-0" Nov 21 14:35:33 crc kubenswrapper[4897]: I1121 14:35:33.385807 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ca85306-d5f3-414f-abd6-6f4623335792-config-data\") pod \"nova-scheduler-0\" (UID: \"4ca85306-d5f3-414f-abd6-6f4623335792\") " pod="openstack/nova-scheduler-0" Nov 21 14:35:33 crc kubenswrapper[4897]: I1121 14:35:33.387118 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 21 14:35:33 crc kubenswrapper[4897]: I1121 14:35:33.399987 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qggzh\" (UniqueName: \"kubernetes.io/projected/4ca85306-d5f3-414f-abd6-6f4623335792-kube-api-access-qggzh\") pod \"nova-scheduler-0\" (UID: \"4ca85306-d5f3-414f-abd6-6f4623335792\") " pod="openstack/nova-scheduler-0" Nov 21 14:35:33 crc kubenswrapper[4897]: I1121 14:35:33.413931 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7877d89589-c28gz"] Nov 21 14:35:33 crc kubenswrapper[4897]: I1121 14:35:33.442066 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7877d89589-c28gz" Nov 21 14:35:33 crc kubenswrapper[4897]: I1121 14:35:33.453476 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7877d89589-c28gz"] Nov 21 14:35:33 crc kubenswrapper[4897]: I1121 14:35:33.474575 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe6f78aa-838c-4bf1-8a19-9726831f2941-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"fe6f78aa-838c-4bf1-8a19-9726831f2941\") " pod="openstack/nova-cell1-novncproxy-0" Nov 21 14:35:33 crc kubenswrapper[4897]: I1121 14:35:33.474656 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6aa3a86a-b083-40ac-a299-5c53fb5710a7-logs\") pod \"nova-metadata-0\" (UID: \"6aa3a86a-b083-40ac-a299-5c53fb5710a7\") " pod="openstack/nova-metadata-0" Nov 21 14:35:33 crc kubenswrapper[4897]: I1121 14:35:33.474682 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6aa3a86a-b083-40ac-a299-5c53fb5710a7-config-data\") pod \"nova-metadata-0\" (UID: \"6aa3a86a-b083-40ac-a299-5c53fb5710a7\") " pod="openstack/nova-metadata-0" Nov 21 14:35:33 crc kubenswrapper[4897]: I1121 14:35:33.474747 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe6f78aa-838c-4bf1-8a19-9726831f2941-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"fe6f78aa-838c-4bf1-8a19-9726831f2941\") " pod="openstack/nova-cell1-novncproxy-0" Nov 21 14:35:33 crc kubenswrapper[4897]: I1121 14:35:33.474810 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6aa3a86a-b083-40ac-a299-5c53fb5710a7-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6aa3a86a-b083-40ac-a299-5c53fb5710a7\") " pod="openstack/nova-metadata-0" Nov 21 14:35:33 crc kubenswrapper[4897]: I1121 14:35:33.474830 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8dfkf\" (UniqueName: \"kubernetes.io/projected/fe6f78aa-838c-4bf1-8a19-9726831f2941-kube-api-access-8dfkf\") pod \"nova-cell1-novncproxy-0\" (UID: \"fe6f78aa-838c-4bf1-8a19-9726831f2941\") " pod="openstack/nova-cell1-novncproxy-0" Nov 21 14:35:33 crc kubenswrapper[4897]: I1121 14:35:33.474867 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dt4d5\" (UniqueName: \"kubernetes.io/projected/6aa3a86a-b083-40ac-a299-5c53fb5710a7-kube-api-access-dt4d5\") pod \"nova-metadata-0\" (UID: \"6aa3a86a-b083-40ac-a299-5c53fb5710a7\") " pod="openstack/nova-metadata-0" Nov 21 14:35:33 crc kubenswrapper[4897]: I1121 14:35:33.478465 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6aa3a86a-b083-40ac-a299-5c53fb5710a7-config-data\") pod \"nova-metadata-0\" (UID: \"6aa3a86a-b083-40ac-a299-5c53fb5710a7\") " pod="openstack/nova-metadata-0" Nov 21 14:35:33 crc kubenswrapper[4897]: I1121 14:35:33.479052 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe6f78aa-838c-4bf1-8a19-9726831f2941-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"fe6f78aa-838c-4bf1-8a19-9726831f2941\") " pod="openstack/nova-cell1-novncproxy-0" Nov 21 14:35:33 crc kubenswrapper[4897]: I1121 14:35:33.480982 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6aa3a86a-b083-40ac-a299-5c53fb5710a7-logs\") pod \"nova-metadata-0\" (UID: \"6aa3a86a-b083-40ac-a299-5c53fb5710a7\") " pod="openstack/nova-metadata-0" Nov 21 14:35:33 crc kubenswrapper[4897]: I1121 14:35:33.481929 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe6f78aa-838c-4bf1-8a19-9726831f2941-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"fe6f78aa-838c-4bf1-8a19-9726831f2941\") " pod="openstack/nova-cell1-novncproxy-0" Nov 21 14:35:33 crc kubenswrapper[4897]: I1121 14:35:33.501271 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6aa3a86a-b083-40ac-a299-5c53fb5710a7-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6aa3a86a-b083-40ac-a299-5c53fb5710a7\") " pod="openstack/nova-metadata-0" Nov 21 14:35:33 crc kubenswrapper[4897]: I1121 14:35:33.508217 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8dfkf\" (UniqueName: \"kubernetes.io/projected/fe6f78aa-838c-4bf1-8a19-9726831f2941-kube-api-access-8dfkf\") pod \"nova-cell1-novncproxy-0\" (UID: \"fe6f78aa-838c-4bf1-8a19-9726831f2941\") " pod="openstack/nova-cell1-novncproxy-0" Nov 21 14:35:33 crc kubenswrapper[4897]: I1121 14:35:33.509934 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dt4d5\" (UniqueName: \"kubernetes.io/projected/6aa3a86a-b083-40ac-a299-5c53fb5710a7-kube-api-access-dt4d5\") pod \"nova-metadata-0\" (UID: \"6aa3a86a-b083-40ac-a299-5c53fb5710a7\") " pod="openstack/nova-metadata-0" Nov 21 14:35:33 crc kubenswrapper[4897]: I1121 14:35:33.534549 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 21 14:35:33 crc kubenswrapper[4897]: I1121 14:35:33.577058 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/190e6358-2083-4549-b5ac-5c35af169b06-ovsdbserver-nb\") pod \"dnsmasq-dns-7877d89589-c28gz\" (UID: \"190e6358-2083-4549-b5ac-5c35af169b06\") " pod="openstack/dnsmasq-dns-7877d89589-c28gz" Nov 21 14:35:33 crc kubenswrapper[4897]: I1121 14:35:33.577135 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4qcr2\" (UniqueName: \"kubernetes.io/projected/190e6358-2083-4549-b5ac-5c35af169b06-kube-api-access-4qcr2\") pod \"dnsmasq-dns-7877d89589-c28gz\" (UID: \"190e6358-2083-4549-b5ac-5c35af169b06\") " pod="openstack/dnsmasq-dns-7877d89589-c28gz" Nov 21 14:35:33 crc kubenswrapper[4897]: I1121 14:35:33.577183 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/190e6358-2083-4549-b5ac-5c35af169b06-dns-svc\") pod \"dnsmasq-dns-7877d89589-c28gz\" (UID: \"190e6358-2083-4549-b5ac-5c35af169b06\") " pod="openstack/dnsmasq-dns-7877d89589-c28gz" Nov 21 14:35:33 crc kubenswrapper[4897]: I1121 14:35:33.577200 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/190e6358-2083-4549-b5ac-5c35af169b06-ovsdbserver-sb\") pod \"dnsmasq-dns-7877d89589-c28gz\" (UID: \"190e6358-2083-4549-b5ac-5c35af169b06\") " pod="openstack/dnsmasq-dns-7877d89589-c28gz" Nov 21 14:35:33 crc kubenswrapper[4897]: I1121 14:35:33.577219 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/190e6358-2083-4549-b5ac-5c35af169b06-config\") pod \"dnsmasq-dns-7877d89589-c28gz\" (UID: \"190e6358-2083-4549-b5ac-5c35af169b06\") " pod="openstack/dnsmasq-dns-7877d89589-c28gz" Nov 21 14:35:33 crc kubenswrapper[4897]: I1121 14:35:33.577245 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/190e6358-2083-4549-b5ac-5c35af169b06-dns-swift-storage-0\") pod \"dnsmasq-dns-7877d89589-c28gz\" (UID: \"190e6358-2083-4549-b5ac-5c35af169b06\") " pod="openstack/dnsmasq-dns-7877d89589-c28gz" Nov 21 14:35:33 crc kubenswrapper[4897]: I1121 14:35:33.622958 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 21 14:35:33 crc kubenswrapper[4897]: I1121 14:35:33.643112 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 21 14:35:33 crc kubenswrapper[4897]: I1121 14:35:33.679712 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/190e6358-2083-4549-b5ac-5c35af169b06-ovsdbserver-nb\") pod \"dnsmasq-dns-7877d89589-c28gz\" (UID: \"190e6358-2083-4549-b5ac-5c35af169b06\") " pod="openstack/dnsmasq-dns-7877d89589-c28gz" Nov 21 14:35:33 crc kubenswrapper[4897]: I1121 14:35:33.679795 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4qcr2\" (UniqueName: \"kubernetes.io/projected/190e6358-2083-4549-b5ac-5c35af169b06-kube-api-access-4qcr2\") pod \"dnsmasq-dns-7877d89589-c28gz\" (UID: \"190e6358-2083-4549-b5ac-5c35af169b06\") " pod="openstack/dnsmasq-dns-7877d89589-c28gz" Nov 21 14:35:33 crc kubenswrapper[4897]: I1121 14:35:33.679851 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/190e6358-2083-4549-b5ac-5c35af169b06-dns-svc\") pod \"dnsmasq-dns-7877d89589-c28gz\" (UID: \"190e6358-2083-4549-b5ac-5c35af169b06\") " pod="openstack/dnsmasq-dns-7877d89589-c28gz" Nov 21 14:35:33 crc kubenswrapper[4897]: I1121 14:35:33.679866 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/190e6358-2083-4549-b5ac-5c35af169b06-ovsdbserver-sb\") pod \"dnsmasq-dns-7877d89589-c28gz\" (UID: \"190e6358-2083-4549-b5ac-5c35af169b06\") " pod="openstack/dnsmasq-dns-7877d89589-c28gz" Nov 21 14:35:33 crc kubenswrapper[4897]: I1121 14:35:33.679881 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/190e6358-2083-4549-b5ac-5c35af169b06-config\") pod \"dnsmasq-dns-7877d89589-c28gz\" (UID: \"190e6358-2083-4549-b5ac-5c35af169b06\") " pod="openstack/dnsmasq-dns-7877d89589-c28gz" Nov 21 14:35:33 crc kubenswrapper[4897]: I1121 14:35:33.679909 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/190e6358-2083-4549-b5ac-5c35af169b06-dns-swift-storage-0\") pod \"dnsmasq-dns-7877d89589-c28gz\" (UID: \"190e6358-2083-4549-b5ac-5c35af169b06\") " pod="openstack/dnsmasq-dns-7877d89589-c28gz" Nov 21 14:35:33 crc kubenswrapper[4897]: I1121 14:35:33.680969 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/190e6358-2083-4549-b5ac-5c35af169b06-dns-swift-storage-0\") pod \"dnsmasq-dns-7877d89589-c28gz\" (UID: \"190e6358-2083-4549-b5ac-5c35af169b06\") " pod="openstack/dnsmasq-dns-7877d89589-c28gz" Nov 21 14:35:33 crc kubenswrapper[4897]: I1121 14:35:33.681495 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/190e6358-2083-4549-b5ac-5c35af169b06-ovsdbserver-nb\") pod \"dnsmasq-dns-7877d89589-c28gz\" (UID: \"190e6358-2083-4549-b5ac-5c35af169b06\") " pod="openstack/dnsmasq-dns-7877d89589-c28gz" Nov 21 14:35:33 crc kubenswrapper[4897]: I1121 14:35:33.682317 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/190e6358-2083-4549-b5ac-5c35af169b06-ovsdbserver-sb\") pod \"dnsmasq-dns-7877d89589-c28gz\" (UID: \"190e6358-2083-4549-b5ac-5c35af169b06\") " pod="openstack/dnsmasq-dns-7877d89589-c28gz" Nov 21 14:35:33 crc kubenswrapper[4897]: I1121 14:35:33.682523 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/190e6358-2083-4549-b5ac-5c35af169b06-dns-svc\") pod \"dnsmasq-dns-7877d89589-c28gz\" (UID: \"190e6358-2083-4549-b5ac-5c35af169b06\") " pod="openstack/dnsmasq-dns-7877d89589-c28gz" Nov 21 14:35:33 crc kubenswrapper[4897]: I1121 14:35:33.682913 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/190e6358-2083-4549-b5ac-5c35af169b06-config\") pod \"dnsmasq-dns-7877d89589-c28gz\" (UID: \"190e6358-2083-4549-b5ac-5c35af169b06\") " pod="openstack/dnsmasq-dns-7877d89589-c28gz" Nov 21 14:35:33 crc kubenswrapper[4897]: I1121 14:35:33.706613 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4qcr2\" (UniqueName: \"kubernetes.io/projected/190e6358-2083-4549-b5ac-5c35af169b06-kube-api-access-4qcr2\") pod \"dnsmasq-dns-7877d89589-c28gz\" (UID: \"190e6358-2083-4549-b5ac-5c35af169b06\") " pod="openstack/dnsmasq-dns-7877d89589-c28gz" Nov 21 14:35:33 crc kubenswrapper[4897]: I1121 14:35:33.721011 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-x8ttp"] Nov 21 14:35:34 crc kubenswrapper[4897]: I1121 14:35:34.000784 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7877d89589-c28gz" Nov 21 14:35:34 crc kubenswrapper[4897]: I1121 14:35:34.001712 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-x8ttp" event={"ID":"518b64a4-d073-462f-ada5-20209e14ff27","Type":"ContainerStarted","Data":"5556ea9e99b6adbd506e3b0d3fa6adbd52b662559d4744d18c013279aa03de59"} Nov 21 14:35:34 crc kubenswrapper[4897]: I1121 14:35:34.107991 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 21 14:35:34 crc kubenswrapper[4897]: W1121 14:35:34.114937 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcf6d459c_961c_4361_a034_4a7c7d5ce7b8.slice/crio-6bbbf0358422b5b0b39552525c776d4c383d4edadd6c977cdb1d9f47a5fdfc55 WatchSource:0}: Error finding container 6bbbf0358422b5b0b39552525c776d4c383d4edadd6c977cdb1d9f47a5fdfc55: Status 404 returned error can't find the container with id 6bbbf0358422b5b0b39552525c776d4c383d4edadd6c977cdb1d9f47a5fdfc55 Nov 21 14:35:34 crc kubenswrapper[4897]: I1121 14:35:34.357374 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 21 14:35:34 crc kubenswrapper[4897]: I1121 14:35:34.434803 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-bpgdn"] Nov 21 14:35:34 crc kubenswrapper[4897]: I1121 14:35:34.436370 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-bpgdn" Nov 21 14:35:34 crc kubenswrapper[4897]: I1121 14:35:34.439024 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Nov 21 14:35:34 crc kubenswrapper[4897]: I1121 14:35:34.439677 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Nov 21 14:35:34 crc kubenswrapper[4897]: I1121 14:35:34.446981 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-bpgdn"] Nov 21 14:35:34 crc kubenswrapper[4897]: I1121 14:35:34.511197 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a558cde2-95bc-4f8d-9daf-90fe592ab9ff-config-data\") pod \"nova-cell1-conductor-db-sync-bpgdn\" (UID: \"a558cde2-95bc-4f8d-9daf-90fe592ab9ff\") " pod="openstack/nova-cell1-conductor-db-sync-bpgdn" Nov 21 14:35:34 crc kubenswrapper[4897]: I1121 14:35:34.511361 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a558cde2-95bc-4f8d-9daf-90fe592ab9ff-scripts\") pod \"nova-cell1-conductor-db-sync-bpgdn\" (UID: \"a558cde2-95bc-4f8d-9daf-90fe592ab9ff\") " pod="openstack/nova-cell1-conductor-db-sync-bpgdn" Nov 21 14:35:34 crc kubenswrapper[4897]: I1121 14:35:34.511387 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-httgf\" (UniqueName: \"kubernetes.io/projected/a558cde2-95bc-4f8d-9daf-90fe592ab9ff-kube-api-access-httgf\") pod \"nova-cell1-conductor-db-sync-bpgdn\" (UID: \"a558cde2-95bc-4f8d-9daf-90fe592ab9ff\") " pod="openstack/nova-cell1-conductor-db-sync-bpgdn" Nov 21 14:35:34 crc kubenswrapper[4897]: I1121 14:35:34.511428 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a558cde2-95bc-4f8d-9daf-90fe592ab9ff-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-bpgdn\" (UID: \"a558cde2-95bc-4f8d-9daf-90fe592ab9ff\") " pod="openstack/nova-cell1-conductor-db-sync-bpgdn" Nov 21 14:35:34 crc kubenswrapper[4897]: I1121 14:35:34.563883 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 21 14:35:34 crc kubenswrapper[4897]: I1121 14:35:34.614543 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a558cde2-95bc-4f8d-9daf-90fe592ab9ff-config-data\") pod \"nova-cell1-conductor-db-sync-bpgdn\" (UID: \"a558cde2-95bc-4f8d-9daf-90fe592ab9ff\") " pod="openstack/nova-cell1-conductor-db-sync-bpgdn" Nov 21 14:35:34 crc kubenswrapper[4897]: I1121 14:35:34.614693 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a558cde2-95bc-4f8d-9daf-90fe592ab9ff-scripts\") pod \"nova-cell1-conductor-db-sync-bpgdn\" (UID: \"a558cde2-95bc-4f8d-9daf-90fe592ab9ff\") " pod="openstack/nova-cell1-conductor-db-sync-bpgdn" Nov 21 14:35:34 crc kubenswrapper[4897]: I1121 14:35:34.614725 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-httgf\" (UniqueName: \"kubernetes.io/projected/a558cde2-95bc-4f8d-9daf-90fe592ab9ff-kube-api-access-httgf\") pod \"nova-cell1-conductor-db-sync-bpgdn\" (UID: \"a558cde2-95bc-4f8d-9daf-90fe592ab9ff\") " pod="openstack/nova-cell1-conductor-db-sync-bpgdn" Nov 21 14:35:34 crc kubenswrapper[4897]: I1121 14:35:34.614772 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a558cde2-95bc-4f8d-9daf-90fe592ab9ff-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-bpgdn\" (UID: \"a558cde2-95bc-4f8d-9daf-90fe592ab9ff\") " pod="openstack/nova-cell1-conductor-db-sync-bpgdn" Nov 21 14:35:34 crc kubenswrapper[4897]: I1121 14:35:34.623479 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a558cde2-95bc-4f8d-9daf-90fe592ab9ff-config-data\") pod \"nova-cell1-conductor-db-sync-bpgdn\" (UID: \"a558cde2-95bc-4f8d-9daf-90fe592ab9ff\") " pod="openstack/nova-cell1-conductor-db-sync-bpgdn" Nov 21 14:35:34 crc kubenswrapper[4897]: I1121 14:35:34.626227 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a558cde2-95bc-4f8d-9daf-90fe592ab9ff-scripts\") pod \"nova-cell1-conductor-db-sync-bpgdn\" (UID: \"a558cde2-95bc-4f8d-9daf-90fe592ab9ff\") " pod="openstack/nova-cell1-conductor-db-sync-bpgdn" Nov 21 14:35:34 crc kubenswrapper[4897]: I1121 14:35:34.627347 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a558cde2-95bc-4f8d-9daf-90fe592ab9ff-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-bpgdn\" (UID: \"a558cde2-95bc-4f8d-9daf-90fe592ab9ff\") " pod="openstack/nova-cell1-conductor-db-sync-bpgdn" Nov 21 14:35:34 crc kubenswrapper[4897]: I1121 14:35:34.641135 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-httgf\" (UniqueName: \"kubernetes.io/projected/a558cde2-95bc-4f8d-9daf-90fe592ab9ff-kube-api-access-httgf\") pod \"nova-cell1-conductor-db-sync-bpgdn\" (UID: \"a558cde2-95bc-4f8d-9daf-90fe592ab9ff\") " pod="openstack/nova-cell1-conductor-db-sync-bpgdn" Nov 21 14:35:34 crc kubenswrapper[4897]: I1121 14:35:34.756197 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-bpgdn" Nov 21 14:35:34 crc kubenswrapper[4897]: I1121 14:35:34.778817 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 21 14:35:34 crc kubenswrapper[4897]: I1121 14:35:34.844600 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7877d89589-c28gz"] Nov 21 14:35:35 crc kubenswrapper[4897]: I1121 14:35:35.024335 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"4ca85306-d5f3-414f-abd6-6f4623335792","Type":"ContainerStarted","Data":"5a9429626e9d62c72e25ccce2999cba08de0e5257e599632a004c8dacd097ed2"} Nov 21 14:35:35 crc kubenswrapper[4897]: I1121 14:35:35.043854 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"cf6d459c-961c-4361-a034-4a7c7d5ce7b8","Type":"ContainerStarted","Data":"6bbbf0358422b5b0b39552525c776d4c383d4edadd6c977cdb1d9f47a5fdfc55"} Nov 21 14:35:35 crc kubenswrapper[4897]: I1121 14:35:35.054369 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-x8ttp" event={"ID":"518b64a4-d073-462f-ada5-20209e14ff27","Type":"ContainerStarted","Data":"5d3558195d06f9c88281f03204b0c131aad447a4332e1b80f7f41f8bf27b346b"} Nov 21 14:35:35 crc kubenswrapper[4897]: I1121 14:35:35.065704 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7877d89589-c28gz" event={"ID":"190e6358-2083-4549-b5ac-5c35af169b06","Type":"ContainerStarted","Data":"4866d456c7234eff14ff58668ec55c79fbc8f0dc56704b57566506437a25d37d"} Nov 21 14:35:35 crc kubenswrapper[4897]: I1121 14:35:35.077862 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6aa3a86a-b083-40ac-a299-5c53fb5710a7","Type":"ContainerStarted","Data":"6f351cc3054eb0b207dea57e57d3ba65e6a51803baa4b2669c359a679da8cfb0"} Nov 21 14:35:35 crc kubenswrapper[4897]: I1121 14:35:35.087087 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"fe6f78aa-838c-4bf1-8a19-9726831f2941","Type":"ContainerStarted","Data":"a3376a83cdcbd3e57aaa30bc8717803a8b3e1b286724380cc1b4bcf035f668a6"} Nov 21 14:35:35 crc kubenswrapper[4897]: I1121 14:35:35.107538 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-x8ttp" podStartSLOduration=3.107496016 podStartE2EDuration="3.107496016s" podCreationTimestamp="2025-11-21 14:35:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:35:35.078073467 +0000 UTC m=+1612.362666942" watchObservedRunningTime="2025-11-21 14:35:35.107496016 +0000 UTC m=+1612.392089491" Nov 21 14:35:35 crc kubenswrapper[4897]: I1121 14:35:35.444143 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-bpgdn"] Nov 21 14:35:36 crc kubenswrapper[4897]: I1121 14:35:36.113378 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-bpgdn" event={"ID":"a558cde2-95bc-4f8d-9daf-90fe592ab9ff","Type":"ContainerStarted","Data":"b85b61c6134535cb1d4f859af67cc0a923a31c62fbd636fe98341334044c0bd0"} Nov 21 14:35:36 crc kubenswrapper[4897]: I1121 14:35:36.113662 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-bpgdn" event={"ID":"a558cde2-95bc-4f8d-9daf-90fe592ab9ff","Type":"ContainerStarted","Data":"12147380ef8f3214b00a59e3095d82485dd0afadb2a3326a37e477478a43c9c3"} Nov 21 14:35:36 crc kubenswrapper[4897]: I1121 14:35:36.118976 4897 generic.go:334] "Generic (PLEG): container finished" podID="190e6358-2083-4549-b5ac-5c35af169b06" containerID="a4982453c7f834bedde18c2a269072073f1527ae983e08fe842d8d47389611db" exitCode=0 Nov 21 14:35:36 crc kubenswrapper[4897]: I1121 14:35:36.119066 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7877d89589-c28gz" event={"ID":"190e6358-2083-4549-b5ac-5c35af169b06","Type":"ContainerDied","Data":"a4982453c7f834bedde18c2a269072073f1527ae983e08fe842d8d47389611db"} Nov 21 14:35:36 crc kubenswrapper[4897]: I1121 14:35:36.134600 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-bpgdn" podStartSLOduration=2.134576886 podStartE2EDuration="2.134576886s" podCreationTimestamp="2025-11-21 14:35:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:35:36.127925737 +0000 UTC m=+1613.412519212" watchObservedRunningTime="2025-11-21 14:35:36.134576886 +0000 UTC m=+1613.419170361" Nov 21 14:35:36 crc kubenswrapper[4897]: I1121 14:35:36.814708 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 21 14:35:36 crc kubenswrapper[4897]: I1121 14:35:36.847381 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 21 14:35:37 crc kubenswrapper[4897]: I1121 14:35:37.143581 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7877d89589-c28gz" event={"ID":"190e6358-2083-4549-b5ac-5c35af169b06","Type":"ContainerStarted","Data":"d2be4b278d182e43defddab52a77b6052ea8408529ad229d4d7dbe3734a7b4c4"} Nov 21 14:35:37 crc kubenswrapper[4897]: I1121 14:35:37.144541 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7877d89589-c28gz" Nov 21 14:35:37 crc kubenswrapper[4897]: I1121 14:35:37.168346 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7877d89589-c28gz" podStartSLOduration=4.168328224 podStartE2EDuration="4.168328224s" podCreationTimestamp="2025-11-21 14:35:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:35:37.160262588 +0000 UTC m=+1614.444856063" watchObservedRunningTime="2025-11-21 14:35:37.168328224 +0000 UTC m=+1614.452921699" Nov 21 14:35:41 crc kubenswrapper[4897]: I1121 14:35:41.191476 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6aa3a86a-b083-40ac-a299-5c53fb5710a7","Type":"ContainerStarted","Data":"93278ec84b059403d33e3e3f8b35727d4c560933cdde69d14051473a02bcb704"} Nov 21 14:35:41 crc kubenswrapper[4897]: I1121 14:35:41.192199 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6aa3a86a-b083-40ac-a299-5c53fb5710a7","Type":"ContainerStarted","Data":"1ad944d20dab71ec3e8e7edfaf9b7e4b95c63e2454e530875a03d408d0abceac"} Nov 21 14:35:41 crc kubenswrapper[4897]: I1121 14:35:41.191880 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="6aa3a86a-b083-40ac-a299-5c53fb5710a7" containerName="nova-metadata-metadata" containerID="cri-o://93278ec84b059403d33e3e3f8b35727d4c560933cdde69d14051473a02bcb704" gracePeriod=30 Nov 21 14:35:41 crc kubenswrapper[4897]: I1121 14:35:41.191613 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="6aa3a86a-b083-40ac-a299-5c53fb5710a7" containerName="nova-metadata-log" containerID="cri-o://1ad944d20dab71ec3e8e7edfaf9b7e4b95c63e2454e530875a03d408d0abceac" gracePeriod=30 Nov 21 14:35:41 crc kubenswrapper[4897]: I1121 14:35:41.193117 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"fe6f78aa-838c-4bf1-8a19-9726831f2941","Type":"ContainerStarted","Data":"9563192811a74d43bb16a893b77b8cc16dbe59707cfe89bae3411c69591d5bed"} Nov 21 14:35:41 crc kubenswrapper[4897]: I1121 14:35:41.193228 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="fe6f78aa-838c-4bf1-8a19-9726831f2941" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://9563192811a74d43bb16a893b77b8cc16dbe59707cfe89bae3411c69591d5bed" gracePeriod=30 Nov 21 14:35:41 crc kubenswrapper[4897]: I1121 14:35:41.197896 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"4ca85306-d5f3-414f-abd6-6f4623335792","Type":"ContainerStarted","Data":"976eada4d6f64a9db0df4c95e99e61c1f4b0e1a254cae5367f8fca2fc8b41061"} Nov 21 14:35:41 crc kubenswrapper[4897]: I1121 14:35:41.201557 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"cf6d459c-961c-4361-a034-4a7c7d5ce7b8","Type":"ContainerStarted","Data":"34bf622d9e31992c9251f7bd8f165c2e2b6a33e70e9c16b91673f884f3228578"} Nov 21 14:35:41 crc kubenswrapper[4897]: I1121 14:35:41.201608 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"cf6d459c-961c-4361-a034-4a7c7d5ce7b8","Type":"ContainerStarted","Data":"162624ca9d5ca88cb13b9ab46129c2083d1c72de99ff20042857a4b9a95a6e82"} Nov 21 14:35:41 crc kubenswrapper[4897]: I1121 14:35:41.227885 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.114877317 podStartE2EDuration="8.227864041s" podCreationTimestamp="2025-11-21 14:35:33 +0000 UTC" firstStartedPulling="2025-11-21 14:35:34.753824449 +0000 UTC m=+1612.038417924" lastFinishedPulling="2025-11-21 14:35:39.866811133 +0000 UTC m=+1617.151404648" observedRunningTime="2025-11-21 14:35:41.214358297 +0000 UTC m=+1618.498951782" watchObservedRunningTime="2025-11-21 14:35:41.227864041 +0000 UTC m=+1618.512457516" Nov 21 14:35:41 crc kubenswrapper[4897]: I1121 14:35:41.266133 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.973042788 podStartE2EDuration="8.266113021s" podCreationTimestamp="2025-11-21 14:35:33 +0000 UTC" firstStartedPulling="2025-11-21 14:35:34.570110592 +0000 UTC m=+1611.854704057" lastFinishedPulling="2025-11-21 14:35:39.863180775 +0000 UTC m=+1617.147774290" observedRunningTime="2025-11-21 14:35:41.240134092 +0000 UTC m=+1618.524727597" watchObservedRunningTime="2025-11-21 14:35:41.266113021 +0000 UTC m=+1618.550706496" Nov 21 14:35:41 crc kubenswrapper[4897]: I1121 14:35:41.269333 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.799397719 podStartE2EDuration="8.269322727s" podCreationTimestamp="2025-11-21 14:35:33 +0000 UTC" firstStartedPulling="2025-11-21 14:35:34.396854654 +0000 UTC m=+1611.681448129" lastFinishedPulling="2025-11-21 14:35:39.866779662 +0000 UTC m=+1617.151373137" observedRunningTime="2025-11-21 14:35:41.256723918 +0000 UTC m=+1618.541317393" watchObservedRunningTime="2025-11-21 14:35:41.269322727 +0000 UTC m=+1618.553916212" Nov 21 14:35:41 crc kubenswrapper[4897]: I1121 14:35:41.289251 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.548563864 podStartE2EDuration="9.289235314s" podCreationTimestamp="2025-11-21 14:35:32 +0000 UTC" firstStartedPulling="2025-11-21 14:35:34.129427701 +0000 UTC m=+1611.414021176" lastFinishedPulling="2025-11-21 14:35:39.870099151 +0000 UTC m=+1617.154692626" observedRunningTime="2025-11-21 14:35:41.274851346 +0000 UTC m=+1618.559444821" watchObservedRunningTime="2025-11-21 14:35:41.289235314 +0000 UTC m=+1618.573828779" Nov 21 14:35:41 crc kubenswrapper[4897]: I1121 14:35:41.940163 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 21 14:35:42 crc kubenswrapper[4897]: I1121 14:35:42.091418 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dt4d5\" (UniqueName: \"kubernetes.io/projected/6aa3a86a-b083-40ac-a299-5c53fb5710a7-kube-api-access-dt4d5\") pod \"6aa3a86a-b083-40ac-a299-5c53fb5710a7\" (UID: \"6aa3a86a-b083-40ac-a299-5c53fb5710a7\") " Nov 21 14:35:42 crc kubenswrapper[4897]: I1121 14:35:42.091469 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6aa3a86a-b083-40ac-a299-5c53fb5710a7-combined-ca-bundle\") pod \"6aa3a86a-b083-40ac-a299-5c53fb5710a7\" (UID: \"6aa3a86a-b083-40ac-a299-5c53fb5710a7\") " Nov 21 14:35:42 crc kubenswrapper[4897]: I1121 14:35:42.091643 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6aa3a86a-b083-40ac-a299-5c53fb5710a7-logs\") pod \"6aa3a86a-b083-40ac-a299-5c53fb5710a7\" (UID: \"6aa3a86a-b083-40ac-a299-5c53fb5710a7\") " Nov 21 14:35:42 crc kubenswrapper[4897]: I1121 14:35:42.091717 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6aa3a86a-b083-40ac-a299-5c53fb5710a7-config-data\") pod \"6aa3a86a-b083-40ac-a299-5c53fb5710a7\" (UID: \"6aa3a86a-b083-40ac-a299-5c53fb5710a7\") " Nov 21 14:35:42 crc kubenswrapper[4897]: I1121 14:35:42.092128 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6aa3a86a-b083-40ac-a299-5c53fb5710a7-logs" (OuterVolumeSpecName: "logs") pod "6aa3a86a-b083-40ac-a299-5c53fb5710a7" (UID: "6aa3a86a-b083-40ac-a299-5c53fb5710a7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:35:42 crc kubenswrapper[4897]: I1121 14:35:42.092372 4897 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6aa3a86a-b083-40ac-a299-5c53fb5710a7-logs\") on node \"crc\" DevicePath \"\"" Nov 21 14:35:42 crc kubenswrapper[4897]: I1121 14:35:42.104351 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6aa3a86a-b083-40ac-a299-5c53fb5710a7-kube-api-access-dt4d5" (OuterVolumeSpecName: "kube-api-access-dt4d5") pod "6aa3a86a-b083-40ac-a299-5c53fb5710a7" (UID: "6aa3a86a-b083-40ac-a299-5c53fb5710a7"). InnerVolumeSpecName "kube-api-access-dt4d5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:35:42 crc kubenswrapper[4897]: I1121 14:35:42.134529 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6aa3a86a-b083-40ac-a299-5c53fb5710a7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6aa3a86a-b083-40ac-a299-5c53fb5710a7" (UID: "6aa3a86a-b083-40ac-a299-5c53fb5710a7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:35:42 crc kubenswrapper[4897]: I1121 14:35:42.135427 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6aa3a86a-b083-40ac-a299-5c53fb5710a7-config-data" (OuterVolumeSpecName: "config-data") pod "6aa3a86a-b083-40ac-a299-5c53fb5710a7" (UID: "6aa3a86a-b083-40ac-a299-5c53fb5710a7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:35:42 crc kubenswrapper[4897]: I1121 14:35:42.194745 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dt4d5\" (UniqueName: \"kubernetes.io/projected/6aa3a86a-b083-40ac-a299-5c53fb5710a7-kube-api-access-dt4d5\") on node \"crc\" DevicePath \"\"" Nov 21 14:35:42 crc kubenswrapper[4897]: I1121 14:35:42.194781 4897 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6aa3a86a-b083-40ac-a299-5c53fb5710a7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 21 14:35:42 crc kubenswrapper[4897]: I1121 14:35:42.194790 4897 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6aa3a86a-b083-40ac-a299-5c53fb5710a7-config-data\") on node \"crc\" DevicePath \"\"" Nov 21 14:35:42 crc kubenswrapper[4897]: I1121 14:35:42.220303 4897 generic.go:334] "Generic (PLEG): container finished" podID="6aa3a86a-b083-40ac-a299-5c53fb5710a7" containerID="93278ec84b059403d33e3e3f8b35727d4c560933cdde69d14051473a02bcb704" exitCode=0 Nov 21 14:35:42 crc kubenswrapper[4897]: I1121 14:35:42.220590 4897 generic.go:334] "Generic (PLEG): container finished" podID="6aa3a86a-b083-40ac-a299-5c53fb5710a7" containerID="1ad944d20dab71ec3e8e7edfaf9b7e4b95c63e2454e530875a03d408d0abceac" exitCode=143 Nov 21 14:35:42 crc kubenswrapper[4897]: I1121 14:35:42.220476 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 21 14:35:42 crc kubenswrapper[4897]: I1121 14:35:42.276112 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6aa3a86a-b083-40ac-a299-5c53fb5710a7","Type":"ContainerDied","Data":"93278ec84b059403d33e3e3f8b35727d4c560933cdde69d14051473a02bcb704"} Nov 21 14:35:42 crc kubenswrapper[4897]: I1121 14:35:42.276165 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6aa3a86a-b083-40ac-a299-5c53fb5710a7","Type":"ContainerDied","Data":"1ad944d20dab71ec3e8e7edfaf9b7e4b95c63e2454e530875a03d408d0abceac"} Nov 21 14:35:42 crc kubenswrapper[4897]: I1121 14:35:42.276176 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6aa3a86a-b083-40ac-a299-5c53fb5710a7","Type":"ContainerDied","Data":"6f351cc3054eb0b207dea57e57d3ba65e6a51803baa4b2669c359a679da8cfb0"} Nov 21 14:35:42 crc kubenswrapper[4897]: I1121 14:35:42.276194 4897 scope.go:117] "RemoveContainer" containerID="93278ec84b059403d33e3e3f8b35727d4c560933cdde69d14051473a02bcb704" Nov 21 14:35:42 crc kubenswrapper[4897]: I1121 14:35:42.329763 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 21 14:35:42 crc kubenswrapper[4897]: I1121 14:35:42.341551 4897 scope.go:117] "RemoveContainer" containerID="1ad944d20dab71ec3e8e7edfaf9b7e4b95c63e2454e530875a03d408d0abceac" Nov 21 14:35:42 crc kubenswrapper[4897]: I1121 14:35:42.354842 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 21 14:35:42 crc kubenswrapper[4897]: I1121 14:35:42.375291 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 21 14:35:42 crc kubenswrapper[4897]: E1121 14:35:42.376052 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6aa3a86a-b083-40ac-a299-5c53fb5710a7" containerName="nova-metadata-metadata" Nov 21 14:35:42 crc kubenswrapper[4897]: I1121 14:35:42.376199 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="6aa3a86a-b083-40ac-a299-5c53fb5710a7" containerName="nova-metadata-metadata" Nov 21 14:35:42 crc kubenswrapper[4897]: E1121 14:35:42.376248 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6aa3a86a-b083-40ac-a299-5c53fb5710a7" containerName="nova-metadata-log" Nov 21 14:35:42 crc kubenswrapper[4897]: I1121 14:35:42.376259 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="6aa3a86a-b083-40ac-a299-5c53fb5710a7" containerName="nova-metadata-log" Nov 21 14:35:42 crc kubenswrapper[4897]: I1121 14:35:42.376687 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="6aa3a86a-b083-40ac-a299-5c53fb5710a7" containerName="nova-metadata-metadata" Nov 21 14:35:42 crc kubenswrapper[4897]: I1121 14:35:42.376759 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="6aa3a86a-b083-40ac-a299-5c53fb5710a7" containerName="nova-metadata-log" Nov 21 14:35:42 crc kubenswrapper[4897]: I1121 14:35:42.378699 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 21 14:35:42 crc kubenswrapper[4897]: I1121 14:35:42.382937 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 21 14:35:42 crc kubenswrapper[4897]: I1121 14:35:42.383096 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 21 14:35:42 crc kubenswrapper[4897]: I1121 14:35:42.408622 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 21 14:35:42 crc kubenswrapper[4897]: I1121 14:35:42.411261 4897 scope.go:117] "RemoveContainer" containerID="93278ec84b059403d33e3e3f8b35727d4c560933cdde69d14051473a02bcb704" Nov 21 14:35:42 crc kubenswrapper[4897]: E1121 14:35:42.411910 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"93278ec84b059403d33e3e3f8b35727d4c560933cdde69d14051473a02bcb704\": container with ID starting with 93278ec84b059403d33e3e3f8b35727d4c560933cdde69d14051473a02bcb704 not found: ID does not exist" containerID="93278ec84b059403d33e3e3f8b35727d4c560933cdde69d14051473a02bcb704" Nov 21 14:35:42 crc kubenswrapper[4897]: I1121 14:35:42.411949 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93278ec84b059403d33e3e3f8b35727d4c560933cdde69d14051473a02bcb704"} err="failed to get container status \"93278ec84b059403d33e3e3f8b35727d4c560933cdde69d14051473a02bcb704\": rpc error: code = NotFound desc = could not find container \"93278ec84b059403d33e3e3f8b35727d4c560933cdde69d14051473a02bcb704\": container with ID starting with 93278ec84b059403d33e3e3f8b35727d4c560933cdde69d14051473a02bcb704 not found: ID does not exist" Nov 21 14:35:42 crc kubenswrapper[4897]: I1121 14:35:42.411978 4897 scope.go:117] "RemoveContainer" containerID="1ad944d20dab71ec3e8e7edfaf9b7e4b95c63e2454e530875a03d408d0abceac" Nov 21 14:35:42 crc kubenswrapper[4897]: E1121 14:35:42.412483 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ad944d20dab71ec3e8e7edfaf9b7e4b95c63e2454e530875a03d408d0abceac\": container with ID starting with 1ad944d20dab71ec3e8e7edfaf9b7e4b95c63e2454e530875a03d408d0abceac not found: ID does not exist" containerID="1ad944d20dab71ec3e8e7edfaf9b7e4b95c63e2454e530875a03d408d0abceac" Nov 21 14:35:42 crc kubenswrapper[4897]: I1121 14:35:42.412546 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ad944d20dab71ec3e8e7edfaf9b7e4b95c63e2454e530875a03d408d0abceac"} err="failed to get container status \"1ad944d20dab71ec3e8e7edfaf9b7e4b95c63e2454e530875a03d408d0abceac\": rpc error: code = NotFound desc = could not find container \"1ad944d20dab71ec3e8e7edfaf9b7e4b95c63e2454e530875a03d408d0abceac\": container with ID starting with 1ad944d20dab71ec3e8e7edfaf9b7e4b95c63e2454e530875a03d408d0abceac not found: ID does not exist" Nov 21 14:35:42 crc kubenswrapper[4897]: I1121 14:35:42.412575 4897 scope.go:117] "RemoveContainer" containerID="93278ec84b059403d33e3e3f8b35727d4c560933cdde69d14051473a02bcb704" Nov 21 14:35:42 crc kubenswrapper[4897]: I1121 14:35:42.412916 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93278ec84b059403d33e3e3f8b35727d4c560933cdde69d14051473a02bcb704"} err="failed to get container status \"93278ec84b059403d33e3e3f8b35727d4c560933cdde69d14051473a02bcb704\": rpc error: code = NotFound desc = could not find container \"93278ec84b059403d33e3e3f8b35727d4c560933cdde69d14051473a02bcb704\": container with ID starting with 93278ec84b059403d33e3e3f8b35727d4c560933cdde69d14051473a02bcb704 not found: ID does not exist" Nov 21 14:35:42 crc kubenswrapper[4897]: I1121 14:35:42.412940 4897 scope.go:117] "RemoveContainer" containerID="1ad944d20dab71ec3e8e7edfaf9b7e4b95c63e2454e530875a03d408d0abceac" Nov 21 14:35:42 crc kubenswrapper[4897]: I1121 14:35:42.413223 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ad944d20dab71ec3e8e7edfaf9b7e4b95c63e2454e530875a03d408d0abceac"} err="failed to get container status \"1ad944d20dab71ec3e8e7edfaf9b7e4b95c63e2454e530875a03d408d0abceac\": rpc error: code = NotFound desc = could not find container \"1ad944d20dab71ec3e8e7edfaf9b7e4b95c63e2454e530875a03d408d0abceac\": container with ID starting with 1ad944d20dab71ec3e8e7edfaf9b7e4b95c63e2454e530875a03d408d0abceac not found: ID does not exist" Nov 21 14:35:42 crc kubenswrapper[4897]: I1121 14:35:42.502943 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/728674e7-ee48-41d2-ad13-7d6f623d99b9-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"728674e7-ee48-41d2-ad13-7d6f623d99b9\") " pod="openstack/nova-metadata-0" Nov 21 14:35:42 crc kubenswrapper[4897]: I1121 14:35:42.503013 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cgjn9\" (UniqueName: \"kubernetes.io/projected/728674e7-ee48-41d2-ad13-7d6f623d99b9-kube-api-access-cgjn9\") pod \"nova-metadata-0\" (UID: \"728674e7-ee48-41d2-ad13-7d6f623d99b9\") " pod="openstack/nova-metadata-0" Nov 21 14:35:42 crc kubenswrapper[4897]: I1121 14:35:42.503272 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/728674e7-ee48-41d2-ad13-7d6f623d99b9-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"728674e7-ee48-41d2-ad13-7d6f623d99b9\") " pod="openstack/nova-metadata-0" Nov 21 14:35:42 crc kubenswrapper[4897]: I1121 14:35:42.503480 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/728674e7-ee48-41d2-ad13-7d6f623d99b9-config-data\") pod \"nova-metadata-0\" (UID: \"728674e7-ee48-41d2-ad13-7d6f623d99b9\") " pod="openstack/nova-metadata-0" Nov 21 14:35:42 crc kubenswrapper[4897]: I1121 14:35:42.503673 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/728674e7-ee48-41d2-ad13-7d6f623d99b9-logs\") pod \"nova-metadata-0\" (UID: \"728674e7-ee48-41d2-ad13-7d6f623d99b9\") " pod="openstack/nova-metadata-0" Nov 21 14:35:42 crc kubenswrapper[4897]: I1121 14:35:42.606226 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/728674e7-ee48-41d2-ad13-7d6f623d99b9-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"728674e7-ee48-41d2-ad13-7d6f623d99b9\") " pod="openstack/nova-metadata-0" Nov 21 14:35:42 crc kubenswrapper[4897]: I1121 14:35:42.606338 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/728674e7-ee48-41d2-ad13-7d6f623d99b9-config-data\") pod \"nova-metadata-0\" (UID: \"728674e7-ee48-41d2-ad13-7d6f623d99b9\") " pod="openstack/nova-metadata-0" Nov 21 14:35:42 crc kubenswrapper[4897]: I1121 14:35:42.606405 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/728674e7-ee48-41d2-ad13-7d6f623d99b9-logs\") pod \"nova-metadata-0\" (UID: \"728674e7-ee48-41d2-ad13-7d6f623d99b9\") " pod="openstack/nova-metadata-0" Nov 21 14:35:42 crc kubenswrapper[4897]: I1121 14:35:42.606451 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/728674e7-ee48-41d2-ad13-7d6f623d99b9-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"728674e7-ee48-41d2-ad13-7d6f623d99b9\") " pod="openstack/nova-metadata-0" Nov 21 14:35:42 crc kubenswrapper[4897]: I1121 14:35:42.606541 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cgjn9\" (UniqueName: \"kubernetes.io/projected/728674e7-ee48-41d2-ad13-7d6f623d99b9-kube-api-access-cgjn9\") pod \"nova-metadata-0\" (UID: \"728674e7-ee48-41d2-ad13-7d6f623d99b9\") " pod="openstack/nova-metadata-0" Nov 21 14:35:42 crc kubenswrapper[4897]: I1121 14:35:42.607016 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/728674e7-ee48-41d2-ad13-7d6f623d99b9-logs\") pod \"nova-metadata-0\" (UID: \"728674e7-ee48-41d2-ad13-7d6f623d99b9\") " pod="openstack/nova-metadata-0" Nov 21 14:35:42 crc kubenswrapper[4897]: I1121 14:35:42.610809 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/728674e7-ee48-41d2-ad13-7d6f623d99b9-config-data\") pod \"nova-metadata-0\" (UID: \"728674e7-ee48-41d2-ad13-7d6f623d99b9\") " pod="openstack/nova-metadata-0" Nov 21 14:35:42 crc kubenswrapper[4897]: I1121 14:35:42.619080 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/728674e7-ee48-41d2-ad13-7d6f623d99b9-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"728674e7-ee48-41d2-ad13-7d6f623d99b9\") " pod="openstack/nova-metadata-0" Nov 21 14:35:42 crc kubenswrapper[4897]: I1121 14:35:42.619222 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/728674e7-ee48-41d2-ad13-7d6f623d99b9-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"728674e7-ee48-41d2-ad13-7d6f623d99b9\") " pod="openstack/nova-metadata-0" Nov 21 14:35:42 crc kubenswrapper[4897]: I1121 14:35:42.630445 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cgjn9\" (UniqueName: \"kubernetes.io/projected/728674e7-ee48-41d2-ad13-7d6f623d99b9-kube-api-access-cgjn9\") pod \"nova-metadata-0\" (UID: \"728674e7-ee48-41d2-ad13-7d6f623d99b9\") " pod="openstack/nova-metadata-0" Nov 21 14:35:42 crc kubenswrapper[4897]: I1121 14:35:42.700226 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 21 14:35:43 crc kubenswrapper[4897]: I1121 14:35:43.211237 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 21 14:35:43 crc kubenswrapper[4897]: I1121 14:35:43.233565 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 21 14:35:43 crc kubenswrapper[4897]: I1121 14:35:43.233737 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 21 14:35:43 crc kubenswrapper[4897]: I1121 14:35:43.234305 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"728674e7-ee48-41d2-ad13-7d6f623d99b9","Type":"ContainerStarted","Data":"5789cf6915ae3a5dfd10140e6500dad99899ebd95b82932f77b3c1ee5bc75688"} Nov 21 14:35:43 crc kubenswrapper[4897]: I1121 14:35:43.535770 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 21 14:35:43 crc kubenswrapper[4897]: I1121 14:35:43.536397 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 21 14:35:43 crc kubenswrapper[4897]: I1121 14:35:43.572191 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 21 14:35:43 crc kubenswrapper[4897]: I1121 14:35:43.624623 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Nov 21 14:35:44 crc kubenswrapper[4897]: I1121 14:35:44.003293 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7877d89589-c28gz" Nov 21 14:35:44 crc kubenswrapper[4897]: I1121 14:35:44.076064 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7d978555f9-wjc8q"] Nov 21 14:35:44 crc kubenswrapper[4897]: I1121 14:35:44.076355 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7d978555f9-wjc8q" podUID="afa69523-63c4-4a94-b94a-61c1343ef10c" containerName="dnsmasq-dns" containerID="cri-o://6cb3731d3191a0096a5051532c43fd96973f7663ef9d3571618a165b3dd1afca" gracePeriod=10 Nov 21 14:35:44 crc kubenswrapper[4897]: I1121 14:35:44.135478 4897 scope.go:117] "RemoveContainer" containerID="c68f1b003d12b4a2964169a85a04e3f0272d86c891f52e6bebd52f3781520d1a" Nov 21 14:35:44 crc kubenswrapper[4897]: E1121 14:35:44.135867 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 14:35:44 crc kubenswrapper[4897]: I1121 14:35:44.151566 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6aa3a86a-b083-40ac-a299-5c53fb5710a7" path="/var/lib/kubelet/pods/6aa3a86a-b083-40ac-a299-5c53fb5710a7/volumes" Nov 21 14:35:44 crc kubenswrapper[4897]: I1121 14:35:44.304706 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"728674e7-ee48-41d2-ad13-7d6f623d99b9","Type":"ContainerStarted","Data":"e5f61c9fbb7f4949782c1efe883d18bd77495f96d811b26f46c530d8c85e12d8"} Nov 21 14:35:44 crc kubenswrapper[4897]: I1121 14:35:44.305140 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"728674e7-ee48-41d2-ad13-7d6f623d99b9","Type":"ContainerStarted","Data":"f2d77335ade2070c2371391c35bd65b08dfc83a9254e99cd03788d9b3e6f05f1"} Nov 21 14:35:44 crc kubenswrapper[4897]: I1121 14:35:44.311680 4897 generic.go:334] "Generic (PLEG): container finished" podID="518b64a4-d073-462f-ada5-20209e14ff27" containerID="5d3558195d06f9c88281f03204b0c131aad447a4332e1b80f7f41f8bf27b346b" exitCode=0 Nov 21 14:35:44 crc kubenswrapper[4897]: I1121 14:35:44.311754 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-x8ttp" event={"ID":"518b64a4-d073-462f-ada5-20209e14ff27","Type":"ContainerDied","Data":"5d3558195d06f9c88281f03204b0c131aad447a4332e1b80f7f41f8bf27b346b"} Nov 21 14:35:44 crc kubenswrapper[4897]: I1121 14:35:44.315123 4897 generic.go:334] "Generic (PLEG): container finished" podID="6d3cad20-fb17-4726-b8fa-9fd5303b5eea" containerID="36d1ba740cdac40a73ac8d07a9d49188d3f951e9e6e9797089e9005483687829" exitCode=0 Nov 21 14:35:44 crc kubenswrapper[4897]: I1121 14:35:44.315200 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dtwql" event={"ID":"6d3cad20-fb17-4726-b8fa-9fd5303b5eea","Type":"ContainerDied","Data":"36d1ba740cdac40a73ac8d07a9d49188d3f951e9e6e9797089e9005483687829"} Nov 21 14:35:44 crc kubenswrapper[4897]: I1121 14:35:44.319653 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="cf6d459c-961c-4361-a034-4a7c7d5ce7b8" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.240:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 21 14:35:44 crc kubenswrapper[4897]: I1121 14:35:44.319726 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="cf6d459c-961c-4361-a034-4a7c7d5ce7b8" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.240:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 21 14:35:44 crc kubenswrapper[4897]: I1121 14:35:44.326301 4897 generic.go:334] "Generic (PLEG): container finished" podID="afa69523-63c4-4a94-b94a-61c1343ef10c" containerID="6cb3731d3191a0096a5051532c43fd96973f7663ef9d3571618a165b3dd1afca" exitCode=0 Nov 21 14:35:44 crc kubenswrapper[4897]: I1121 14:35:44.329597 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.329578322 podStartE2EDuration="2.329578322s" podCreationTimestamp="2025-11-21 14:35:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:35:44.322544052 +0000 UTC m=+1621.607137617" watchObservedRunningTime="2025-11-21 14:35:44.329578322 +0000 UTC m=+1621.614171797" Nov 21 14:35:44 crc kubenswrapper[4897]: I1121 14:35:44.330810 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d978555f9-wjc8q" event={"ID":"afa69523-63c4-4a94-b94a-61c1343ef10c","Type":"ContainerDied","Data":"6cb3731d3191a0096a5051532c43fd96973f7663ef9d3571618a165b3dd1afca"} Nov 21 14:35:44 crc kubenswrapper[4897]: I1121 14:35:44.367478 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 21 14:35:44 crc kubenswrapper[4897]: I1121 14:35:44.812136 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d978555f9-wjc8q" Nov 21 14:35:44 crc kubenswrapper[4897]: I1121 14:35:44.979195 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/afa69523-63c4-4a94-b94a-61c1343ef10c-dns-svc\") pod \"afa69523-63c4-4a94-b94a-61c1343ef10c\" (UID: \"afa69523-63c4-4a94-b94a-61c1343ef10c\") " Nov 21 14:35:44 crc kubenswrapper[4897]: I1121 14:35:44.979254 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/afa69523-63c4-4a94-b94a-61c1343ef10c-ovsdbserver-sb\") pod \"afa69523-63c4-4a94-b94a-61c1343ef10c\" (UID: \"afa69523-63c4-4a94-b94a-61c1343ef10c\") " Nov 21 14:35:44 crc kubenswrapper[4897]: I1121 14:35:44.979342 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/afa69523-63c4-4a94-b94a-61c1343ef10c-ovsdbserver-nb\") pod \"afa69523-63c4-4a94-b94a-61c1343ef10c\" (UID: \"afa69523-63c4-4a94-b94a-61c1343ef10c\") " Nov 21 14:35:44 crc kubenswrapper[4897]: I1121 14:35:44.979490 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/afa69523-63c4-4a94-b94a-61c1343ef10c-dns-swift-storage-0\") pod \"afa69523-63c4-4a94-b94a-61c1343ef10c\" (UID: \"afa69523-63c4-4a94-b94a-61c1343ef10c\") " Nov 21 14:35:44 crc kubenswrapper[4897]: I1121 14:35:44.979596 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5lx7z\" (UniqueName: \"kubernetes.io/projected/afa69523-63c4-4a94-b94a-61c1343ef10c-kube-api-access-5lx7z\") pod \"afa69523-63c4-4a94-b94a-61c1343ef10c\" (UID: \"afa69523-63c4-4a94-b94a-61c1343ef10c\") " Nov 21 14:35:44 crc kubenswrapper[4897]: I1121 14:35:44.979662 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/afa69523-63c4-4a94-b94a-61c1343ef10c-config\") pod \"afa69523-63c4-4a94-b94a-61c1343ef10c\" (UID: \"afa69523-63c4-4a94-b94a-61c1343ef10c\") " Nov 21 14:35:44 crc kubenswrapper[4897]: I1121 14:35:44.991687 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/afa69523-63c4-4a94-b94a-61c1343ef10c-kube-api-access-5lx7z" (OuterVolumeSpecName: "kube-api-access-5lx7z") pod "afa69523-63c4-4a94-b94a-61c1343ef10c" (UID: "afa69523-63c4-4a94-b94a-61c1343ef10c"). InnerVolumeSpecName "kube-api-access-5lx7z". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:35:45 crc kubenswrapper[4897]: I1121 14:35:45.053206 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/afa69523-63c4-4a94-b94a-61c1343ef10c-config" (OuterVolumeSpecName: "config") pod "afa69523-63c4-4a94-b94a-61c1343ef10c" (UID: "afa69523-63c4-4a94-b94a-61c1343ef10c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:35:45 crc kubenswrapper[4897]: I1121 14:35:45.053363 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/afa69523-63c4-4a94-b94a-61c1343ef10c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "afa69523-63c4-4a94-b94a-61c1343ef10c" (UID: "afa69523-63c4-4a94-b94a-61c1343ef10c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:35:45 crc kubenswrapper[4897]: I1121 14:35:45.065904 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/afa69523-63c4-4a94-b94a-61c1343ef10c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "afa69523-63c4-4a94-b94a-61c1343ef10c" (UID: "afa69523-63c4-4a94-b94a-61c1343ef10c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:35:45 crc kubenswrapper[4897]: I1121 14:35:45.077347 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/afa69523-63c4-4a94-b94a-61c1343ef10c-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "afa69523-63c4-4a94-b94a-61c1343ef10c" (UID: "afa69523-63c4-4a94-b94a-61c1343ef10c"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:35:45 crc kubenswrapper[4897]: I1121 14:35:45.084240 4897 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/afa69523-63c4-4a94-b94a-61c1343ef10c-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 21 14:35:45 crc kubenswrapper[4897]: I1121 14:35:45.084277 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5lx7z\" (UniqueName: \"kubernetes.io/projected/afa69523-63c4-4a94-b94a-61c1343ef10c-kube-api-access-5lx7z\") on node \"crc\" DevicePath \"\"" Nov 21 14:35:45 crc kubenswrapper[4897]: I1121 14:35:45.084291 4897 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/afa69523-63c4-4a94-b94a-61c1343ef10c-config\") on node \"crc\" DevicePath \"\"" Nov 21 14:35:45 crc kubenswrapper[4897]: I1121 14:35:45.084305 4897 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/afa69523-63c4-4a94-b94a-61c1343ef10c-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 21 14:35:45 crc kubenswrapper[4897]: I1121 14:35:45.084316 4897 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/afa69523-63c4-4a94-b94a-61c1343ef10c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 21 14:35:45 crc kubenswrapper[4897]: I1121 14:35:45.097453 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/afa69523-63c4-4a94-b94a-61c1343ef10c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "afa69523-63c4-4a94-b94a-61c1343ef10c" (UID: "afa69523-63c4-4a94-b94a-61c1343ef10c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:35:45 crc kubenswrapper[4897]: I1121 14:35:45.188296 4897 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/afa69523-63c4-4a94-b94a-61c1343ef10c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 21 14:35:45 crc kubenswrapper[4897]: I1121 14:35:45.339303 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dtwql" event={"ID":"6d3cad20-fb17-4726-b8fa-9fd5303b5eea","Type":"ContainerStarted","Data":"5ee18c9557dc1781cb41a1a2df7bdb7b7161886e8a11d31888240d1445a4be95"} Nov 21 14:35:45 crc kubenswrapper[4897]: I1121 14:35:45.342326 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d978555f9-wjc8q" event={"ID":"afa69523-63c4-4a94-b94a-61c1343ef10c","Type":"ContainerDied","Data":"ce254d218e4e45e561a94f1873bfd2ea63536085260967c8f8df6325db3bcfec"} Nov 21 14:35:45 crc kubenswrapper[4897]: I1121 14:35:45.342376 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d978555f9-wjc8q" Nov 21 14:35:45 crc kubenswrapper[4897]: I1121 14:35:45.342389 4897 scope.go:117] "RemoveContainer" containerID="6cb3731d3191a0096a5051532c43fd96973f7663ef9d3571618a165b3dd1afca" Nov 21 14:35:45 crc kubenswrapper[4897]: I1121 14:35:45.410641 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7d978555f9-wjc8q"] Nov 21 14:35:45 crc kubenswrapper[4897]: I1121 14:35:45.422891 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7d978555f9-wjc8q"] Nov 21 14:35:45 crc kubenswrapper[4897]: I1121 14:35:45.491353 4897 scope.go:117] "RemoveContainer" containerID="1597597d5344b5251bb3061739b1b61fdd9a42c65f23f537f22dac42ac7b2194" Nov 21 14:35:45 crc kubenswrapper[4897]: I1121 14:35:45.824313 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-x8ttp" Nov 21 14:35:46 crc kubenswrapper[4897]: I1121 14:35:46.005085 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qx8n8\" (UniqueName: \"kubernetes.io/projected/518b64a4-d073-462f-ada5-20209e14ff27-kube-api-access-qx8n8\") pod \"518b64a4-d073-462f-ada5-20209e14ff27\" (UID: \"518b64a4-d073-462f-ada5-20209e14ff27\") " Nov 21 14:35:46 crc kubenswrapper[4897]: I1121 14:35:46.005245 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/518b64a4-d073-462f-ada5-20209e14ff27-combined-ca-bundle\") pod \"518b64a4-d073-462f-ada5-20209e14ff27\" (UID: \"518b64a4-d073-462f-ada5-20209e14ff27\") " Nov 21 14:35:46 crc kubenswrapper[4897]: I1121 14:35:46.005305 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/518b64a4-d073-462f-ada5-20209e14ff27-scripts\") pod \"518b64a4-d073-462f-ada5-20209e14ff27\" (UID: \"518b64a4-d073-462f-ada5-20209e14ff27\") " Nov 21 14:35:46 crc kubenswrapper[4897]: I1121 14:35:46.005350 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/518b64a4-d073-462f-ada5-20209e14ff27-config-data\") pod \"518b64a4-d073-462f-ada5-20209e14ff27\" (UID: \"518b64a4-d073-462f-ada5-20209e14ff27\") " Nov 21 14:35:46 crc kubenswrapper[4897]: I1121 14:35:46.011763 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/518b64a4-d073-462f-ada5-20209e14ff27-kube-api-access-qx8n8" (OuterVolumeSpecName: "kube-api-access-qx8n8") pod "518b64a4-d073-462f-ada5-20209e14ff27" (UID: "518b64a4-d073-462f-ada5-20209e14ff27"). InnerVolumeSpecName "kube-api-access-qx8n8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:35:46 crc kubenswrapper[4897]: I1121 14:35:46.012288 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/518b64a4-d073-462f-ada5-20209e14ff27-scripts" (OuterVolumeSpecName: "scripts") pod "518b64a4-d073-462f-ada5-20209e14ff27" (UID: "518b64a4-d073-462f-ada5-20209e14ff27"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:35:46 crc kubenswrapper[4897]: I1121 14:35:46.038024 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/518b64a4-d073-462f-ada5-20209e14ff27-config-data" (OuterVolumeSpecName: "config-data") pod "518b64a4-d073-462f-ada5-20209e14ff27" (UID: "518b64a4-d073-462f-ada5-20209e14ff27"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:35:46 crc kubenswrapper[4897]: I1121 14:35:46.038934 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/518b64a4-d073-462f-ada5-20209e14ff27-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "518b64a4-d073-462f-ada5-20209e14ff27" (UID: "518b64a4-d073-462f-ada5-20209e14ff27"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:35:46 crc kubenswrapper[4897]: I1121 14:35:46.109164 4897 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/518b64a4-d073-462f-ada5-20209e14ff27-scripts\") on node \"crc\" DevicePath \"\"" Nov 21 14:35:46 crc kubenswrapper[4897]: I1121 14:35:46.109218 4897 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/518b64a4-d073-462f-ada5-20209e14ff27-config-data\") on node \"crc\" DevicePath \"\"" Nov 21 14:35:46 crc kubenswrapper[4897]: I1121 14:35:46.109234 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qx8n8\" (UniqueName: \"kubernetes.io/projected/518b64a4-d073-462f-ada5-20209e14ff27-kube-api-access-qx8n8\") on node \"crc\" DevicePath \"\"" Nov 21 14:35:46 crc kubenswrapper[4897]: I1121 14:35:46.109246 4897 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/518b64a4-d073-462f-ada5-20209e14ff27-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 21 14:35:46 crc kubenswrapper[4897]: I1121 14:35:46.110631 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="afa69523-63c4-4a94-b94a-61c1343ef10c" path="/var/lib/kubelet/pods/afa69523-63c4-4a94-b94a-61c1343ef10c/volumes" Nov 21 14:35:46 crc kubenswrapper[4897]: I1121 14:35:46.360485 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-x8ttp" event={"ID":"518b64a4-d073-462f-ada5-20209e14ff27","Type":"ContainerDied","Data":"5556ea9e99b6adbd506e3b0d3fa6adbd52b662559d4744d18c013279aa03de59"} Nov 21 14:35:46 crc kubenswrapper[4897]: I1121 14:35:46.360540 4897 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5556ea9e99b6adbd506e3b0d3fa6adbd52b662559d4744d18c013279aa03de59" Nov 21 14:35:46 crc kubenswrapper[4897]: I1121 14:35:46.360550 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-x8ttp" Nov 21 14:35:46 crc kubenswrapper[4897]: I1121 14:35:46.524498 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 21 14:35:46 crc kubenswrapper[4897]: I1121 14:35:46.524759 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="cf6d459c-961c-4361-a034-4a7c7d5ce7b8" containerName="nova-api-log" containerID="cri-o://162624ca9d5ca88cb13b9ab46129c2083d1c72de99ff20042857a4b9a95a6e82" gracePeriod=30 Nov 21 14:35:46 crc kubenswrapper[4897]: I1121 14:35:46.524818 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="cf6d459c-961c-4361-a034-4a7c7d5ce7b8" containerName="nova-api-api" containerID="cri-o://34bf622d9e31992c9251f7bd8f165c2e2b6a33e70e9c16b91673f884f3228578" gracePeriod=30 Nov 21 14:35:46 crc kubenswrapper[4897]: I1121 14:35:46.537948 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 21 14:35:46 crc kubenswrapper[4897]: I1121 14:35:46.538149 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="4ca85306-d5f3-414f-abd6-6f4623335792" containerName="nova-scheduler-scheduler" containerID="cri-o://976eada4d6f64a9db0df4c95e99e61c1f4b0e1a254cae5367f8fca2fc8b41061" gracePeriod=30 Nov 21 14:35:46 crc kubenswrapper[4897]: I1121 14:35:46.560331 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 21 14:35:46 crc kubenswrapper[4897]: I1121 14:35:46.561226 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="728674e7-ee48-41d2-ad13-7d6f623d99b9" containerName="nova-metadata-metadata" containerID="cri-o://e5f61c9fbb7f4949782c1efe883d18bd77495f96d811b26f46c530d8c85e12d8" gracePeriod=30 Nov 21 14:35:46 crc kubenswrapper[4897]: I1121 14:35:46.561167 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="728674e7-ee48-41d2-ad13-7d6f623d99b9" containerName="nova-metadata-log" containerID="cri-o://f2d77335ade2070c2371391c35bd65b08dfc83a9254e99cd03788d9b3e6f05f1" gracePeriod=30 Nov 21 14:35:47 crc kubenswrapper[4897]: I1121 14:35:47.253733 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 21 14:35:47 crc kubenswrapper[4897]: I1121 14:35:47.375784 4897 generic.go:334] "Generic (PLEG): container finished" podID="728674e7-ee48-41d2-ad13-7d6f623d99b9" containerID="e5f61c9fbb7f4949782c1efe883d18bd77495f96d811b26f46c530d8c85e12d8" exitCode=0 Nov 21 14:35:47 crc kubenswrapper[4897]: I1121 14:35:47.376151 4897 generic.go:334] "Generic (PLEG): container finished" podID="728674e7-ee48-41d2-ad13-7d6f623d99b9" containerID="f2d77335ade2070c2371391c35bd65b08dfc83a9254e99cd03788d9b3e6f05f1" exitCode=143 Nov 21 14:35:47 crc kubenswrapper[4897]: I1121 14:35:47.376223 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"728674e7-ee48-41d2-ad13-7d6f623d99b9","Type":"ContainerDied","Data":"e5f61c9fbb7f4949782c1efe883d18bd77495f96d811b26f46c530d8c85e12d8"} Nov 21 14:35:47 crc kubenswrapper[4897]: I1121 14:35:47.376262 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"728674e7-ee48-41d2-ad13-7d6f623d99b9","Type":"ContainerDied","Data":"f2d77335ade2070c2371391c35bd65b08dfc83a9254e99cd03788d9b3e6f05f1"} Nov 21 14:35:47 crc kubenswrapper[4897]: I1121 14:35:47.376280 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"728674e7-ee48-41d2-ad13-7d6f623d99b9","Type":"ContainerDied","Data":"5789cf6915ae3a5dfd10140e6500dad99899ebd95b82932f77b3c1ee5bc75688"} Nov 21 14:35:47 crc kubenswrapper[4897]: I1121 14:35:47.376302 4897 scope.go:117] "RemoveContainer" containerID="e5f61c9fbb7f4949782c1efe883d18bd77495f96d811b26f46c530d8c85e12d8" Nov 21 14:35:47 crc kubenswrapper[4897]: I1121 14:35:47.376621 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 21 14:35:47 crc kubenswrapper[4897]: I1121 14:35:47.382033 4897 generic.go:334] "Generic (PLEG): container finished" podID="cf6d459c-961c-4361-a034-4a7c7d5ce7b8" containerID="162624ca9d5ca88cb13b9ab46129c2083d1c72de99ff20042857a4b9a95a6e82" exitCode=143 Nov 21 14:35:47 crc kubenswrapper[4897]: I1121 14:35:47.382074 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"cf6d459c-961c-4361-a034-4a7c7d5ce7b8","Type":"ContainerDied","Data":"162624ca9d5ca88cb13b9ab46129c2083d1c72de99ff20042857a4b9a95a6e82"} Nov 21 14:35:47 crc kubenswrapper[4897]: I1121 14:35:47.404695 4897 scope.go:117] "RemoveContainer" containerID="f2d77335ade2070c2371391c35bd65b08dfc83a9254e99cd03788d9b3e6f05f1" Nov 21 14:35:47 crc kubenswrapper[4897]: I1121 14:35:47.427352 4897 scope.go:117] "RemoveContainer" containerID="e5f61c9fbb7f4949782c1efe883d18bd77495f96d811b26f46c530d8c85e12d8" Nov 21 14:35:47 crc kubenswrapper[4897]: E1121 14:35:47.427912 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e5f61c9fbb7f4949782c1efe883d18bd77495f96d811b26f46c530d8c85e12d8\": container with ID starting with e5f61c9fbb7f4949782c1efe883d18bd77495f96d811b26f46c530d8c85e12d8 not found: ID does not exist" containerID="e5f61c9fbb7f4949782c1efe883d18bd77495f96d811b26f46c530d8c85e12d8" Nov 21 14:35:47 crc kubenswrapper[4897]: I1121 14:35:47.427950 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5f61c9fbb7f4949782c1efe883d18bd77495f96d811b26f46c530d8c85e12d8"} err="failed to get container status \"e5f61c9fbb7f4949782c1efe883d18bd77495f96d811b26f46c530d8c85e12d8\": rpc error: code = NotFound desc = could not find container \"e5f61c9fbb7f4949782c1efe883d18bd77495f96d811b26f46c530d8c85e12d8\": container with ID starting with e5f61c9fbb7f4949782c1efe883d18bd77495f96d811b26f46c530d8c85e12d8 not found: ID does not exist" Nov 21 14:35:47 crc kubenswrapper[4897]: I1121 14:35:47.427976 4897 scope.go:117] "RemoveContainer" containerID="f2d77335ade2070c2371391c35bd65b08dfc83a9254e99cd03788d9b3e6f05f1" Nov 21 14:35:47 crc kubenswrapper[4897]: E1121 14:35:47.428242 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f2d77335ade2070c2371391c35bd65b08dfc83a9254e99cd03788d9b3e6f05f1\": container with ID starting with f2d77335ade2070c2371391c35bd65b08dfc83a9254e99cd03788d9b3e6f05f1 not found: ID does not exist" containerID="f2d77335ade2070c2371391c35bd65b08dfc83a9254e99cd03788d9b3e6f05f1" Nov 21 14:35:47 crc kubenswrapper[4897]: I1121 14:35:47.428273 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2d77335ade2070c2371391c35bd65b08dfc83a9254e99cd03788d9b3e6f05f1"} err="failed to get container status \"f2d77335ade2070c2371391c35bd65b08dfc83a9254e99cd03788d9b3e6f05f1\": rpc error: code = NotFound desc = could not find container \"f2d77335ade2070c2371391c35bd65b08dfc83a9254e99cd03788d9b3e6f05f1\": container with ID starting with f2d77335ade2070c2371391c35bd65b08dfc83a9254e99cd03788d9b3e6f05f1 not found: ID does not exist" Nov 21 14:35:47 crc kubenswrapper[4897]: I1121 14:35:47.428292 4897 scope.go:117] "RemoveContainer" containerID="e5f61c9fbb7f4949782c1efe883d18bd77495f96d811b26f46c530d8c85e12d8" Nov 21 14:35:47 crc kubenswrapper[4897]: I1121 14:35:47.428495 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5f61c9fbb7f4949782c1efe883d18bd77495f96d811b26f46c530d8c85e12d8"} err="failed to get container status \"e5f61c9fbb7f4949782c1efe883d18bd77495f96d811b26f46c530d8c85e12d8\": rpc error: code = NotFound desc = could not find container \"e5f61c9fbb7f4949782c1efe883d18bd77495f96d811b26f46c530d8c85e12d8\": container with ID starting with e5f61c9fbb7f4949782c1efe883d18bd77495f96d811b26f46c530d8c85e12d8 not found: ID does not exist" Nov 21 14:35:47 crc kubenswrapper[4897]: I1121 14:35:47.428540 4897 scope.go:117] "RemoveContainer" containerID="f2d77335ade2070c2371391c35bd65b08dfc83a9254e99cd03788d9b3e6f05f1" Nov 21 14:35:47 crc kubenswrapper[4897]: I1121 14:35:47.429009 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2d77335ade2070c2371391c35bd65b08dfc83a9254e99cd03788d9b3e6f05f1"} err="failed to get container status \"f2d77335ade2070c2371391c35bd65b08dfc83a9254e99cd03788d9b3e6f05f1\": rpc error: code = NotFound desc = could not find container \"f2d77335ade2070c2371391c35bd65b08dfc83a9254e99cd03788d9b3e6f05f1\": container with ID starting with f2d77335ade2070c2371391c35bd65b08dfc83a9254e99cd03788d9b3e6f05f1 not found: ID does not exist" Nov 21 14:35:47 crc kubenswrapper[4897]: I1121 14:35:47.443114 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/728674e7-ee48-41d2-ad13-7d6f623d99b9-logs\") pod \"728674e7-ee48-41d2-ad13-7d6f623d99b9\" (UID: \"728674e7-ee48-41d2-ad13-7d6f623d99b9\") " Nov 21 14:35:47 crc kubenswrapper[4897]: I1121 14:35:47.443228 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/728674e7-ee48-41d2-ad13-7d6f623d99b9-nova-metadata-tls-certs\") pod \"728674e7-ee48-41d2-ad13-7d6f623d99b9\" (UID: \"728674e7-ee48-41d2-ad13-7d6f623d99b9\") " Nov 21 14:35:47 crc kubenswrapper[4897]: I1121 14:35:47.443346 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/728674e7-ee48-41d2-ad13-7d6f623d99b9-config-data\") pod \"728674e7-ee48-41d2-ad13-7d6f623d99b9\" (UID: \"728674e7-ee48-41d2-ad13-7d6f623d99b9\") " Nov 21 14:35:47 crc kubenswrapper[4897]: I1121 14:35:47.443379 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/728674e7-ee48-41d2-ad13-7d6f623d99b9-combined-ca-bundle\") pod \"728674e7-ee48-41d2-ad13-7d6f623d99b9\" (UID: \"728674e7-ee48-41d2-ad13-7d6f623d99b9\") " Nov 21 14:35:47 crc kubenswrapper[4897]: I1121 14:35:47.443415 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cgjn9\" (UniqueName: \"kubernetes.io/projected/728674e7-ee48-41d2-ad13-7d6f623d99b9-kube-api-access-cgjn9\") pod \"728674e7-ee48-41d2-ad13-7d6f623d99b9\" (UID: \"728674e7-ee48-41d2-ad13-7d6f623d99b9\") " Nov 21 14:35:47 crc kubenswrapper[4897]: I1121 14:35:47.444232 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/728674e7-ee48-41d2-ad13-7d6f623d99b9-logs" (OuterVolumeSpecName: "logs") pod "728674e7-ee48-41d2-ad13-7d6f623d99b9" (UID: "728674e7-ee48-41d2-ad13-7d6f623d99b9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:35:47 crc kubenswrapper[4897]: I1121 14:35:47.449683 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/728674e7-ee48-41d2-ad13-7d6f623d99b9-kube-api-access-cgjn9" (OuterVolumeSpecName: "kube-api-access-cgjn9") pod "728674e7-ee48-41d2-ad13-7d6f623d99b9" (UID: "728674e7-ee48-41d2-ad13-7d6f623d99b9"). InnerVolumeSpecName "kube-api-access-cgjn9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:35:47 crc kubenswrapper[4897]: I1121 14:35:47.477200 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/728674e7-ee48-41d2-ad13-7d6f623d99b9-config-data" (OuterVolumeSpecName: "config-data") pod "728674e7-ee48-41d2-ad13-7d6f623d99b9" (UID: "728674e7-ee48-41d2-ad13-7d6f623d99b9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:35:47 crc kubenswrapper[4897]: I1121 14:35:47.480032 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/728674e7-ee48-41d2-ad13-7d6f623d99b9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "728674e7-ee48-41d2-ad13-7d6f623d99b9" (UID: "728674e7-ee48-41d2-ad13-7d6f623d99b9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:35:47 crc kubenswrapper[4897]: I1121 14:35:47.515368 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/728674e7-ee48-41d2-ad13-7d6f623d99b9-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "728674e7-ee48-41d2-ad13-7d6f623d99b9" (UID: "728674e7-ee48-41d2-ad13-7d6f623d99b9"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:35:47 crc kubenswrapper[4897]: I1121 14:35:47.547095 4897 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/728674e7-ee48-41d2-ad13-7d6f623d99b9-logs\") on node \"crc\" DevicePath \"\"" Nov 21 14:35:47 crc kubenswrapper[4897]: I1121 14:35:47.547135 4897 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/728674e7-ee48-41d2-ad13-7d6f623d99b9-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 21 14:35:47 crc kubenswrapper[4897]: I1121 14:35:47.547148 4897 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/728674e7-ee48-41d2-ad13-7d6f623d99b9-config-data\") on node \"crc\" DevicePath \"\"" Nov 21 14:35:47 crc kubenswrapper[4897]: I1121 14:35:47.547160 4897 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/728674e7-ee48-41d2-ad13-7d6f623d99b9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 21 14:35:47 crc kubenswrapper[4897]: I1121 14:35:47.547169 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cgjn9\" (UniqueName: \"kubernetes.io/projected/728674e7-ee48-41d2-ad13-7d6f623d99b9-kube-api-access-cgjn9\") on node \"crc\" DevicePath \"\"" Nov 21 14:35:47 crc kubenswrapper[4897]: I1121 14:35:47.716158 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 21 14:35:47 crc kubenswrapper[4897]: I1121 14:35:47.734784 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 21 14:35:47 crc kubenswrapper[4897]: I1121 14:35:47.756336 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 21 14:35:47 crc kubenswrapper[4897]: E1121 14:35:47.757056 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afa69523-63c4-4a94-b94a-61c1343ef10c" containerName="dnsmasq-dns" Nov 21 14:35:47 crc kubenswrapper[4897]: I1121 14:35:47.757157 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="afa69523-63c4-4a94-b94a-61c1343ef10c" containerName="dnsmasq-dns" Nov 21 14:35:47 crc kubenswrapper[4897]: E1121 14:35:47.757226 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="728674e7-ee48-41d2-ad13-7d6f623d99b9" containerName="nova-metadata-log" Nov 21 14:35:47 crc kubenswrapper[4897]: I1121 14:35:47.757276 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="728674e7-ee48-41d2-ad13-7d6f623d99b9" containerName="nova-metadata-log" Nov 21 14:35:47 crc kubenswrapper[4897]: E1121 14:35:47.757362 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="518b64a4-d073-462f-ada5-20209e14ff27" containerName="nova-manage" Nov 21 14:35:47 crc kubenswrapper[4897]: I1121 14:35:47.757420 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="518b64a4-d073-462f-ada5-20209e14ff27" containerName="nova-manage" Nov 21 14:35:47 crc kubenswrapper[4897]: E1121 14:35:47.757498 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afa69523-63c4-4a94-b94a-61c1343ef10c" containerName="init" Nov 21 14:35:47 crc kubenswrapper[4897]: I1121 14:35:47.757578 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="afa69523-63c4-4a94-b94a-61c1343ef10c" containerName="init" Nov 21 14:35:47 crc kubenswrapper[4897]: E1121 14:35:47.757642 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="728674e7-ee48-41d2-ad13-7d6f623d99b9" containerName="nova-metadata-metadata" Nov 21 14:35:47 crc kubenswrapper[4897]: I1121 14:35:47.757692 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="728674e7-ee48-41d2-ad13-7d6f623d99b9" containerName="nova-metadata-metadata" Nov 21 14:35:47 crc kubenswrapper[4897]: I1121 14:35:47.757974 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="afa69523-63c4-4a94-b94a-61c1343ef10c" containerName="dnsmasq-dns" Nov 21 14:35:47 crc kubenswrapper[4897]: I1121 14:35:47.758057 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="728674e7-ee48-41d2-ad13-7d6f623d99b9" containerName="nova-metadata-log" Nov 21 14:35:47 crc kubenswrapper[4897]: I1121 14:35:47.758127 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="728674e7-ee48-41d2-ad13-7d6f623d99b9" containerName="nova-metadata-metadata" Nov 21 14:35:47 crc kubenswrapper[4897]: I1121 14:35:47.758221 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="518b64a4-d073-462f-ada5-20209e14ff27" containerName="nova-manage" Nov 21 14:35:47 crc kubenswrapper[4897]: I1121 14:35:47.761052 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 21 14:35:47 crc kubenswrapper[4897]: I1121 14:35:47.767176 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 21 14:35:47 crc kubenswrapper[4897]: I1121 14:35:47.767821 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 21 14:35:47 crc kubenswrapper[4897]: I1121 14:35:47.770295 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 21 14:35:47 crc kubenswrapper[4897]: I1121 14:35:47.963111 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6810bdc5-077d-44fc-a616-3582d0929d42-logs\") pod \"nova-metadata-0\" (UID: \"6810bdc5-077d-44fc-a616-3582d0929d42\") " pod="openstack/nova-metadata-0" Nov 21 14:35:47 crc kubenswrapper[4897]: I1121 14:35:47.963172 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6wvhk\" (UniqueName: \"kubernetes.io/projected/6810bdc5-077d-44fc-a616-3582d0929d42-kube-api-access-6wvhk\") pod \"nova-metadata-0\" (UID: \"6810bdc5-077d-44fc-a616-3582d0929d42\") " pod="openstack/nova-metadata-0" Nov 21 14:35:47 crc kubenswrapper[4897]: I1121 14:35:47.963208 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6810bdc5-077d-44fc-a616-3582d0929d42-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"6810bdc5-077d-44fc-a616-3582d0929d42\") " pod="openstack/nova-metadata-0" Nov 21 14:35:47 crc kubenswrapper[4897]: I1121 14:35:47.963294 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6810bdc5-077d-44fc-a616-3582d0929d42-config-data\") pod \"nova-metadata-0\" (UID: \"6810bdc5-077d-44fc-a616-3582d0929d42\") " pod="openstack/nova-metadata-0" Nov 21 14:35:47 crc kubenswrapper[4897]: I1121 14:35:47.963425 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6810bdc5-077d-44fc-a616-3582d0929d42-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6810bdc5-077d-44fc-a616-3582d0929d42\") " pod="openstack/nova-metadata-0" Nov 21 14:35:48 crc kubenswrapper[4897]: I1121 14:35:48.065476 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6810bdc5-077d-44fc-a616-3582d0929d42-logs\") pod \"nova-metadata-0\" (UID: \"6810bdc5-077d-44fc-a616-3582d0929d42\") " pod="openstack/nova-metadata-0" Nov 21 14:35:48 crc kubenswrapper[4897]: I1121 14:35:48.065567 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6wvhk\" (UniqueName: \"kubernetes.io/projected/6810bdc5-077d-44fc-a616-3582d0929d42-kube-api-access-6wvhk\") pod \"nova-metadata-0\" (UID: \"6810bdc5-077d-44fc-a616-3582d0929d42\") " pod="openstack/nova-metadata-0" Nov 21 14:35:48 crc kubenswrapper[4897]: I1121 14:35:48.065617 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6810bdc5-077d-44fc-a616-3582d0929d42-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"6810bdc5-077d-44fc-a616-3582d0929d42\") " pod="openstack/nova-metadata-0" Nov 21 14:35:48 crc kubenswrapper[4897]: I1121 14:35:48.065708 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6810bdc5-077d-44fc-a616-3582d0929d42-config-data\") pod \"nova-metadata-0\" (UID: \"6810bdc5-077d-44fc-a616-3582d0929d42\") " pod="openstack/nova-metadata-0" Nov 21 14:35:48 crc kubenswrapper[4897]: I1121 14:35:48.066317 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6810bdc5-077d-44fc-a616-3582d0929d42-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6810bdc5-077d-44fc-a616-3582d0929d42\") " pod="openstack/nova-metadata-0" Nov 21 14:35:48 crc kubenswrapper[4897]: I1121 14:35:48.066727 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6810bdc5-077d-44fc-a616-3582d0929d42-logs\") pod \"nova-metadata-0\" (UID: \"6810bdc5-077d-44fc-a616-3582d0929d42\") " pod="openstack/nova-metadata-0" Nov 21 14:35:48 crc kubenswrapper[4897]: I1121 14:35:48.073485 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6810bdc5-077d-44fc-a616-3582d0929d42-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"6810bdc5-077d-44fc-a616-3582d0929d42\") " pod="openstack/nova-metadata-0" Nov 21 14:35:48 crc kubenswrapper[4897]: I1121 14:35:48.073585 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6810bdc5-077d-44fc-a616-3582d0929d42-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6810bdc5-077d-44fc-a616-3582d0929d42\") " pod="openstack/nova-metadata-0" Nov 21 14:35:48 crc kubenswrapper[4897]: I1121 14:35:48.074010 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6810bdc5-077d-44fc-a616-3582d0929d42-config-data\") pod \"nova-metadata-0\" (UID: \"6810bdc5-077d-44fc-a616-3582d0929d42\") " pod="openstack/nova-metadata-0" Nov 21 14:35:48 crc kubenswrapper[4897]: I1121 14:35:48.082715 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6wvhk\" (UniqueName: \"kubernetes.io/projected/6810bdc5-077d-44fc-a616-3582d0929d42-kube-api-access-6wvhk\") pod \"nova-metadata-0\" (UID: \"6810bdc5-077d-44fc-a616-3582d0929d42\") " pod="openstack/nova-metadata-0" Nov 21 14:35:48 crc kubenswrapper[4897]: I1121 14:35:48.085948 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 21 14:35:48 crc kubenswrapper[4897]: I1121 14:35:48.101541 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="728674e7-ee48-41d2-ad13-7d6f623d99b9" path="/var/lib/kubelet/pods/728674e7-ee48-41d2-ad13-7d6f623d99b9/volumes" Nov 21 14:35:48 crc kubenswrapper[4897]: E1121 14:35:48.283430 4897 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podafa69523_63c4_4a94_b94a_61c1343ef10c.slice/crio-conmon-6cb3731d3191a0096a5051532c43fd96973f7663ef9d3571618a165b3dd1afca.scope\": RecentStats: unable to find data in memory cache]" Nov 21 14:35:48 crc kubenswrapper[4897]: E1121 14:35:48.283586 4897 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podafa69523_63c4_4a94_b94a_61c1343ef10c.slice/crio-conmon-6cb3731d3191a0096a5051532c43fd96973f7663ef9d3571618a165b3dd1afca.scope\": RecentStats: unable to find data in memory cache]" Nov 21 14:35:48 crc kubenswrapper[4897]: I1121 14:35:48.397349 4897 generic.go:334] "Generic (PLEG): container finished" podID="4ca85306-d5f3-414f-abd6-6f4623335792" containerID="976eada4d6f64a9db0df4c95e99e61c1f4b0e1a254cae5367f8fca2fc8b41061" exitCode=0 Nov 21 14:35:48 crc kubenswrapper[4897]: I1121 14:35:48.397448 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"4ca85306-d5f3-414f-abd6-6f4623335792","Type":"ContainerDied","Data":"976eada4d6f64a9db0df4c95e99e61c1f4b0e1a254cae5367f8fca2fc8b41061"} Nov 21 14:35:48 crc kubenswrapper[4897]: E1121 14:35:48.536183 4897 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 976eada4d6f64a9db0df4c95e99e61c1f4b0e1a254cae5367f8fca2fc8b41061 is running failed: container process not found" containerID="976eada4d6f64a9db0df4c95e99e61c1f4b0e1a254cae5367f8fca2fc8b41061" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 21 14:35:48 crc kubenswrapper[4897]: E1121 14:35:48.536608 4897 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 976eada4d6f64a9db0df4c95e99e61c1f4b0e1a254cae5367f8fca2fc8b41061 is running failed: container process not found" containerID="976eada4d6f64a9db0df4c95e99e61c1f4b0e1a254cae5367f8fca2fc8b41061" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 21 14:35:48 crc kubenswrapper[4897]: E1121 14:35:48.537188 4897 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 976eada4d6f64a9db0df4c95e99e61c1f4b0e1a254cae5367f8fca2fc8b41061 is running failed: container process not found" containerID="976eada4d6f64a9db0df4c95e99e61c1f4b0e1a254cae5367f8fca2fc8b41061" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 21 14:35:48 crc kubenswrapper[4897]: E1121 14:35:48.537269 4897 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 976eada4d6f64a9db0df4c95e99e61c1f4b0e1a254cae5367f8fca2fc8b41061 is running failed: container process not found" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="4ca85306-d5f3-414f-abd6-6f4623335792" containerName="nova-scheduler-scheduler" Nov 21 14:35:48 crc kubenswrapper[4897]: I1121 14:35:48.605015 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 21 14:35:49 crc kubenswrapper[4897]: I1121 14:35:49.000174 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 21 14:35:49 crc kubenswrapper[4897]: I1121 14:35:49.203360 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ca85306-d5f3-414f-abd6-6f4623335792-combined-ca-bundle\") pod \"4ca85306-d5f3-414f-abd6-6f4623335792\" (UID: \"4ca85306-d5f3-414f-abd6-6f4623335792\") " Nov 21 14:35:49 crc kubenswrapper[4897]: I1121 14:35:49.203690 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ca85306-d5f3-414f-abd6-6f4623335792-config-data\") pod \"4ca85306-d5f3-414f-abd6-6f4623335792\" (UID: \"4ca85306-d5f3-414f-abd6-6f4623335792\") " Nov 21 14:35:49 crc kubenswrapper[4897]: I1121 14:35:49.203946 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qggzh\" (UniqueName: \"kubernetes.io/projected/4ca85306-d5f3-414f-abd6-6f4623335792-kube-api-access-qggzh\") pod \"4ca85306-d5f3-414f-abd6-6f4623335792\" (UID: \"4ca85306-d5f3-414f-abd6-6f4623335792\") " Nov 21 14:35:49 crc kubenswrapper[4897]: I1121 14:35:49.212843 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ca85306-d5f3-414f-abd6-6f4623335792-kube-api-access-qggzh" (OuterVolumeSpecName: "kube-api-access-qggzh") pod "4ca85306-d5f3-414f-abd6-6f4623335792" (UID: "4ca85306-d5f3-414f-abd6-6f4623335792"). InnerVolumeSpecName "kube-api-access-qggzh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:35:49 crc kubenswrapper[4897]: I1121 14:35:49.252133 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ca85306-d5f3-414f-abd6-6f4623335792-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4ca85306-d5f3-414f-abd6-6f4623335792" (UID: "4ca85306-d5f3-414f-abd6-6f4623335792"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:35:49 crc kubenswrapper[4897]: I1121 14:35:49.252161 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ca85306-d5f3-414f-abd6-6f4623335792-config-data" (OuterVolumeSpecName: "config-data") pod "4ca85306-d5f3-414f-abd6-6f4623335792" (UID: "4ca85306-d5f3-414f-abd6-6f4623335792"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:35:49 crc kubenswrapper[4897]: I1121 14:35:49.306432 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qggzh\" (UniqueName: \"kubernetes.io/projected/4ca85306-d5f3-414f-abd6-6f4623335792-kube-api-access-qggzh\") on node \"crc\" DevicePath \"\"" Nov 21 14:35:49 crc kubenswrapper[4897]: I1121 14:35:49.306468 4897 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ca85306-d5f3-414f-abd6-6f4623335792-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 21 14:35:49 crc kubenswrapper[4897]: I1121 14:35:49.306478 4897 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ca85306-d5f3-414f-abd6-6f4623335792-config-data\") on node \"crc\" DevicePath \"\"" Nov 21 14:35:49 crc kubenswrapper[4897]: I1121 14:35:49.410789 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"4ca85306-d5f3-414f-abd6-6f4623335792","Type":"ContainerDied","Data":"5a9429626e9d62c72e25ccce2999cba08de0e5257e599632a004c8dacd097ed2"} Nov 21 14:35:49 crc kubenswrapper[4897]: I1121 14:35:49.410819 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 21 14:35:49 crc kubenswrapper[4897]: I1121 14:35:49.411131 4897 scope.go:117] "RemoveContainer" containerID="976eada4d6f64a9db0df4c95e99e61c1f4b0e1a254cae5367f8fca2fc8b41061" Nov 21 14:35:49 crc kubenswrapper[4897]: I1121 14:35:49.413084 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6810bdc5-077d-44fc-a616-3582d0929d42","Type":"ContainerStarted","Data":"f4856114b27cbe08f59a71c5c4ba4eb464c61da14a4b1fbfd1dc2d65a22619ae"} Nov 21 14:35:49 crc kubenswrapper[4897]: I1121 14:35:49.413112 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6810bdc5-077d-44fc-a616-3582d0929d42","Type":"ContainerStarted","Data":"8fb72facf97f0167bacd8eb084a6f636d166cfb4100b89a35018d53ad964a621"} Nov 21 14:35:49 crc kubenswrapper[4897]: I1121 14:35:49.413122 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6810bdc5-077d-44fc-a616-3582d0929d42","Type":"ContainerStarted","Data":"aad9611cad00d40b8f5680fbd9d96866afb7427c68d5f8b904be2beee52320bf"} Nov 21 14:35:49 crc kubenswrapper[4897]: I1121 14:35:49.449930 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.449906881 podStartE2EDuration="2.449906881s" podCreationTimestamp="2025-11-21 14:35:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:35:49.441969397 +0000 UTC m=+1626.726562862" watchObservedRunningTime="2025-11-21 14:35:49.449906881 +0000 UTC m=+1626.734500366" Nov 21 14:35:49 crc kubenswrapper[4897]: E1121 14:35:49.467342 4897 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podafa69523_63c4_4a94_b94a_61c1343ef10c.slice/crio-conmon-6cb3731d3191a0096a5051532c43fd96973f7663ef9d3571618a165b3dd1afca.scope\": RecentStats: unable to find data in memory cache]" Nov 21 14:35:49 crc kubenswrapper[4897]: I1121 14:35:49.489838 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 21 14:35:49 crc kubenswrapper[4897]: I1121 14:35:49.530271 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Nov 21 14:35:49 crc kubenswrapper[4897]: I1121 14:35:49.544544 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 21 14:35:49 crc kubenswrapper[4897]: E1121 14:35:49.545090 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ca85306-d5f3-414f-abd6-6f4623335792" containerName="nova-scheduler-scheduler" Nov 21 14:35:49 crc kubenswrapper[4897]: I1121 14:35:49.545109 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ca85306-d5f3-414f-abd6-6f4623335792" containerName="nova-scheduler-scheduler" Nov 21 14:35:49 crc kubenswrapper[4897]: I1121 14:35:49.545327 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ca85306-d5f3-414f-abd6-6f4623335792" containerName="nova-scheduler-scheduler" Nov 21 14:35:49 crc kubenswrapper[4897]: I1121 14:35:49.546275 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 21 14:35:49 crc kubenswrapper[4897]: I1121 14:35:49.548576 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 21 14:35:49 crc kubenswrapper[4897]: I1121 14:35:49.557450 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 21 14:35:49 crc kubenswrapper[4897]: I1121 14:35:49.716427 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l4mm4\" (UniqueName: \"kubernetes.io/projected/7ee39fa1-44f9-46e8-9206-39d5ad1e2e6c-kube-api-access-l4mm4\") pod \"nova-scheduler-0\" (UID: \"7ee39fa1-44f9-46e8-9206-39d5ad1e2e6c\") " pod="openstack/nova-scheduler-0" Nov 21 14:35:49 crc kubenswrapper[4897]: I1121 14:35:49.716646 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ee39fa1-44f9-46e8-9206-39d5ad1e2e6c-config-data\") pod \"nova-scheduler-0\" (UID: \"7ee39fa1-44f9-46e8-9206-39d5ad1e2e6c\") " pod="openstack/nova-scheduler-0" Nov 21 14:35:49 crc kubenswrapper[4897]: I1121 14:35:49.716743 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ee39fa1-44f9-46e8-9206-39d5ad1e2e6c-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"7ee39fa1-44f9-46e8-9206-39d5ad1e2e6c\") " pod="openstack/nova-scheduler-0" Nov 21 14:35:49 crc kubenswrapper[4897]: I1121 14:35:49.819095 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l4mm4\" (UniqueName: \"kubernetes.io/projected/7ee39fa1-44f9-46e8-9206-39d5ad1e2e6c-kube-api-access-l4mm4\") pod \"nova-scheduler-0\" (UID: \"7ee39fa1-44f9-46e8-9206-39d5ad1e2e6c\") " pod="openstack/nova-scheduler-0" Nov 21 14:35:49 crc kubenswrapper[4897]: I1121 14:35:49.819171 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ee39fa1-44f9-46e8-9206-39d5ad1e2e6c-config-data\") pod \"nova-scheduler-0\" (UID: \"7ee39fa1-44f9-46e8-9206-39d5ad1e2e6c\") " pod="openstack/nova-scheduler-0" Nov 21 14:35:49 crc kubenswrapper[4897]: I1121 14:35:49.819200 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ee39fa1-44f9-46e8-9206-39d5ad1e2e6c-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"7ee39fa1-44f9-46e8-9206-39d5ad1e2e6c\") " pod="openstack/nova-scheduler-0" Nov 21 14:35:49 crc kubenswrapper[4897]: I1121 14:35:49.824848 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ee39fa1-44f9-46e8-9206-39d5ad1e2e6c-config-data\") pod \"nova-scheduler-0\" (UID: \"7ee39fa1-44f9-46e8-9206-39d5ad1e2e6c\") " pod="openstack/nova-scheduler-0" Nov 21 14:35:49 crc kubenswrapper[4897]: I1121 14:35:49.826120 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ee39fa1-44f9-46e8-9206-39d5ad1e2e6c-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"7ee39fa1-44f9-46e8-9206-39d5ad1e2e6c\") " pod="openstack/nova-scheduler-0" Nov 21 14:35:49 crc kubenswrapper[4897]: I1121 14:35:49.841293 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l4mm4\" (UniqueName: \"kubernetes.io/projected/7ee39fa1-44f9-46e8-9206-39d5ad1e2e6c-kube-api-access-l4mm4\") pod \"nova-scheduler-0\" (UID: \"7ee39fa1-44f9-46e8-9206-39d5ad1e2e6c\") " pod="openstack/nova-scheduler-0" Nov 21 14:35:49 crc kubenswrapper[4897]: I1121 14:35:49.875236 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 21 14:35:50 crc kubenswrapper[4897]: I1121 14:35:50.120542 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ca85306-d5f3-414f-abd6-6f4623335792" path="/var/lib/kubelet/pods/4ca85306-d5f3-414f-abd6-6f4623335792/volumes" Nov 21 14:35:50 crc kubenswrapper[4897]: I1121 14:35:50.311861 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 21 14:35:50 crc kubenswrapper[4897]: I1121 14:35:50.431986 4897 generic.go:334] "Generic (PLEG): container finished" podID="cf6d459c-961c-4361-a034-4a7c7d5ce7b8" containerID="34bf622d9e31992c9251f7bd8f165c2e2b6a33e70e9c16b91673f884f3228578" exitCode=0 Nov 21 14:35:50 crc kubenswrapper[4897]: I1121 14:35:50.432057 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"cf6d459c-961c-4361-a034-4a7c7d5ce7b8","Type":"ContainerDied","Data":"34bf622d9e31992c9251f7bd8f165c2e2b6a33e70e9c16b91673f884f3228578"} Nov 21 14:35:50 crc kubenswrapper[4897]: I1121 14:35:50.432065 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 21 14:35:50 crc kubenswrapper[4897]: I1121 14:35:50.432091 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"cf6d459c-961c-4361-a034-4a7c7d5ce7b8","Type":"ContainerDied","Data":"6bbbf0358422b5b0b39552525c776d4c383d4edadd6c977cdb1d9f47a5fdfc55"} Nov 21 14:35:50 crc kubenswrapper[4897]: I1121 14:35:50.432121 4897 scope.go:117] "RemoveContainer" containerID="34bf622d9e31992c9251f7bd8f165c2e2b6a33e70e9c16b91673f884f3228578" Nov 21 14:35:50 crc kubenswrapper[4897]: I1121 14:35:50.439436 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf6d459c-961c-4361-a034-4a7c7d5ce7b8-combined-ca-bundle\") pod \"cf6d459c-961c-4361-a034-4a7c7d5ce7b8\" (UID: \"cf6d459c-961c-4361-a034-4a7c7d5ce7b8\") " Nov 21 14:35:50 crc kubenswrapper[4897]: I1121 14:35:50.439580 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf6d459c-961c-4361-a034-4a7c7d5ce7b8-config-data\") pod \"cf6d459c-961c-4361-a034-4a7c7d5ce7b8\" (UID: \"cf6d459c-961c-4361-a034-4a7c7d5ce7b8\") " Nov 21 14:35:50 crc kubenswrapper[4897]: I1121 14:35:50.439640 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cf6d459c-961c-4361-a034-4a7c7d5ce7b8-logs\") pod \"cf6d459c-961c-4361-a034-4a7c7d5ce7b8\" (UID: \"cf6d459c-961c-4361-a034-4a7c7d5ce7b8\") " Nov 21 14:35:50 crc kubenswrapper[4897]: I1121 14:35:50.439722 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6bsn7\" (UniqueName: \"kubernetes.io/projected/cf6d459c-961c-4361-a034-4a7c7d5ce7b8-kube-api-access-6bsn7\") pod \"cf6d459c-961c-4361-a034-4a7c7d5ce7b8\" (UID: \"cf6d459c-961c-4361-a034-4a7c7d5ce7b8\") " Nov 21 14:35:50 crc kubenswrapper[4897]: I1121 14:35:50.444331 4897 generic.go:334] "Generic (PLEG): container finished" podID="53e2a389-78b0-463a-82df-dc3d1d870b7b" containerID="b82fb5422034aef083739986bf603d009ad47087d5635868123e0241717761b3" exitCode=137 Nov 21 14:35:50 crc kubenswrapper[4897]: I1121 14:35:50.444408 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"53e2a389-78b0-463a-82df-dc3d1d870b7b","Type":"ContainerDied","Data":"b82fb5422034aef083739986bf603d009ad47087d5635868123e0241717761b3"} Nov 21 14:35:50 crc kubenswrapper[4897]: I1121 14:35:50.457072 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cf6d459c-961c-4361-a034-4a7c7d5ce7b8-logs" (OuterVolumeSpecName: "logs") pod "cf6d459c-961c-4361-a034-4a7c7d5ce7b8" (UID: "cf6d459c-961c-4361-a034-4a7c7d5ce7b8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:35:50 crc kubenswrapper[4897]: I1121 14:35:50.471483 4897 scope.go:117] "RemoveContainer" containerID="162624ca9d5ca88cb13b9ab46129c2083d1c72de99ff20042857a4b9a95a6e82" Nov 21 14:35:50 crc kubenswrapper[4897]: I1121 14:35:50.472313 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf6d459c-961c-4361-a034-4a7c7d5ce7b8-kube-api-access-6bsn7" (OuterVolumeSpecName: "kube-api-access-6bsn7") pod "cf6d459c-961c-4361-a034-4a7c7d5ce7b8" (UID: "cf6d459c-961c-4361-a034-4a7c7d5ce7b8"). InnerVolumeSpecName "kube-api-access-6bsn7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:35:50 crc kubenswrapper[4897]: I1121 14:35:50.483343 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf6d459c-961c-4361-a034-4a7c7d5ce7b8-config-data" (OuterVolumeSpecName: "config-data") pod "cf6d459c-961c-4361-a034-4a7c7d5ce7b8" (UID: "cf6d459c-961c-4361-a034-4a7c7d5ce7b8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:35:50 crc kubenswrapper[4897]: E1121 14:35:50.484257 4897 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podafa69523_63c4_4a94_b94a_61c1343ef10c.slice/crio-conmon-6cb3731d3191a0096a5051532c43fd96973f7663ef9d3571618a165b3dd1afca.scope\": RecentStats: unable to find data in memory cache]" Nov 21 14:35:50 crc kubenswrapper[4897]: I1121 14:35:50.492729 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf6d459c-961c-4361-a034-4a7c7d5ce7b8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cf6d459c-961c-4361-a034-4a7c7d5ce7b8" (UID: "cf6d459c-961c-4361-a034-4a7c7d5ce7b8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:35:50 crc kubenswrapper[4897]: I1121 14:35:50.542236 4897 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf6d459c-961c-4361-a034-4a7c7d5ce7b8-config-data\") on node \"crc\" DevicePath \"\"" Nov 21 14:35:50 crc kubenswrapper[4897]: I1121 14:35:50.542268 4897 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cf6d459c-961c-4361-a034-4a7c7d5ce7b8-logs\") on node \"crc\" DevicePath \"\"" Nov 21 14:35:50 crc kubenswrapper[4897]: I1121 14:35:50.542282 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6bsn7\" (UniqueName: \"kubernetes.io/projected/cf6d459c-961c-4361-a034-4a7c7d5ce7b8-kube-api-access-6bsn7\") on node \"crc\" DevicePath \"\"" Nov 21 14:35:50 crc kubenswrapper[4897]: I1121 14:35:50.542291 4897 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf6d459c-961c-4361-a034-4a7c7d5ce7b8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 21 14:35:50 crc kubenswrapper[4897]: I1121 14:35:50.555612 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 21 14:35:50 crc kubenswrapper[4897]: I1121 14:35:50.605086 4897 scope.go:117] "RemoveContainer" containerID="34bf622d9e31992c9251f7bd8f165c2e2b6a33e70e9c16b91673f884f3228578" Nov 21 14:35:50 crc kubenswrapper[4897]: E1121 14:35:50.605745 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"34bf622d9e31992c9251f7bd8f165c2e2b6a33e70e9c16b91673f884f3228578\": container with ID starting with 34bf622d9e31992c9251f7bd8f165c2e2b6a33e70e9c16b91673f884f3228578 not found: ID does not exist" containerID="34bf622d9e31992c9251f7bd8f165c2e2b6a33e70e9c16b91673f884f3228578" Nov 21 14:35:50 crc kubenswrapper[4897]: I1121 14:35:50.605805 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"34bf622d9e31992c9251f7bd8f165c2e2b6a33e70e9c16b91673f884f3228578"} err="failed to get container status \"34bf622d9e31992c9251f7bd8f165c2e2b6a33e70e9c16b91673f884f3228578\": rpc error: code = NotFound desc = could not find container \"34bf622d9e31992c9251f7bd8f165c2e2b6a33e70e9c16b91673f884f3228578\": container with ID starting with 34bf622d9e31992c9251f7bd8f165c2e2b6a33e70e9c16b91673f884f3228578 not found: ID does not exist" Nov 21 14:35:50 crc kubenswrapper[4897]: I1121 14:35:50.605826 4897 scope.go:117] "RemoveContainer" containerID="162624ca9d5ca88cb13b9ab46129c2083d1c72de99ff20042857a4b9a95a6e82" Nov 21 14:35:50 crc kubenswrapper[4897]: E1121 14:35:50.606486 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"162624ca9d5ca88cb13b9ab46129c2083d1c72de99ff20042857a4b9a95a6e82\": container with ID starting with 162624ca9d5ca88cb13b9ab46129c2083d1c72de99ff20042857a4b9a95a6e82 not found: ID does not exist" containerID="162624ca9d5ca88cb13b9ab46129c2083d1c72de99ff20042857a4b9a95a6e82" Nov 21 14:35:50 crc kubenswrapper[4897]: I1121 14:35:50.606534 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"162624ca9d5ca88cb13b9ab46129c2083d1c72de99ff20042857a4b9a95a6e82"} err="failed to get container status \"162624ca9d5ca88cb13b9ab46129c2083d1c72de99ff20042857a4b9a95a6e82\": rpc error: code = NotFound desc = could not find container \"162624ca9d5ca88cb13b9ab46129c2083d1c72de99ff20042857a4b9a95a6e82\": container with ID starting with 162624ca9d5ca88cb13b9ab46129c2083d1c72de99ff20042857a4b9a95a6e82 not found: ID does not exist" Nov 21 14:35:50 crc kubenswrapper[4897]: I1121 14:35:50.727754 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Nov 21 14:35:50 crc kubenswrapper[4897]: I1121 14:35:50.805345 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 21 14:35:50 crc kubenswrapper[4897]: I1121 14:35:50.826127 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 21 14:35:50 crc kubenswrapper[4897]: I1121 14:35:50.837613 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 21 14:35:50 crc kubenswrapper[4897]: E1121 14:35:50.838300 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53e2a389-78b0-463a-82df-dc3d1d870b7b" containerName="aodh-evaluator" Nov 21 14:35:50 crc kubenswrapper[4897]: I1121 14:35:50.838351 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="53e2a389-78b0-463a-82df-dc3d1d870b7b" containerName="aodh-evaluator" Nov 21 14:35:50 crc kubenswrapper[4897]: E1121 14:35:50.838379 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53e2a389-78b0-463a-82df-dc3d1d870b7b" containerName="aodh-listener" Nov 21 14:35:50 crc kubenswrapper[4897]: I1121 14:35:50.838388 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="53e2a389-78b0-463a-82df-dc3d1d870b7b" containerName="aodh-listener" Nov 21 14:35:50 crc kubenswrapper[4897]: E1121 14:35:50.838408 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53e2a389-78b0-463a-82df-dc3d1d870b7b" containerName="aodh-notifier" Nov 21 14:35:50 crc kubenswrapper[4897]: I1121 14:35:50.838417 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="53e2a389-78b0-463a-82df-dc3d1d870b7b" containerName="aodh-notifier" Nov 21 14:35:50 crc kubenswrapper[4897]: E1121 14:35:50.838427 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf6d459c-961c-4361-a034-4a7c7d5ce7b8" containerName="nova-api-log" Nov 21 14:35:50 crc kubenswrapper[4897]: I1121 14:35:50.838439 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf6d459c-961c-4361-a034-4a7c7d5ce7b8" containerName="nova-api-log" Nov 21 14:35:50 crc kubenswrapper[4897]: E1121 14:35:50.838466 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf6d459c-961c-4361-a034-4a7c7d5ce7b8" containerName="nova-api-api" Nov 21 14:35:50 crc kubenswrapper[4897]: I1121 14:35:50.838473 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf6d459c-961c-4361-a034-4a7c7d5ce7b8" containerName="nova-api-api" Nov 21 14:35:50 crc kubenswrapper[4897]: E1121 14:35:50.838483 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53e2a389-78b0-463a-82df-dc3d1d870b7b" containerName="aodh-api" Nov 21 14:35:50 crc kubenswrapper[4897]: I1121 14:35:50.838489 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="53e2a389-78b0-463a-82df-dc3d1d870b7b" containerName="aodh-api" Nov 21 14:35:50 crc kubenswrapper[4897]: I1121 14:35:50.838771 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="53e2a389-78b0-463a-82df-dc3d1d870b7b" containerName="aodh-evaluator" Nov 21 14:35:50 crc kubenswrapper[4897]: I1121 14:35:50.838805 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="53e2a389-78b0-463a-82df-dc3d1d870b7b" containerName="aodh-listener" Nov 21 14:35:50 crc kubenswrapper[4897]: I1121 14:35:50.838821 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="53e2a389-78b0-463a-82df-dc3d1d870b7b" containerName="aodh-notifier" Nov 21 14:35:50 crc kubenswrapper[4897]: I1121 14:35:50.838832 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf6d459c-961c-4361-a034-4a7c7d5ce7b8" containerName="nova-api-log" Nov 21 14:35:50 crc kubenswrapper[4897]: I1121 14:35:50.838850 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="53e2a389-78b0-463a-82df-dc3d1d870b7b" containerName="aodh-api" Nov 21 14:35:50 crc kubenswrapper[4897]: I1121 14:35:50.838866 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf6d459c-961c-4361-a034-4a7c7d5ce7b8" containerName="nova-api-api" Nov 21 14:35:50 crc kubenswrapper[4897]: I1121 14:35:50.840805 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 21 14:35:50 crc kubenswrapper[4897]: I1121 14:35:50.848447 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53e2a389-78b0-463a-82df-dc3d1d870b7b-combined-ca-bundle\") pod \"53e2a389-78b0-463a-82df-dc3d1d870b7b\" (UID: \"53e2a389-78b0-463a-82df-dc3d1d870b7b\") " Nov 21 14:35:50 crc kubenswrapper[4897]: I1121 14:35:50.848661 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53e2a389-78b0-463a-82df-dc3d1d870b7b-scripts\") pod \"53e2a389-78b0-463a-82df-dc3d1d870b7b\" (UID: \"53e2a389-78b0-463a-82df-dc3d1d870b7b\") " Nov 21 14:35:50 crc kubenswrapper[4897]: I1121 14:35:50.848762 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53e2a389-78b0-463a-82df-dc3d1d870b7b-config-data\") pod \"53e2a389-78b0-463a-82df-dc3d1d870b7b\" (UID: \"53e2a389-78b0-463a-82df-dc3d1d870b7b\") " Nov 21 14:35:50 crc kubenswrapper[4897]: I1121 14:35:50.848879 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hmg45\" (UniqueName: \"kubernetes.io/projected/53e2a389-78b0-463a-82df-dc3d1d870b7b-kube-api-access-hmg45\") pod \"53e2a389-78b0-463a-82df-dc3d1d870b7b\" (UID: \"53e2a389-78b0-463a-82df-dc3d1d870b7b\") " Nov 21 14:35:50 crc kubenswrapper[4897]: I1121 14:35:50.849269 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 21 14:35:50 crc kubenswrapper[4897]: I1121 14:35:50.849570 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 21 14:35:50 crc kubenswrapper[4897]: I1121 14:35:50.854628 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53e2a389-78b0-463a-82df-dc3d1d870b7b-kube-api-access-hmg45" (OuterVolumeSpecName: "kube-api-access-hmg45") pod "53e2a389-78b0-463a-82df-dc3d1d870b7b" (UID: "53e2a389-78b0-463a-82df-dc3d1d870b7b"). InnerVolumeSpecName "kube-api-access-hmg45". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:35:50 crc kubenswrapper[4897]: I1121 14:35:50.855023 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53e2a389-78b0-463a-82df-dc3d1d870b7b-scripts" (OuterVolumeSpecName: "scripts") pod "53e2a389-78b0-463a-82df-dc3d1d870b7b" (UID: "53e2a389-78b0-463a-82df-dc3d1d870b7b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:35:50 crc kubenswrapper[4897]: I1121 14:35:50.951257 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hmfwh\" (UniqueName: \"kubernetes.io/projected/d3a7078f-ec31-46d5-aadb-b119532328b8-kube-api-access-hmfwh\") pod \"nova-api-0\" (UID: \"d3a7078f-ec31-46d5-aadb-b119532328b8\") " pod="openstack/nova-api-0" Nov 21 14:35:50 crc kubenswrapper[4897]: I1121 14:35:50.951337 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d3a7078f-ec31-46d5-aadb-b119532328b8-logs\") pod \"nova-api-0\" (UID: \"d3a7078f-ec31-46d5-aadb-b119532328b8\") " pod="openstack/nova-api-0" Nov 21 14:35:50 crc kubenswrapper[4897]: I1121 14:35:50.951572 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3a7078f-ec31-46d5-aadb-b119532328b8-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d3a7078f-ec31-46d5-aadb-b119532328b8\") " pod="openstack/nova-api-0" Nov 21 14:35:50 crc kubenswrapper[4897]: I1121 14:35:50.952021 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3a7078f-ec31-46d5-aadb-b119532328b8-config-data\") pod \"nova-api-0\" (UID: \"d3a7078f-ec31-46d5-aadb-b119532328b8\") " pod="openstack/nova-api-0" Nov 21 14:35:50 crc kubenswrapper[4897]: I1121 14:35:50.952715 4897 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53e2a389-78b0-463a-82df-dc3d1d870b7b-scripts\") on node \"crc\" DevicePath \"\"" Nov 21 14:35:50 crc kubenswrapper[4897]: I1121 14:35:50.952751 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hmg45\" (UniqueName: \"kubernetes.io/projected/53e2a389-78b0-463a-82df-dc3d1d870b7b-kube-api-access-hmg45\") on node \"crc\" DevicePath \"\"" Nov 21 14:35:50 crc kubenswrapper[4897]: I1121 14:35:50.979599 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53e2a389-78b0-463a-82df-dc3d1d870b7b-config-data" (OuterVolumeSpecName: "config-data") pod "53e2a389-78b0-463a-82df-dc3d1d870b7b" (UID: "53e2a389-78b0-463a-82df-dc3d1d870b7b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:35:50 crc kubenswrapper[4897]: I1121 14:35:50.980177 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53e2a389-78b0-463a-82df-dc3d1d870b7b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "53e2a389-78b0-463a-82df-dc3d1d870b7b" (UID: "53e2a389-78b0-463a-82df-dc3d1d870b7b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:35:51 crc kubenswrapper[4897]: I1121 14:35:51.054387 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3a7078f-ec31-46d5-aadb-b119532328b8-config-data\") pod \"nova-api-0\" (UID: \"d3a7078f-ec31-46d5-aadb-b119532328b8\") " pod="openstack/nova-api-0" Nov 21 14:35:51 crc kubenswrapper[4897]: I1121 14:35:51.054472 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hmfwh\" (UniqueName: \"kubernetes.io/projected/d3a7078f-ec31-46d5-aadb-b119532328b8-kube-api-access-hmfwh\") pod \"nova-api-0\" (UID: \"d3a7078f-ec31-46d5-aadb-b119532328b8\") " pod="openstack/nova-api-0" Nov 21 14:35:51 crc kubenswrapper[4897]: I1121 14:35:51.054544 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d3a7078f-ec31-46d5-aadb-b119532328b8-logs\") pod \"nova-api-0\" (UID: \"d3a7078f-ec31-46d5-aadb-b119532328b8\") " pod="openstack/nova-api-0" Nov 21 14:35:51 crc kubenswrapper[4897]: I1121 14:35:51.054599 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3a7078f-ec31-46d5-aadb-b119532328b8-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d3a7078f-ec31-46d5-aadb-b119532328b8\") " pod="openstack/nova-api-0" Nov 21 14:35:51 crc kubenswrapper[4897]: I1121 14:35:51.054691 4897 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53e2a389-78b0-463a-82df-dc3d1d870b7b-config-data\") on node \"crc\" DevicePath \"\"" Nov 21 14:35:51 crc kubenswrapper[4897]: I1121 14:35:51.054707 4897 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53e2a389-78b0-463a-82df-dc3d1d870b7b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 21 14:35:51 crc kubenswrapper[4897]: I1121 14:35:51.055434 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d3a7078f-ec31-46d5-aadb-b119532328b8-logs\") pod \"nova-api-0\" (UID: \"d3a7078f-ec31-46d5-aadb-b119532328b8\") " pod="openstack/nova-api-0" Nov 21 14:35:51 crc kubenswrapper[4897]: I1121 14:35:51.058697 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3a7078f-ec31-46d5-aadb-b119532328b8-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d3a7078f-ec31-46d5-aadb-b119532328b8\") " pod="openstack/nova-api-0" Nov 21 14:35:51 crc kubenswrapper[4897]: I1121 14:35:51.059878 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3a7078f-ec31-46d5-aadb-b119532328b8-config-data\") pod \"nova-api-0\" (UID: \"d3a7078f-ec31-46d5-aadb-b119532328b8\") " pod="openstack/nova-api-0" Nov 21 14:35:51 crc kubenswrapper[4897]: I1121 14:35:51.071255 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hmfwh\" (UniqueName: \"kubernetes.io/projected/d3a7078f-ec31-46d5-aadb-b119532328b8-kube-api-access-hmfwh\") pod \"nova-api-0\" (UID: \"d3a7078f-ec31-46d5-aadb-b119532328b8\") " pod="openstack/nova-api-0" Nov 21 14:35:51 crc kubenswrapper[4897]: I1121 14:35:51.164729 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 21 14:35:51 crc kubenswrapper[4897]: I1121 14:35:51.473693 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"53e2a389-78b0-463a-82df-dc3d1d870b7b","Type":"ContainerDied","Data":"71192c01fc3c53970c06f13f68ea3d7f4f6667b0812854d8bae5a4272c250d4b"} Nov 21 14:35:51 crc kubenswrapper[4897]: I1121 14:35:51.474305 4897 scope.go:117] "RemoveContainer" containerID="b82fb5422034aef083739986bf603d009ad47087d5635868123e0241717761b3" Nov 21 14:35:51 crc kubenswrapper[4897]: I1121 14:35:51.473724 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Nov 21 14:35:51 crc kubenswrapper[4897]: I1121 14:35:51.479018 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"7ee39fa1-44f9-46e8-9206-39d5ad1e2e6c","Type":"ContainerStarted","Data":"fdaf635e274efe1f2c24cf94ed630f0e027ba9d8b0078089b84a1f4dfb4949ba"} Nov 21 14:35:51 crc kubenswrapper[4897]: I1121 14:35:51.479063 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"7ee39fa1-44f9-46e8-9206-39d5ad1e2e6c","Type":"ContainerStarted","Data":"ba416c5a2341719fb456ec3197d8e3416e6a5195c4e1c7cc12249e52f9daa04a"} Nov 21 14:35:51 crc kubenswrapper[4897]: I1121 14:35:51.508850 4897 scope.go:117] "RemoveContainer" containerID="d2f404c2708b34449a7d699e98fd40e367eaae057219a0725d1c6c79b5893c1e" Nov 21 14:35:51 crc kubenswrapper[4897]: I1121 14:35:51.517434 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.517389106 podStartE2EDuration="2.517389106s" podCreationTimestamp="2025-11-21 14:35:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:35:51.502353611 +0000 UTC m=+1628.786947106" watchObservedRunningTime="2025-11-21 14:35:51.517389106 +0000 UTC m=+1628.801982591" Nov 21 14:35:51 crc kubenswrapper[4897]: I1121 14:35:51.548753 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Nov 21 14:35:51 crc kubenswrapper[4897]: I1121 14:35:51.552631 4897 scope.go:117] "RemoveContainer" containerID="51ef72dcaa68128a85131783b7460b847ef18d69b8d518d355a8e2fc52cea669" Nov 21 14:35:51 crc kubenswrapper[4897]: I1121 14:35:51.573742 4897 scope.go:117] "RemoveContainer" containerID="739afdf043be77117d422b7a6c54fdb232351f5cf29a3121b8e40c85d307efbc" Nov 21 14:35:51 crc kubenswrapper[4897]: I1121 14:35:51.574899 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-0"] Nov 21 14:35:51 crc kubenswrapper[4897]: I1121 14:35:51.585426 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Nov 21 14:35:51 crc kubenswrapper[4897]: I1121 14:35:51.588212 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Nov 21 14:35:51 crc kubenswrapper[4897]: I1121 14:35:51.590721 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Nov 21 14:35:51 crc kubenswrapper[4897]: I1121 14:35:51.591772 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-public-svc" Nov 21 14:35:51 crc kubenswrapper[4897]: I1121 14:35:51.591852 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-internal-svc" Nov 21 14:35:51 crc kubenswrapper[4897]: I1121 14:35:51.591809 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Nov 21 14:35:51 crc kubenswrapper[4897]: I1121 14:35:51.592246 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-m57c8" Nov 21 14:35:51 crc kubenswrapper[4897]: I1121 14:35:51.598539 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Nov 21 14:35:51 crc kubenswrapper[4897]: I1121 14:35:51.661875 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 21 14:35:51 crc kubenswrapper[4897]: I1121 14:35:51.774704 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d5dac4b-9e6c-41d5-9d10-87015da49a79-internal-tls-certs\") pod \"aodh-0\" (UID: \"1d5dac4b-9e6c-41d5-9d10-87015da49a79\") " pod="openstack/aodh-0" Nov 21 14:35:51 crc kubenswrapper[4897]: I1121 14:35:51.774790 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d5dac4b-9e6c-41d5-9d10-87015da49a79-combined-ca-bundle\") pod \"aodh-0\" (UID: \"1d5dac4b-9e6c-41d5-9d10-87015da49a79\") " pod="openstack/aodh-0" Nov 21 14:35:51 crc kubenswrapper[4897]: I1121 14:35:51.774810 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d5dac4b-9e6c-41d5-9d10-87015da49a79-config-data\") pod \"aodh-0\" (UID: \"1d5dac4b-9e6c-41d5-9d10-87015da49a79\") " pod="openstack/aodh-0" Nov 21 14:35:51 crc kubenswrapper[4897]: I1121 14:35:51.774956 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d5dac4b-9e6c-41d5-9d10-87015da49a79-scripts\") pod \"aodh-0\" (UID: \"1d5dac4b-9e6c-41d5-9d10-87015da49a79\") " pod="openstack/aodh-0" Nov 21 14:35:51 crc kubenswrapper[4897]: I1121 14:35:51.774991 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d5dac4b-9e6c-41d5-9d10-87015da49a79-public-tls-certs\") pod \"aodh-0\" (UID: \"1d5dac4b-9e6c-41d5-9d10-87015da49a79\") " pod="openstack/aodh-0" Nov 21 14:35:51 crc kubenswrapper[4897]: I1121 14:35:51.775025 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6dmkw\" (UniqueName: \"kubernetes.io/projected/1d5dac4b-9e6c-41d5-9d10-87015da49a79-kube-api-access-6dmkw\") pod \"aodh-0\" (UID: \"1d5dac4b-9e6c-41d5-9d10-87015da49a79\") " pod="openstack/aodh-0" Nov 21 14:35:51 crc kubenswrapper[4897]: I1121 14:35:51.877042 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d5dac4b-9e6c-41d5-9d10-87015da49a79-scripts\") pod \"aodh-0\" (UID: \"1d5dac4b-9e6c-41d5-9d10-87015da49a79\") " pod="openstack/aodh-0" Nov 21 14:35:51 crc kubenswrapper[4897]: I1121 14:35:51.877529 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d5dac4b-9e6c-41d5-9d10-87015da49a79-public-tls-certs\") pod \"aodh-0\" (UID: \"1d5dac4b-9e6c-41d5-9d10-87015da49a79\") " pod="openstack/aodh-0" Nov 21 14:35:51 crc kubenswrapper[4897]: I1121 14:35:51.877643 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6dmkw\" (UniqueName: \"kubernetes.io/projected/1d5dac4b-9e6c-41d5-9d10-87015da49a79-kube-api-access-6dmkw\") pod \"aodh-0\" (UID: \"1d5dac4b-9e6c-41d5-9d10-87015da49a79\") " pod="openstack/aodh-0" Nov 21 14:35:51 crc kubenswrapper[4897]: I1121 14:35:51.877761 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d5dac4b-9e6c-41d5-9d10-87015da49a79-internal-tls-certs\") pod \"aodh-0\" (UID: \"1d5dac4b-9e6c-41d5-9d10-87015da49a79\") " pod="openstack/aodh-0" Nov 21 14:35:51 crc kubenswrapper[4897]: I1121 14:35:51.877871 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d5dac4b-9e6c-41d5-9d10-87015da49a79-combined-ca-bundle\") pod \"aodh-0\" (UID: \"1d5dac4b-9e6c-41d5-9d10-87015da49a79\") " pod="openstack/aodh-0" Nov 21 14:35:51 crc kubenswrapper[4897]: I1121 14:35:51.877939 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d5dac4b-9e6c-41d5-9d10-87015da49a79-config-data\") pod \"aodh-0\" (UID: \"1d5dac4b-9e6c-41d5-9d10-87015da49a79\") " pod="openstack/aodh-0" Nov 21 14:35:51 crc kubenswrapper[4897]: I1121 14:35:51.884963 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d5dac4b-9e6c-41d5-9d10-87015da49a79-public-tls-certs\") pod \"aodh-0\" (UID: \"1d5dac4b-9e6c-41d5-9d10-87015da49a79\") " pod="openstack/aodh-0" Nov 21 14:35:51 crc kubenswrapper[4897]: I1121 14:35:51.885722 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d5dac4b-9e6c-41d5-9d10-87015da49a79-config-data\") pod \"aodh-0\" (UID: \"1d5dac4b-9e6c-41d5-9d10-87015da49a79\") " pod="openstack/aodh-0" Nov 21 14:35:51 crc kubenswrapper[4897]: I1121 14:35:51.886002 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d5dac4b-9e6c-41d5-9d10-87015da49a79-scripts\") pod \"aodh-0\" (UID: \"1d5dac4b-9e6c-41d5-9d10-87015da49a79\") " pod="openstack/aodh-0" Nov 21 14:35:51 crc kubenswrapper[4897]: I1121 14:35:51.887997 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d5dac4b-9e6c-41d5-9d10-87015da49a79-combined-ca-bundle\") pod \"aodh-0\" (UID: \"1d5dac4b-9e6c-41d5-9d10-87015da49a79\") " pod="openstack/aodh-0" Nov 21 14:35:51 crc kubenswrapper[4897]: I1121 14:35:51.889150 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d5dac4b-9e6c-41d5-9d10-87015da49a79-internal-tls-certs\") pod \"aodh-0\" (UID: \"1d5dac4b-9e6c-41d5-9d10-87015da49a79\") " pod="openstack/aodh-0" Nov 21 14:35:51 crc kubenswrapper[4897]: I1121 14:35:51.900379 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6dmkw\" (UniqueName: \"kubernetes.io/projected/1d5dac4b-9e6c-41d5-9d10-87015da49a79-kube-api-access-6dmkw\") pod \"aodh-0\" (UID: \"1d5dac4b-9e6c-41d5-9d10-87015da49a79\") " pod="openstack/aodh-0" Nov 21 14:35:51 crc kubenswrapper[4897]: I1121 14:35:51.914055 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Nov 21 14:35:52 crc kubenswrapper[4897]: I1121 14:35:52.108221 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="53e2a389-78b0-463a-82df-dc3d1d870b7b" path="/var/lib/kubelet/pods/53e2a389-78b0-463a-82df-dc3d1d870b7b/volumes" Nov 21 14:35:52 crc kubenswrapper[4897]: I1121 14:35:52.110498 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cf6d459c-961c-4361-a034-4a7c7d5ce7b8" path="/var/lib/kubelet/pods/cf6d459c-961c-4361-a034-4a7c7d5ce7b8/volumes" Nov 21 14:35:52 crc kubenswrapper[4897]: I1121 14:35:52.450853 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Nov 21 14:35:52 crc kubenswrapper[4897]: W1121 14:35:52.459491 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1d5dac4b_9e6c_41d5_9d10_87015da49a79.slice/crio-d635a11d4cbbb9a1e3a5b53d856af24ad0a256da54c7d352f9f4735e521b150b WatchSource:0}: Error finding container d635a11d4cbbb9a1e3a5b53d856af24ad0a256da54c7d352f9f4735e521b150b: Status 404 returned error can't find the container with id d635a11d4cbbb9a1e3a5b53d856af24ad0a256da54c7d352f9f4735e521b150b Nov 21 14:35:52 crc kubenswrapper[4897]: I1121 14:35:52.492837 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d3a7078f-ec31-46d5-aadb-b119532328b8","Type":"ContainerStarted","Data":"785d3b3f0d81f47cf04eb76ef2e2d4232e551f90cd5d334cb72df9b89716f361"} Nov 21 14:35:52 crc kubenswrapper[4897]: I1121 14:35:52.492880 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d3a7078f-ec31-46d5-aadb-b119532328b8","Type":"ContainerStarted","Data":"0dceb413fddf48e1b6387a57eb97ad30211325e57d6120065cb6b2ff4baee214"} Nov 21 14:35:52 crc kubenswrapper[4897]: I1121 14:35:52.492892 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d3a7078f-ec31-46d5-aadb-b119532328b8","Type":"ContainerStarted","Data":"21066b7db277e30b514dd6581465dea216ce805e03982ae699fafa33de8c6c5f"} Nov 21 14:35:52 crc kubenswrapper[4897]: I1121 14:35:52.494740 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"1d5dac4b-9e6c-41d5-9d10-87015da49a79","Type":"ContainerStarted","Data":"d635a11d4cbbb9a1e3a5b53d856af24ad0a256da54c7d352f9f4735e521b150b"} Nov 21 14:35:52 crc kubenswrapper[4897]: I1121 14:35:52.516951 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.516927758 podStartE2EDuration="2.516927758s" podCreationTimestamp="2025-11-21 14:35:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:35:52.508520391 +0000 UTC m=+1629.793113876" watchObservedRunningTime="2025-11-21 14:35:52.516927758 +0000 UTC m=+1629.801521243" Nov 21 14:35:53 crc kubenswrapper[4897]: I1121 14:35:53.086928 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 21 14:35:53 crc kubenswrapper[4897]: I1121 14:35:53.086987 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 21 14:35:53 crc kubenswrapper[4897]: I1121 14:35:53.510496 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"1d5dac4b-9e6c-41d5-9d10-87015da49a79","Type":"ContainerStarted","Data":"e65e756fbb3117c206622407abfc9990d581b756bb24310f36d0c24d49164dc0"} Nov 21 14:35:54 crc kubenswrapper[4897]: I1121 14:35:54.275313 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 21 14:35:54 crc kubenswrapper[4897]: I1121 14:35:54.521995 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"1d5dac4b-9e6c-41d5-9d10-87015da49a79","Type":"ContainerStarted","Data":"0b3e7824d93c1825d32ebe2895717aed9691091cdb0492f9fc2fed0b92d32dd9"} Nov 21 14:35:54 crc kubenswrapper[4897]: I1121 14:35:54.875659 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 21 14:35:55 crc kubenswrapper[4897]: I1121 14:35:55.109759 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-dtwql" Nov 21 14:35:55 crc kubenswrapper[4897]: I1121 14:35:55.110068 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-dtwql" Nov 21 14:35:55 crc kubenswrapper[4897]: I1121 14:35:55.535490 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"1d5dac4b-9e6c-41d5-9d10-87015da49a79","Type":"ContainerStarted","Data":"88984a2b2c0a657d25b890e97a861f232ed8a051dddc4d79a706956451832fce"} Nov 21 14:35:56 crc kubenswrapper[4897]: I1121 14:35:56.163335 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-dtwql" podUID="6d3cad20-fb17-4726-b8fa-9fd5303b5eea" containerName="registry-server" probeResult="failure" output=< Nov 21 14:35:56 crc kubenswrapper[4897]: timeout: failed to connect service ":50051" within 1s Nov 21 14:35:56 crc kubenswrapper[4897]: > Nov 21 14:35:57 crc kubenswrapper[4897]: I1121 14:35:57.089954 4897 scope.go:117] "RemoveContainer" containerID="c68f1b003d12b4a2964169a85a04e3f0272d86c891f52e6bebd52f3781520d1a" Nov 21 14:35:57 crc kubenswrapper[4897]: E1121 14:35:57.091139 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 14:35:57 crc kubenswrapper[4897]: I1121 14:35:57.564778 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"1d5dac4b-9e6c-41d5-9d10-87015da49a79","Type":"ContainerStarted","Data":"715e768e1b2e43e144c6d93c0e95b4d92949f9b90ed622f285ee9f286c09f298"} Nov 21 14:35:57 crc kubenswrapper[4897]: I1121 14:35:57.608664 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=2.040856277 podStartE2EDuration="6.600126016s" podCreationTimestamp="2025-11-21 14:35:51 +0000 UTC" firstStartedPulling="2025-11-21 14:35:52.462022888 +0000 UTC m=+1629.746616373" lastFinishedPulling="2025-11-21 14:35:57.021292637 +0000 UTC m=+1634.305886112" observedRunningTime="2025-11-21 14:35:57.588071151 +0000 UTC m=+1634.872664626" watchObservedRunningTime="2025-11-21 14:35:57.600126016 +0000 UTC m=+1634.884719511" Nov 21 14:35:58 crc kubenswrapper[4897]: I1121 14:35:58.086125 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 21 14:35:58 crc kubenswrapper[4897]: I1121 14:35:58.086454 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 21 14:35:58 crc kubenswrapper[4897]: I1121 14:35:58.577860 4897 generic.go:334] "Generic (PLEG): container finished" podID="a558cde2-95bc-4f8d-9daf-90fe592ab9ff" containerID="b85b61c6134535cb1d4f859af67cc0a923a31c62fbd636fe98341334044c0bd0" exitCode=0 Nov 21 14:35:58 crc kubenswrapper[4897]: I1121 14:35:58.577947 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-bpgdn" event={"ID":"a558cde2-95bc-4f8d-9daf-90fe592ab9ff","Type":"ContainerDied","Data":"b85b61c6134535cb1d4f859af67cc0a923a31c62fbd636fe98341334044c0bd0"} Nov 21 14:35:59 crc kubenswrapper[4897]: I1121 14:35:59.103712 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="6810bdc5-077d-44fc-a616-3582d0929d42" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.247:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 21 14:35:59 crc kubenswrapper[4897]: I1121 14:35:59.103769 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="6810bdc5-077d-44fc-a616-3582d0929d42" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.247:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 21 14:35:59 crc kubenswrapper[4897]: I1121 14:35:59.875790 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 21 14:35:59 crc kubenswrapper[4897]: I1121 14:35:59.949984 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 21 14:36:00 crc kubenswrapper[4897]: I1121 14:36:00.154172 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-bpgdn" Nov 21 14:36:00 crc kubenswrapper[4897]: I1121 14:36:00.291025 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-httgf\" (UniqueName: \"kubernetes.io/projected/a558cde2-95bc-4f8d-9daf-90fe592ab9ff-kube-api-access-httgf\") pod \"a558cde2-95bc-4f8d-9daf-90fe592ab9ff\" (UID: \"a558cde2-95bc-4f8d-9daf-90fe592ab9ff\") " Nov 21 14:36:00 crc kubenswrapper[4897]: I1121 14:36:00.291086 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a558cde2-95bc-4f8d-9daf-90fe592ab9ff-combined-ca-bundle\") pod \"a558cde2-95bc-4f8d-9daf-90fe592ab9ff\" (UID: \"a558cde2-95bc-4f8d-9daf-90fe592ab9ff\") " Nov 21 14:36:00 crc kubenswrapper[4897]: I1121 14:36:00.291179 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a558cde2-95bc-4f8d-9daf-90fe592ab9ff-config-data\") pod \"a558cde2-95bc-4f8d-9daf-90fe592ab9ff\" (UID: \"a558cde2-95bc-4f8d-9daf-90fe592ab9ff\") " Nov 21 14:36:00 crc kubenswrapper[4897]: I1121 14:36:00.291254 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a558cde2-95bc-4f8d-9daf-90fe592ab9ff-scripts\") pod \"a558cde2-95bc-4f8d-9daf-90fe592ab9ff\" (UID: \"a558cde2-95bc-4f8d-9daf-90fe592ab9ff\") " Nov 21 14:36:00 crc kubenswrapper[4897]: I1121 14:36:00.297721 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a558cde2-95bc-4f8d-9daf-90fe592ab9ff-kube-api-access-httgf" (OuterVolumeSpecName: "kube-api-access-httgf") pod "a558cde2-95bc-4f8d-9daf-90fe592ab9ff" (UID: "a558cde2-95bc-4f8d-9daf-90fe592ab9ff"). InnerVolumeSpecName "kube-api-access-httgf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:36:00 crc kubenswrapper[4897]: I1121 14:36:00.300458 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a558cde2-95bc-4f8d-9daf-90fe592ab9ff-scripts" (OuterVolumeSpecName: "scripts") pod "a558cde2-95bc-4f8d-9daf-90fe592ab9ff" (UID: "a558cde2-95bc-4f8d-9daf-90fe592ab9ff"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:36:00 crc kubenswrapper[4897]: I1121 14:36:00.323803 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a558cde2-95bc-4f8d-9daf-90fe592ab9ff-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a558cde2-95bc-4f8d-9daf-90fe592ab9ff" (UID: "a558cde2-95bc-4f8d-9daf-90fe592ab9ff"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:36:00 crc kubenswrapper[4897]: I1121 14:36:00.337088 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a558cde2-95bc-4f8d-9daf-90fe592ab9ff-config-data" (OuterVolumeSpecName: "config-data") pod "a558cde2-95bc-4f8d-9daf-90fe592ab9ff" (UID: "a558cde2-95bc-4f8d-9daf-90fe592ab9ff"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:36:00 crc kubenswrapper[4897]: I1121 14:36:00.394772 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-httgf\" (UniqueName: \"kubernetes.io/projected/a558cde2-95bc-4f8d-9daf-90fe592ab9ff-kube-api-access-httgf\") on node \"crc\" DevicePath \"\"" Nov 21 14:36:00 crc kubenswrapper[4897]: I1121 14:36:00.394813 4897 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a558cde2-95bc-4f8d-9daf-90fe592ab9ff-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 21 14:36:00 crc kubenswrapper[4897]: I1121 14:36:00.394824 4897 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a558cde2-95bc-4f8d-9daf-90fe592ab9ff-config-data\") on node \"crc\" DevicePath \"\"" Nov 21 14:36:00 crc kubenswrapper[4897]: I1121 14:36:00.394833 4897 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a558cde2-95bc-4f8d-9daf-90fe592ab9ff-scripts\") on node \"crc\" DevicePath \"\"" Nov 21 14:36:00 crc kubenswrapper[4897]: I1121 14:36:00.608193 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-bpgdn" Nov 21 14:36:00 crc kubenswrapper[4897]: I1121 14:36:00.608545 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-bpgdn" event={"ID":"a558cde2-95bc-4f8d-9daf-90fe592ab9ff","Type":"ContainerDied","Data":"12147380ef8f3214b00a59e3095d82485dd0afadb2a3326a37e477478a43c9c3"} Nov 21 14:36:00 crc kubenswrapper[4897]: I1121 14:36:00.608582 4897 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="12147380ef8f3214b00a59e3095d82485dd0afadb2a3326a37e477478a43c9c3" Nov 21 14:36:00 crc kubenswrapper[4897]: I1121 14:36:00.656861 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 21 14:36:00 crc kubenswrapper[4897]: I1121 14:36:00.687064 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 21 14:36:00 crc kubenswrapper[4897]: E1121 14:36:00.687722 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a558cde2-95bc-4f8d-9daf-90fe592ab9ff" containerName="nova-cell1-conductor-db-sync" Nov 21 14:36:00 crc kubenswrapper[4897]: I1121 14:36:00.687747 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="a558cde2-95bc-4f8d-9daf-90fe592ab9ff" containerName="nova-cell1-conductor-db-sync" Nov 21 14:36:00 crc kubenswrapper[4897]: I1121 14:36:00.688107 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="a558cde2-95bc-4f8d-9daf-90fe592ab9ff" containerName="nova-cell1-conductor-db-sync" Nov 21 14:36:00 crc kubenswrapper[4897]: I1121 14:36:00.689247 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 21 14:36:00 crc kubenswrapper[4897]: I1121 14:36:00.691293 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Nov 21 14:36:00 crc kubenswrapper[4897]: I1121 14:36:00.722673 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 21 14:36:00 crc kubenswrapper[4897]: I1121 14:36:00.802367 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e987828-bd16-47fd-8b58-02271724fa07-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"2e987828-bd16-47fd-8b58-02271724fa07\") " pod="openstack/nova-cell1-conductor-0" Nov 21 14:36:00 crc kubenswrapper[4897]: I1121 14:36:00.802457 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e987828-bd16-47fd-8b58-02271724fa07-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"2e987828-bd16-47fd-8b58-02271724fa07\") " pod="openstack/nova-cell1-conductor-0" Nov 21 14:36:00 crc kubenswrapper[4897]: I1121 14:36:00.802853 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xnvj6\" (UniqueName: \"kubernetes.io/projected/2e987828-bd16-47fd-8b58-02271724fa07-kube-api-access-xnvj6\") pod \"nova-cell1-conductor-0\" (UID: \"2e987828-bd16-47fd-8b58-02271724fa07\") " pod="openstack/nova-cell1-conductor-0" Nov 21 14:36:00 crc kubenswrapper[4897]: E1121 14:36:00.834406 4897 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podafa69523_63c4_4a94_b94a_61c1343ef10c.slice/crio-conmon-6cb3731d3191a0096a5051532c43fd96973f7663ef9d3571618a165b3dd1afca.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda558cde2_95bc_4f8d_9daf_90fe592ab9ff.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda558cde2_95bc_4f8d_9daf_90fe592ab9ff.slice/crio-12147380ef8f3214b00a59e3095d82485dd0afadb2a3326a37e477478a43c9c3\": RecentStats: unable to find data in memory cache]" Nov 21 14:36:00 crc kubenswrapper[4897]: I1121 14:36:00.905239 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e987828-bd16-47fd-8b58-02271724fa07-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"2e987828-bd16-47fd-8b58-02271724fa07\") " pod="openstack/nova-cell1-conductor-0" Nov 21 14:36:00 crc kubenswrapper[4897]: I1121 14:36:00.905308 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e987828-bd16-47fd-8b58-02271724fa07-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"2e987828-bd16-47fd-8b58-02271724fa07\") " pod="openstack/nova-cell1-conductor-0" Nov 21 14:36:00 crc kubenswrapper[4897]: I1121 14:36:00.905408 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xnvj6\" (UniqueName: \"kubernetes.io/projected/2e987828-bd16-47fd-8b58-02271724fa07-kube-api-access-xnvj6\") pod \"nova-cell1-conductor-0\" (UID: \"2e987828-bd16-47fd-8b58-02271724fa07\") " pod="openstack/nova-cell1-conductor-0" Nov 21 14:36:00 crc kubenswrapper[4897]: I1121 14:36:00.911804 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e987828-bd16-47fd-8b58-02271724fa07-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"2e987828-bd16-47fd-8b58-02271724fa07\") " pod="openstack/nova-cell1-conductor-0" Nov 21 14:36:00 crc kubenswrapper[4897]: I1121 14:36:00.912383 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e987828-bd16-47fd-8b58-02271724fa07-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"2e987828-bd16-47fd-8b58-02271724fa07\") " pod="openstack/nova-cell1-conductor-0" Nov 21 14:36:00 crc kubenswrapper[4897]: I1121 14:36:00.931190 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xnvj6\" (UniqueName: \"kubernetes.io/projected/2e987828-bd16-47fd-8b58-02271724fa07-kube-api-access-xnvj6\") pod \"nova-cell1-conductor-0\" (UID: \"2e987828-bd16-47fd-8b58-02271724fa07\") " pod="openstack/nova-cell1-conductor-0" Nov 21 14:36:01 crc kubenswrapper[4897]: I1121 14:36:01.013730 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 21 14:36:01 crc kubenswrapper[4897]: I1121 14:36:01.166251 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 21 14:36:01 crc kubenswrapper[4897]: I1121 14:36:01.167057 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 21 14:36:01 crc kubenswrapper[4897]: I1121 14:36:01.523448 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 21 14:36:01 crc kubenswrapper[4897]: I1121 14:36:01.637601 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"2e987828-bd16-47fd-8b58-02271724fa07","Type":"ContainerStarted","Data":"6c098de17397dca990f076c434a443b3b3c3d545a859a484aaf7de283066cec0"} Nov 21 14:36:02 crc kubenswrapper[4897]: I1121 14:36:02.013123 4897 scope.go:117] "RemoveContainer" containerID="1f2d47fd9bfccad73e9b004362d0525c4684960507722d851013a098334c21fd" Nov 21 14:36:02 crc kubenswrapper[4897]: I1121 14:36:02.051969 4897 scope.go:117] "RemoveContainer" containerID="f19dbbb5857a286582bed5bb5631a58480e595744701974407d574911f1accf4" Nov 21 14:36:02 crc kubenswrapper[4897]: I1121 14:36:02.109265 4897 scope.go:117] "RemoveContainer" containerID="45c47b725cdc3b5aa9ec4a77953186f3bd2971e1387a0ceaa67234ada9133531" Nov 21 14:36:02 crc kubenswrapper[4897]: I1121 14:36:02.157201 4897 scope.go:117] "RemoveContainer" containerID="3e1bff1bbd0769c15d6aa8951ca8e1ab60f58ea1d63e68a36263aec05a1a078e" Nov 21 14:36:02 crc kubenswrapper[4897]: I1121 14:36:02.249794 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="d3a7078f-ec31-46d5-aadb-b119532328b8" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.249:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 21 14:36:02 crc kubenswrapper[4897]: I1121 14:36:02.250310 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="d3a7078f-ec31-46d5-aadb-b119532328b8" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.249:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 21 14:36:02 crc kubenswrapper[4897]: I1121 14:36:02.658874 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"2e987828-bd16-47fd-8b58-02271724fa07","Type":"ContainerStarted","Data":"b1d21689d218347a4faf5f59bcf92fce8a82b0eeb33f90c40242d7311cb08f48"} Nov 21 14:36:02 crc kubenswrapper[4897]: I1121 14:36:02.660704 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Nov 21 14:36:02 crc kubenswrapper[4897]: I1121 14:36:02.680307 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.680281413 podStartE2EDuration="2.680281413s" podCreationTimestamp="2025-11-21 14:36:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:36:02.673018298 +0000 UTC m=+1639.957611783" watchObservedRunningTime="2025-11-21 14:36:02.680281413 +0000 UTC m=+1639.964874888" Nov 21 14:36:04 crc kubenswrapper[4897]: E1121 14:36:04.468140 4897 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podafa69523_63c4_4a94_b94a_61c1343ef10c.slice/crio-conmon-6cb3731d3191a0096a5051532c43fd96973f7663ef9d3571618a165b3dd1afca.scope\": RecentStats: unable to find data in memory cache]" Nov 21 14:36:06 crc kubenswrapper[4897]: I1121 14:36:06.044177 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Nov 21 14:36:06 crc kubenswrapper[4897]: I1121 14:36:06.161401 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-dtwql" podUID="6d3cad20-fb17-4726-b8fa-9fd5303b5eea" containerName="registry-server" probeResult="failure" output=< Nov 21 14:36:06 crc kubenswrapper[4897]: timeout: failed to connect service ":50051" within 1s Nov 21 14:36:06 crc kubenswrapper[4897]: > Nov 21 14:36:08 crc kubenswrapper[4897]: I1121 14:36:08.090081 4897 scope.go:117] "RemoveContainer" containerID="c68f1b003d12b4a2964169a85a04e3f0272d86c891f52e6bebd52f3781520d1a" Nov 21 14:36:08 crc kubenswrapper[4897]: E1121 14:36:08.090814 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 14:36:08 crc kubenswrapper[4897]: I1121 14:36:08.117367 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 21 14:36:08 crc kubenswrapper[4897]: I1121 14:36:08.117546 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 21 14:36:08 crc kubenswrapper[4897]: I1121 14:36:08.128237 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 21 14:36:08 crc kubenswrapper[4897]: I1121 14:36:08.129017 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 21 14:36:11 crc kubenswrapper[4897]: E1121 14:36:11.166672 4897 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podafa69523_63c4_4a94_b94a_61c1343ef10c.slice/crio-conmon-6cb3731d3191a0096a5051532c43fd96973f7663ef9d3571618a165b3dd1afca.scope\": RecentStats: unable to find data in memory cache]" Nov 21 14:36:11 crc kubenswrapper[4897]: I1121 14:36:11.168361 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 21 14:36:11 crc kubenswrapper[4897]: I1121 14:36:11.168862 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 21 14:36:11 crc kubenswrapper[4897]: I1121 14:36:11.169179 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 21 14:36:11 crc kubenswrapper[4897]: I1121 14:36:11.171447 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 21 14:36:11 crc kubenswrapper[4897]: I1121 14:36:11.699831 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 21 14:36:11 crc kubenswrapper[4897]: I1121 14:36:11.754324 4897 generic.go:334] "Generic (PLEG): container finished" podID="fe6f78aa-838c-4bf1-8a19-9726831f2941" containerID="9563192811a74d43bb16a893b77b8cc16dbe59707cfe89bae3411c69591d5bed" exitCode=137 Nov 21 14:36:11 crc kubenswrapper[4897]: I1121 14:36:11.754619 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 21 14:36:11 crc kubenswrapper[4897]: I1121 14:36:11.755713 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"fe6f78aa-838c-4bf1-8a19-9726831f2941","Type":"ContainerDied","Data":"9563192811a74d43bb16a893b77b8cc16dbe59707cfe89bae3411c69591d5bed"} Nov 21 14:36:11 crc kubenswrapper[4897]: I1121 14:36:11.755762 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 21 14:36:11 crc kubenswrapper[4897]: I1121 14:36:11.755782 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"fe6f78aa-838c-4bf1-8a19-9726831f2941","Type":"ContainerDied","Data":"a3376a83cdcbd3e57aaa30bc8717803a8b3e1b286724380cc1b4bcf035f668a6"} Nov 21 14:36:11 crc kubenswrapper[4897]: I1121 14:36:11.755797 4897 scope.go:117] "RemoveContainer" containerID="9563192811a74d43bb16a893b77b8cc16dbe59707cfe89bae3411c69591d5bed" Nov 21 14:36:11 crc kubenswrapper[4897]: I1121 14:36:11.759813 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 21 14:36:11 crc kubenswrapper[4897]: I1121 14:36:11.794498 4897 scope.go:117] "RemoveContainer" containerID="9563192811a74d43bb16a893b77b8cc16dbe59707cfe89bae3411c69591d5bed" Nov 21 14:36:11 crc kubenswrapper[4897]: E1121 14:36:11.795191 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9563192811a74d43bb16a893b77b8cc16dbe59707cfe89bae3411c69591d5bed\": container with ID starting with 9563192811a74d43bb16a893b77b8cc16dbe59707cfe89bae3411c69591d5bed not found: ID does not exist" containerID="9563192811a74d43bb16a893b77b8cc16dbe59707cfe89bae3411c69591d5bed" Nov 21 14:36:11 crc kubenswrapper[4897]: I1121 14:36:11.795258 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9563192811a74d43bb16a893b77b8cc16dbe59707cfe89bae3411c69591d5bed"} err="failed to get container status \"9563192811a74d43bb16a893b77b8cc16dbe59707cfe89bae3411c69591d5bed\": rpc error: code = NotFound desc = could not find container \"9563192811a74d43bb16a893b77b8cc16dbe59707cfe89bae3411c69591d5bed\": container with ID starting with 9563192811a74d43bb16a893b77b8cc16dbe59707cfe89bae3411c69591d5bed not found: ID does not exist" Nov 21 14:36:11 crc kubenswrapper[4897]: I1121 14:36:11.799713 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe6f78aa-838c-4bf1-8a19-9726831f2941-combined-ca-bundle\") pod \"fe6f78aa-838c-4bf1-8a19-9726831f2941\" (UID: \"fe6f78aa-838c-4bf1-8a19-9726831f2941\") " Nov 21 14:36:11 crc kubenswrapper[4897]: I1121 14:36:11.800034 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe6f78aa-838c-4bf1-8a19-9726831f2941-config-data\") pod \"fe6f78aa-838c-4bf1-8a19-9726831f2941\" (UID: \"fe6f78aa-838c-4bf1-8a19-9726831f2941\") " Nov 21 14:36:11 crc kubenswrapper[4897]: I1121 14:36:11.800178 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8dfkf\" (UniqueName: \"kubernetes.io/projected/fe6f78aa-838c-4bf1-8a19-9726831f2941-kube-api-access-8dfkf\") pod \"fe6f78aa-838c-4bf1-8a19-9726831f2941\" (UID: \"fe6f78aa-838c-4bf1-8a19-9726831f2941\") " Nov 21 14:36:11 crc kubenswrapper[4897]: I1121 14:36:11.814138 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fe6f78aa-838c-4bf1-8a19-9726831f2941-kube-api-access-8dfkf" (OuterVolumeSpecName: "kube-api-access-8dfkf") pod "fe6f78aa-838c-4bf1-8a19-9726831f2941" (UID: "fe6f78aa-838c-4bf1-8a19-9726831f2941"). InnerVolumeSpecName "kube-api-access-8dfkf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:36:11 crc kubenswrapper[4897]: I1121 14:36:11.860587 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe6f78aa-838c-4bf1-8a19-9726831f2941-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fe6f78aa-838c-4bf1-8a19-9726831f2941" (UID: "fe6f78aa-838c-4bf1-8a19-9726831f2941"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:36:11 crc kubenswrapper[4897]: I1121 14:36:11.878926 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe6f78aa-838c-4bf1-8a19-9726831f2941-config-data" (OuterVolumeSpecName: "config-data") pod "fe6f78aa-838c-4bf1-8a19-9726831f2941" (UID: "fe6f78aa-838c-4bf1-8a19-9726831f2941"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:36:11 crc kubenswrapper[4897]: I1121 14:36:11.903184 4897 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe6f78aa-838c-4bf1-8a19-9726831f2941-config-data\") on node \"crc\" DevicePath \"\"" Nov 21 14:36:11 crc kubenswrapper[4897]: I1121 14:36:11.903219 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8dfkf\" (UniqueName: \"kubernetes.io/projected/fe6f78aa-838c-4bf1-8a19-9726831f2941-kube-api-access-8dfkf\") on node \"crc\" DevicePath \"\"" Nov 21 14:36:11 crc kubenswrapper[4897]: I1121 14:36:11.903229 4897 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe6f78aa-838c-4bf1-8a19-9726831f2941-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 21 14:36:11 crc kubenswrapper[4897]: I1121 14:36:11.950746 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6d99f6bc7f-k7pkd"] Nov 21 14:36:11 crc kubenswrapper[4897]: E1121 14:36:11.951279 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe6f78aa-838c-4bf1-8a19-9726831f2941" containerName="nova-cell1-novncproxy-novncproxy" Nov 21 14:36:11 crc kubenswrapper[4897]: I1121 14:36:11.951291 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe6f78aa-838c-4bf1-8a19-9726831f2941" containerName="nova-cell1-novncproxy-novncproxy" Nov 21 14:36:11 crc kubenswrapper[4897]: I1121 14:36:11.951480 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe6f78aa-838c-4bf1-8a19-9726831f2941" containerName="nova-cell1-novncproxy-novncproxy" Nov 21 14:36:11 crc kubenswrapper[4897]: I1121 14:36:11.953888 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d99f6bc7f-k7pkd" Nov 21 14:36:11 crc kubenswrapper[4897]: I1121 14:36:11.972206 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d99f6bc7f-k7pkd"] Nov 21 14:36:12 crc kubenswrapper[4897]: I1121 14:36:12.111927 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/139262e5-8014-495f-b877-8a355481b63e-ovsdbserver-sb\") pod \"dnsmasq-dns-6d99f6bc7f-k7pkd\" (UID: \"139262e5-8014-495f-b877-8a355481b63e\") " pod="openstack/dnsmasq-dns-6d99f6bc7f-k7pkd" Nov 21 14:36:12 crc kubenswrapper[4897]: I1121 14:36:12.112181 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/139262e5-8014-495f-b877-8a355481b63e-config\") pod \"dnsmasq-dns-6d99f6bc7f-k7pkd\" (UID: \"139262e5-8014-495f-b877-8a355481b63e\") " pod="openstack/dnsmasq-dns-6d99f6bc7f-k7pkd" Nov 21 14:36:12 crc kubenswrapper[4897]: I1121 14:36:12.112267 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hj92f\" (UniqueName: \"kubernetes.io/projected/139262e5-8014-495f-b877-8a355481b63e-kube-api-access-hj92f\") pod \"dnsmasq-dns-6d99f6bc7f-k7pkd\" (UID: \"139262e5-8014-495f-b877-8a355481b63e\") " pod="openstack/dnsmasq-dns-6d99f6bc7f-k7pkd" Nov 21 14:36:12 crc kubenswrapper[4897]: I1121 14:36:12.112298 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/139262e5-8014-495f-b877-8a355481b63e-dns-svc\") pod \"dnsmasq-dns-6d99f6bc7f-k7pkd\" (UID: \"139262e5-8014-495f-b877-8a355481b63e\") " pod="openstack/dnsmasq-dns-6d99f6bc7f-k7pkd" Nov 21 14:36:12 crc kubenswrapper[4897]: I1121 14:36:12.112336 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/139262e5-8014-495f-b877-8a355481b63e-ovsdbserver-nb\") pod \"dnsmasq-dns-6d99f6bc7f-k7pkd\" (UID: \"139262e5-8014-495f-b877-8a355481b63e\") " pod="openstack/dnsmasq-dns-6d99f6bc7f-k7pkd" Nov 21 14:36:12 crc kubenswrapper[4897]: I1121 14:36:12.112372 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/139262e5-8014-495f-b877-8a355481b63e-dns-swift-storage-0\") pod \"dnsmasq-dns-6d99f6bc7f-k7pkd\" (UID: \"139262e5-8014-495f-b877-8a355481b63e\") " pod="openstack/dnsmasq-dns-6d99f6bc7f-k7pkd" Nov 21 14:36:12 crc kubenswrapper[4897]: I1121 14:36:12.117580 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 21 14:36:12 crc kubenswrapper[4897]: I1121 14:36:12.126732 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 21 14:36:12 crc kubenswrapper[4897]: I1121 14:36:12.137881 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 21 14:36:12 crc kubenswrapper[4897]: I1121 14:36:12.139489 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 21 14:36:12 crc kubenswrapper[4897]: I1121 14:36:12.144499 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Nov 21 14:36:12 crc kubenswrapper[4897]: I1121 14:36:12.145056 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Nov 21 14:36:12 crc kubenswrapper[4897]: I1121 14:36:12.145452 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Nov 21 14:36:12 crc kubenswrapper[4897]: I1121 14:36:12.174177 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 21 14:36:12 crc kubenswrapper[4897]: I1121 14:36:12.215021 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9kzdg\" (UniqueName: \"kubernetes.io/projected/4c89616e-027a-47ec-a845-a5e2a519156d-kube-api-access-9kzdg\") pod \"nova-cell1-novncproxy-0\" (UID: \"4c89616e-027a-47ec-a845-a5e2a519156d\") " pod="openstack/nova-cell1-novncproxy-0" Nov 21 14:36:12 crc kubenswrapper[4897]: I1121 14:36:12.215134 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/139262e5-8014-495f-b877-8a355481b63e-config\") pod \"dnsmasq-dns-6d99f6bc7f-k7pkd\" (UID: \"139262e5-8014-495f-b877-8a355481b63e\") " pod="openstack/dnsmasq-dns-6d99f6bc7f-k7pkd" Nov 21 14:36:12 crc kubenswrapper[4897]: I1121 14:36:12.215235 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hj92f\" (UniqueName: \"kubernetes.io/projected/139262e5-8014-495f-b877-8a355481b63e-kube-api-access-hj92f\") pod \"dnsmasq-dns-6d99f6bc7f-k7pkd\" (UID: \"139262e5-8014-495f-b877-8a355481b63e\") " pod="openstack/dnsmasq-dns-6d99f6bc7f-k7pkd" Nov 21 14:36:12 crc kubenswrapper[4897]: I1121 14:36:12.215259 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/139262e5-8014-495f-b877-8a355481b63e-dns-svc\") pod \"dnsmasq-dns-6d99f6bc7f-k7pkd\" (UID: \"139262e5-8014-495f-b877-8a355481b63e\") " pod="openstack/dnsmasq-dns-6d99f6bc7f-k7pkd" Nov 21 14:36:12 crc kubenswrapper[4897]: I1121 14:36:12.215290 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/139262e5-8014-495f-b877-8a355481b63e-ovsdbserver-nb\") pod \"dnsmasq-dns-6d99f6bc7f-k7pkd\" (UID: \"139262e5-8014-495f-b877-8a355481b63e\") " pod="openstack/dnsmasq-dns-6d99f6bc7f-k7pkd" Nov 21 14:36:12 crc kubenswrapper[4897]: I1121 14:36:12.215315 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/139262e5-8014-495f-b877-8a355481b63e-dns-swift-storage-0\") pod \"dnsmasq-dns-6d99f6bc7f-k7pkd\" (UID: \"139262e5-8014-495f-b877-8a355481b63e\") " pod="openstack/dnsmasq-dns-6d99f6bc7f-k7pkd" Nov 21 14:36:12 crc kubenswrapper[4897]: I1121 14:36:12.215411 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/4c89616e-027a-47ec-a845-a5e2a519156d-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"4c89616e-027a-47ec-a845-a5e2a519156d\") " pod="openstack/nova-cell1-novncproxy-0" Nov 21 14:36:12 crc kubenswrapper[4897]: I1121 14:36:12.215451 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c89616e-027a-47ec-a845-a5e2a519156d-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"4c89616e-027a-47ec-a845-a5e2a519156d\") " pod="openstack/nova-cell1-novncproxy-0" Nov 21 14:36:12 crc kubenswrapper[4897]: I1121 14:36:12.215476 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/4c89616e-027a-47ec-a845-a5e2a519156d-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"4c89616e-027a-47ec-a845-a5e2a519156d\") " pod="openstack/nova-cell1-novncproxy-0" Nov 21 14:36:12 crc kubenswrapper[4897]: I1121 14:36:12.215607 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c89616e-027a-47ec-a845-a5e2a519156d-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"4c89616e-027a-47ec-a845-a5e2a519156d\") " pod="openstack/nova-cell1-novncproxy-0" Nov 21 14:36:12 crc kubenswrapper[4897]: I1121 14:36:12.215665 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/139262e5-8014-495f-b877-8a355481b63e-ovsdbserver-sb\") pod \"dnsmasq-dns-6d99f6bc7f-k7pkd\" (UID: \"139262e5-8014-495f-b877-8a355481b63e\") " pod="openstack/dnsmasq-dns-6d99f6bc7f-k7pkd" Nov 21 14:36:12 crc kubenswrapper[4897]: I1121 14:36:12.218143 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/139262e5-8014-495f-b877-8a355481b63e-ovsdbserver-nb\") pod \"dnsmasq-dns-6d99f6bc7f-k7pkd\" (UID: \"139262e5-8014-495f-b877-8a355481b63e\") " pod="openstack/dnsmasq-dns-6d99f6bc7f-k7pkd" Nov 21 14:36:12 crc kubenswrapper[4897]: I1121 14:36:12.224270 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/139262e5-8014-495f-b877-8a355481b63e-config\") pod \"dnsmasq-dns-6d99f6bc7f-k7pkd\" (UID: \"139262e5-8014-495f-b877-8a355481b63e\") " pod="openstack/dnsmasq-dns-6d99f6bc7f-k7pkd" Nov 21 14:36:12 crc kubenswrapper[4897]: I1121 14:36:12.224467 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/139262e5-8014-495f-b877-8a355481b63e-dns-svc\") pod \"dnsmasq-dns-6d99f6bc7f-k7pkd\" (UID: \"139262e5-8014-495f-b877-8a355481b63e\") " pod="openstack/dnsmasq-dns-6d99f6bc7f-k7pkd" Nov 21 14:36:12 crc kubenswrapper[4897]: I1121 14:36:12.224615 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/139262e5-8014-495f-b877-8a355481b63e-ovsdbserver-sb\") pod \"dnsmasq-dns-6d99f6bc7f-k7pkd\" (UID: \"139262e5-8014-495f-b877-8a355481b63e\") " pod="openstack/dnsmasq-dns-6d99f6bc7f-k7pkd" Nov 21 14:36:12 crc kubenswrapper[4897]: I1121 14:36:12.225116 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/139262e5-8014-495f-b877-8a355481b63e-dns-swift-storage-0\") pod \"dnsmasq-dns-6d99f6bc7f-k7pkd\" (UID: \"139262e5-8014-495f-b877-8a355481b63e\") " pod="openstack/dnsmasq-dns-6d99f6bc7f-k7pkd" Nov 21 14:36:12 crc kubenswrapper[4897]: I1121 14:36:12.233560 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hj92f\" (UniqueName: \"kubernetes.io/projected/139262e5-8014-495f-b877-8a355481b63e-kube-api-access-hj92f\") pod \"dnsmasq-dns-6d99f6bc7f-k7pkd\" (UID: \"139262e5-8014-495f-b877-8a355481b63e\") " pod="openstack/dnsmasq-dns-6d99f6bc7f-k7pkd" Nov 21 14:36:12 crc kubenswrapper[4897]: I1121 14:36:12.294938 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d99f6bc7f-k7pkd" Nov 21 14:36:12 crc kubenswrapper[4897]: I1121 14:36:12.319716 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9kzdg\" (UniqueName: \"kubernetes.io/projected/4c89616e-027a-47ec-a845-a5e2a519156d-kube-api-access-9kzdg\") pod \"nova-cell1-novncproxy-0\" (UID: \"4c89616e-027a-47ec-a845-a5e2a519156d\") " pod="openstack/nova-cell1-novncproxy-0" Nov 21 14:36:12 crc kubenswrapper[4897]: I1121 14:36:12.321941 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/4c89616e-027a-47ec-a845-a5e2a519156d-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"4c89616e-027a-47ec-a845-a5e2a519156d\") " pod="openstack/nova-cell1-novncproxy-0" Nov 21 14:36:12 crc kubenswrapper[4897]: I1121 14:36:12.322016 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c89616e-027a-47ec-a845-a5e2a519156d-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"4c89616e-027a-47ec-a845-a5e2a519156d\") " pod="openstack/nova-cell1-novncproxy-0" Nov 21 14:36:12 crc kubenswrapper[4897]: I1121 14:36:12.322042 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/4c89616e-027a-47ec-a845-a5e2a519156d-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"4c89616e-027a-47ec-a845-a5e2a519156d\") " pod="openstack/nova-cell1-novncproxy-0" Nov 21 14:36:12 crc kubenswrapper[4897]: I1121 14:36:12.322154 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c89616e-027a-47ec-a845-a5e2a519156d-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"4c89616e-027a-47ec-a845-a5e2a519156d\") " pod="openstack/nova-cell1-novncproxy-0" Nov 21 14:36:12 crc kubenswrapper[4897]: I1121 14:36:12.326313 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c89616e-027a-47ec-a845-a5e2a519156d-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"4c89616e-027a-47ec-a845-a5e2a519156d\") " pod="openstack/nova-cell1-novncproxy-0" Nov 21 14:36:12 crc kubenswrapper[4897]: I1121 14:36:12.329052 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/4c89616e-027a-47ec-a845-a5e2a519156d-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"4c89616e-027a-47ec-a845-a5e2a519156d\") " pod="openstack/nova-cell1-novncproxy-0" Nov 21 14:36:12 crc kubenswrapper[4897]: I1121 14:36:12.332453 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c89616e-027a-47ec-a845-a5e2a519156d-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"4c89616e-027a-47ec-a845-a5e2a519156d\") " pod="openstack/nova-cell1-novncproxy-0" Nov 21 14:36:12 crc kubenswrapper[4897]: I1121 14:36:12.336080 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/4c89616e-027a-47ec-a845-a5e2a519156d-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"4c89616e-027a-47ec-a845-a5e2a519156d\") " pod="openstack/nova-cell1-novncproxy-0" Nov 21 14:36:12 crc kubenswrapper[4897]: I1121 14:36:12.346465 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9kzdg\" (UniqueName: \"kubernetes.io/projected/4c89616e-027a-47ec-a845-a5e2a519156d-kube-api-access-9kzdg\") pod \"nova-cell1-novncproxy-0\" (UID: \"4c89616e-027a-47ec-a845-a5e2a519156d\") " pod="openstack/nova-cell1-novncproxy-0" Nov 21 14:36:12 crc kubenswrapper[4897]: I1121 14:36:12.488270 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 21 14:36:12 crc kubenswrapper[4897]: I1121 14:36:12.829922 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d99f6bc7f-k7pkd"] Nov 21 14:36:13 crc kubenswrapper[4897]: I1121 14:36:13.063062 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 21 14:36:13 crc kubenswrapper[4897]: I1121 14:36:13.786224 4897 generic.go:334] "Generic (PLEG): container finished" podID="139262e5-8014-495f-b877-8a355481b63e" containerID="3f71d8a78981c2b1871f233e3fb130b2a5ddbb19d9a79b4a171c893b7f7a6432" exitCode=0 Nov 21 14:36:13 crc kubenswrapper[4897]: I1121 14:36:13.786298 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d99f6bc7f-k7pkd" event={"ID":"139262e5-8014-495f-b877-8a355481b63e","Type":"ContainerDied","Data":"3f71d8a78981c2b1871f233e3fb130b2a5ddbb19d9a79b4a171c893b7f7a6432"} Nov 21 14:36:13 crc kubenswrapper[4897]: I1121 14:36:13.786694 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d99f6bc7f-k7pkd" event={"ID":"139262e5-8014-495f-b877-8a355481b63e","Type":"ContainerStarted","Data":"85e7d1770a28110921693cfd5e03b12873fbe5da7589b8d80b2aabb15c6a8239"} Nov 21 14:36:13 crc kubenswrapper[4897]: I1121 14:36:13.788991 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"4c89616e-027a-47ec-a845-a5e2a519156d","Type":"ContainerStarted","Data":"3283038b15af47d77ceaa7d9b5d4c6918ba463cc02c59ca756d8ad16a1feea0c"} Nov 21 14:36:13 crc kubenswrapper[4897]: I1121 14:36:13.789037 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"4c89616e-027a-47ec-a845-a5e2a519156d","Type":"ContainerStarted","Data":"c506e4b32c69aa04a76db26abd4c2723980b15e20ffeebe373168c61d465de55"} Nov 21 14:36:13 crc kubenswrapper[4897]: I1121 14:36:13.873701 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=1.8736671029999998 podStartE2EDuration="1.873667103s" podCreationTimestamp="2025-11-21 14:36:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:36:13.846854531 +0000 UTC m=+1651.131448026" watchObservedRunningTime="2025-11-21 14:36:13.873667103 +0000 UTC m=+1651.158260578" Nov 21 14:36:14 crc kubenswrapper[4897]: I1121 14:36:14.103886 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fe6f78aa-838c-4bf1-8a19-9726831f2941" path="/var/lib/kubelet/pods/fe6f78aa-838c-4bf1-8a19-9726831f2941/volumes" Nov 21 14:36:14 crc kubenswrapper[4897]: I1121 14:36:14.306627 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 21 14:36:14 crc kubenswrapper[4897]: I1121 14:36:14.513207 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 21 14:36:14 crc kubenswrapper[4897]: I1121 14:36:14.520561 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="efed1737-5de8-4313-9127-724bbd239f1a" containerName="ceilometer-central-agent" containerID="cri-o://5c1ffe3f03b119bb9d7d244acd512729212c30aafbda3158b86b3b645cdc77e5" gracePeriod=30 Nov 21 14:36:14 crc kubenswrapper[4897]: I1121 14:36:14.520659 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="efed1737-5de8-4313-9127-724bbd239f1a" containerName="proxy-httpd" containerID="cri-o://76bd36bbbb963c0c157f72905e2682ad3a3471d6d16c1c47c51a3e87bf44837e" gracePeriod=30 Nov 21 14:36:14 crc kubenswrapper[4897]: I1121 14:36:14.520687 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="efed1737-5de8-4313-9127-724bbd239f1a" containerName="ceilometer-notification-agent" containerID="cri-o://37c0701bd55a3a69856270d224fe1f7c29a3befbcc9c45fb1e1596e394a69933" gracePeriod=30 Nov 21 14:36:14 crc kubenswrapper[4897]: I1121 14:36:14.520619 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="efed1737-5de8-4313-9127-724bbd239f1a" containerName="sg-core" containerID="cri-o://aba81a8a1224fdc2f5618aeb6eb3be0ce285ed4b26b0e1d6ceaf5c836c7536a1" gracePeriod=30 Nov 21 14:36:14 crc kubenswrapper[4897]: I1121 14:36:14.802926 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d99f6bc7f-k7pkd" event={"ID":"139262e5-8014-495f-b877-8a355481b63e","Type":"ContainerStarted","Data":"68ad1c7ead606d090872be2914843b6f3b65ad129c1977e39d9cb84fe5b3e99b"} Nov 21 14:36:14 crc kubenswrapper[4897]: I1121 14:36:14.803446 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6d99f6bc7f-k7pkd" Nov 21 14:36:14 crc kubenswrapper[4897]: I1121 14:36:14.806867 4897 generic.go:334] "Generic (PLEG): container finished" podID="efed1737-5de8-4313-9127-724bbd239f1a" containerID="76bd36bbbb963c0c157f72905e2682ad3a3471d6d16c1c47c51a3e87bf44837e" exitCode=0 Nov 21 14:36:14 crc kubenswrapper[4897]: I1121 14:36:14.806902 4897 generic.go:334] "Generic (PLEG): container finished" podID="efed1737-5de8-4313-9127-724bbd239f1a" containerID="aba81a8a1224fdc2f5618aeb6eb3be0ce285ed4b26b0e1d6ceaf5c836c7536a1" exitCode=2 Nov 21 14:36:14 crc kubenswrapper[4897]: I1121 14:36:14.807111 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"efed1737-5de8-4313-9127-724bbd239f1a","Type":"ContainerDied","Data":"76bd36bbbb963c0c157f72905e2682ad3a3471d6d16c1c47c51a3e87bf44837e"} Nov 21 14:36:14 crc kubenswrapper[4897]: I1121 14:36:14.807174 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"efed1737-5de8-4313-9127-724bbd239f1a","Type":"ContainerDied","Data":"aba81a8a1224fdc2f5618aeb6eb3be0ce285ed4b26b0e1d6ceaf5c836c7536a1"} Nov 21 14:36:14 crc kubenswrapper[4897]: I1121 14:36:14.807360 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="d3a7078f-ec31-46d5-aadb-b119532328b8" containerName="nova-api-log" containerID="cri-o://0dceb413fddf48e1b6387a57eb97ad30211325e57d6120065cb6b2ff4baee214" gracePeriod=30 Nov 21 14:36:14 crc kubenswrapper[4897]: I1121 14:36:14.807558 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="d3a7078f-ec31-46d5-aadb-b119532328b8" containerName="nova-api-api" containerID="cri-o://785d3b3f0d81f47cf04eb76ef2e2d4232e551f90cd5d334cb72df9b89716f361" gracePeriod=30 Nov 21 14:36:14 crc kubenswrapper[4897]: I1121 14:36:14.846200 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6d99f6bc7f-k7pkd" podStartSLOduration=3.846176176 podStartE2EDuration="3.846176176s" podCreationTimestamp="2025-11-21 14:36:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:36:14.832197469 +0000 UTC m=+1652.116790944" watchObservedRunningTime="2025-11-21 14:36:14.846176176 +0000 UTC m=+1652.130769651" Nov 21 14:36:15 crc kubenswrapper[4897]: I1121 14:36:15.825898 4897 generic.go:334] "Generic (PLEG): container finished" podID="efed1737-5de8-4313-9127-724bbd239f1a" containerID="37c0701bd55a3a69856270d224fe1f7c29a3befbcc9c45fb1e1596e394a69933" exitCode=0 Nov 21 14:36:15 crc kubenswrapper[4897]: I1121 14:36:15.825953 4897 generic.go:334] "Generic (PLEG): container finished" podID="efed1737-5de8-4313-9127-724bbd239f1a" containerID="5c1ffe3f03b119bb9d7d244acd512729212c30aafbda3158b86b3b645cdc77e5" exitCode=0 Nov 21 14:36:15 crc kubenswrapper[4897]: I1121 14:36:15.825966 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"efed1737-5de8-4313-9127-724bbd239f1a","Type":"ContainerDied","Data":"37c0701bd55a3a69856270d224fe1f7c29a3befbcc9c45fb1e1596e394a69933"} Nov 21 14:36:15 crc kubenswrapper[4897]: I1121 14:36:15.826021 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"efed1737-5de8-4313-9127-724bbd239f1a","Type":"ContainerDied","Data":"5c1ffe3f03b119bb9d7d244acd512729212c30aafbda3158b86b3b645cdc77e5"} Nov 21 14:36:15 crc kubenswrapper[4897]: I1121 14:36:15.829213 4897 generic.go:334] "Generic (PLEG): container finished" podID="d3a7078f-ec31-46d5-aadb-b119532328b8" containerID="0dceb413fddf48e1b6387a57eb97ad30211325e57d6120065cb6b2ff4baee214" exitCode=143 Nov 21 14:36:15 crc kubenswrapper[4897]: I1121 14:36:15.829318 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d3a7078f-ec31-46d5-aadb-b119532328b8","Type":"ContainerDied","Data":"0dceb413fddf48e1b6387a57eb97ad30211325e57d6120065cb6b2ff4baee214"} Nov 21 14:36:16 crc kubenswrapper[4897]: I1121 14:36:16.166190 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-dtwql" podUID="6d3cad20-fb17-4726-b8fa-9fd5303b5eea" containerName="registry-server" probeResult="failure" output=< Nov 21 14:36:16 crc kubenswrapper[4897]: timeout: failed to connect service ":50051" within 1s Nov 21 14:36:16 crc kubenswrapper[4897]: > Nov 21 14:36:16 crc kubenswrapper[4897]: I1121 14:36:16.453650 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 21 14:36:16 crc kubenswrapper[4897]: I1121 14:36:16.527661 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/efed1737-5de8-4313-9127-724bbd239f1a-sg-core-conf-yaml\") pod \"efed1737-5de8-4313-9127-724bbd239f1a\" (UID: \"efed1737-5de8-4313-9127-724bbd239f1a\") " Nov 21 14:36:16 crc kubenswrapper[4897]: I1121 14:36:16.527761 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/efed1737-5de8-4313-9127-724bbd239f1a-scripts\") pod \"efed1737-5de8-4313-9127-724bbd239f1a\" (UID: \"efed1737-5de8-4313-9127-724bbd239f1a\") " Nov 21 14:36:16 crc kubenswrapper[4897]: I1121 14:36:16.527925 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/efed1737-5de8-4313-9127-724bbd239f1a-run-httpd\") pod \"efed1737-5de8-4313-9127-724bbd239f1a\" (UID: \"efed1737-5de8-4313-9127-724bbd239f1a\") " Nov 21 14:36:16 crc kubenswrapper[4897]: I1121 14:36:16.528009 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-86hmv\" (UniqueName: \"kubernetes.io/projected/efed1737-5de8-4313-9127-724bbd239f1a-kube-api-access-86hmv\") pod \"efed1737-5de8-4313-9127-724bbd239f1a\" (UID: \"efed1737-5de8-4313-9127-724bbd239f1a\") " Nov 21 14:36:16 crc kubenswrapper[4897]: I1121 14:36:16.528093 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/efed1737-5de8-4313-9127-724bbd239f1a-config-data\") pod \"efed1737-5de8-4313-9127-724bbd239f1a\" (UID: \"efed1737-5de8-4313-9127-724bbd239f1a\") " Nov 21 14:36:16 crc kubenswrapper[4897]: I1121 14:36:16.528118 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/efed1737-5de8-4313-9127-724bbd239f1a-ceilometer-tls-certs\") pod \"efed1737-5de8-4313-9127-724bbd239f1a\" (UID: \"efed1737-5de8-4313-9127-724bbd239f1a\") " Nov 21 14:36:16 crc kubenswrapper[4897]: I1121 14:36:16.528182 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/efed1737-5de8-4313-9127-724bbd239f1a-combined-ca-bundle\") pod \"efed1737-5de8-4313-9127-724bbd239f1a\" (UID: \"efed1737-5de8-4313-9127-724bbd239f1a\") " Nov 21 14:36:16 crc kubenswrapper[4897]: I1121 14:36:16.528210 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/efed1737-5de8-4313-9127-724bbd239f1a-log-httpd\") pod \"efed1737-5de8-4313-9127-724bbd239f1a\" (UID: \"efed1737-5de8-4313-9127-724bbd239f1a\") " Nov 21 14:36:16 crc kubenswrapper[4897]: I1121 14:36:16.529025 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/efed1737-5de8-4313-9127-724bbd239f1a-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "efed1737-5de8-4313-9127-724bbd239f1a" (UID: "efed1737-5de8-4313-9127-724bbd239f1a"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:36:16 crc kubenswrapper[4897]: I1121 14:36:16.529355 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/efed1737-5de8-4313-9127-724bbd239f1a-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "efed1737-5de8-4313-9127-724bbd239f1a" (UID: "efed1737-5de8-4313-9127-724bbd239f1a"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:36:16 crc kubenswrapper[4897]: I1121 14:36:16.537569 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efed1737-5de8-4313-9127-724bbd239f1a-kube-api-access-86hmv" (OuterVolumeSpecName: "kube-api-access-86hmv") pod "efed1737-5de8-4313-9127-724bbd239f1a" (UID: "efed1737-5de8-4313-9127-724bbd239f1a"). InnerVolumeSpecName "kube-api-access-86hmv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:36:16 crc kubenswrapper[4897]: I1121 14:36:16.538141 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efed1737-5de8-4313-9127-724bbd239f1a-scripts" (OuterVolumeSpecName: "scripts") pod "efed1737-5de8-4313-9127-724bbd239f1a" (UID: "efed1737-5de8-4313-9127-724bbd239f1a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:36:16 crc kubenswrapper[4897]: I1121 14:36:16.588575 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efed1737-5de8-4313-9127-724bbd239f1a-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "efed1737-5de8-4313-9127-724bbd239f1a" (UID: "efed1737-5de8-4313-9127-724bbd239f1a"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:36:16 crc kubenswrapper[4897]: I1121 14:36:16.622354 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efed1737-5de8-4313-9127-724bbd239f1a-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "efed1737-5de8-4313-9127-724bbd239f1a" (UID: "efed1737-5de8-4313-9127-724bbd239f1a"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:36:16 crc kubenswrapper[4897]: I1121 14:36:16.631613 4897 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/efed1737-5de8-4313-9127-724bbd239f1a-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 21 14:36:16 crc kubenswrapper[4897]: I1121 14:36:16.631646 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-86hmv\" (UniqueName: \"kubernetes.io/projected/efed1737-5de8-4313-9127-724bbd239f1a-kube-api-access-86hmv\") on node \"crc\" DevicePath \"\"" Nov 21 14:36:16 crc kubenswrapper[4897]: I1121 14:36:16.631658 4897 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/efed1737-5de8-4313-9127-724bbd239f1a-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 21 14:36:16 crc kubenswrapper[4897]: I1121 14:36:16.631667 4897 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/efed1737-5de8-4313-9127-724bbd239f1a-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 21 14:36:16 crc kubenswrapper[4897]: I1121 14:36:16.631676 4897 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/efed1737-5de8-4313-9127-724bbd239f1a-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 21 14:36:16 crc kubenswrapper[4897]: I1121 14:36:16.631683 4897 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/efed1737-5de8-4313-9127-724bbd239f1a-scripts\") on node \"crc\" DevicePath \"\"" Nov 21 14:36:16 crc kubenswrapper[4897]: I1121 14:36:16.651634 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efed1737-5de8-4313-9127-724bbd239f1a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "efed1737-5de8-4313-9127-724bbd239f1a" (UID: "efed1737-5de8-4313-9127-724bbd239f1a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:36:16 crc kubenswrapper[4897]: I1121 14:36:16.671234 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efed1737-5de8-4313-9127-724bbd239f1a-config-data" (OuterVolumeSpecName: "config-data") pod "efed1737-5de8-4313-9127-724bbd239f1a" (UID: "efed1737-5de8-4313-9127-724bbd239f1a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:36:16 crc kubenswrapper[4897]: I1121 14:36:16.734471 4897 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/efed1737-5de8-4313-9127-724bbd239f1a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 21 14:36:16 crc kubenswrapper[4897]: I1121 14:36:16.734551 4897 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/efed1737-5de8-4313-9127-724bbd239f1a-config-data\") on node \"crc\" DevicePath \"\"" Nov 21 14:36:16 crc kubenswrapper[4897]: I1121 14:36:16.843963 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"efed1737-5de8-4313-9127-724bbd239f1a","Type":"ContainerDied","Data":"159cd7aa94cac5b32c8dbee75041c735a73b20d7d9cac3a941696bf88c467720"} Nov 21 14:36:16 crc kubenswrapper[4897]: I1121 14:36:16.844036 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 21 14:36:16 crc kubenswrapper[4897]: I1121 14:36:16.844037 4897 scope.go:117] "RemoveContainer" containerID="76bd36bbbb963c0c157f72905e2682ad3a3471d6d16c1c47c51a3e87bf44837e" Nov 21 14:36:16 crc kubenswrapper[4897]: I1121 14:36:16.933134 4897 scope.go:117] "RemoveContainer" containerID="aba81a8a1224fdc2f5618aeb6eb3be0ce285ed4b26b0e1d6ceaf5c836c7536a1" Nov 21 14:36:16 crc kubenswrapper[4897]: I1121 14:36:16.956589 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 21 14:36:16 crc kubenswrapper[4897]: I1121 14:36:16.986686 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 21 14:36:17 crc kubenswrapper[4897]: I1121 14:36:17.026548 4897 scope.go:117] "RemoveContainer" containerID="37c0701bd55a3a69856270d224fe1f7c29a3befbcc9c45fb1e1596e394a69933" Nov 21 14:36:17 crc kubenswrapper[4897]: I1121 14:36:17.026569 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 21 14:36:17 crc kubenswrapper[4897]: E1121 14:36:17.027310 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efed1737-5de8-4313-9127-724bbd239f1a" containerName="proxy-httpd" Nov 21 14:36:17 crc kubenswrapper[4897]: I1121 14:36:17.027326 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="efed1737-5de8-4313-9127-724bbd239f1a" containerName="proxy-httpd" Nov 21 14:36:17 crc kubenswrapper[4897]: E1121 14:36:17.027363 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efed1737-5de8-4313-9127-724bbd239f1a" containerName="sg-core" Nov 21 14:36:17 crc kubenswrapper[4897]: I1121 14:36:17.027375 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="efed1737-5de8-4313-9127-724bbd239f1a" containerName="sg-core" Nov 21 14:36:17 crc kubenswrapper[4897]: E1121 14:36:17.027387 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efed1737-5de8-4313-9127-724bbd239f1a" containerName="ceilometer-central-agent" Nov 21 14:36:17 crc kubenswrapper[4897]: I1121 14:36:17.027393 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="efed1737-5de8-4313-9127-724bbd239f1a" containerName="ceilometer-central-agent" Nov 21 14:36:17 crc kubenswrapper[4897]: E1121 14:36:17.027418 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efed1737-5de8-4313-9127-724bbd239f1a" containerName="ceilometer-notification-agent" Nov 21 14:36:17 crc kubenswrapper[4897]: I1121 14:36:17.027424 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="efed1737-5de8-4313-9127-724bbd239f1a" containerName="ceilometer-notification-agent" Nov 21 14:36:17 crc kubenswrapper[4897]: I1121 14:36:17.027656 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="efed1737-5de8-4313-9127-724bbd239f1a" containerName="sg-core" Nov 21 14:36:17 crc kubenswrapper[4897]: I1121 14:36:17.027689 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="efed1737-5de8-4313-9127-724bbd239f1a" containerName="ceilometer-central-agent" Nov 21 14:36:17 crc kubenswrapper[4897]: I1121 14:36:17.027714 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="efed1737-5de8-4313-9127-724bbd239f1a" containerName="ceilometer-notification-agent" Nov 21 14:36:17 crc kubenswrapper[4897]: I1121 14:36:17.027722 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="efed1737-5de8-4313-9127-724bbd239f1a" containerName="proxy-httpd" Nov 21 14:36:17 crc kubenswrapper[4897]: I1121 14:36:17.030211 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 21 14:36:17 crc kubenswrapper[4897]: I1121 14:36:17.033346 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 21 14:36:17 crc kubenswrapper[4897]: I1121 14:36:17.033388 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 21 14:36:17 crc kubenswrapper[4897]: I1121 14:36:17.037088 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 21 14:36:17 crc kubenswrapper[4897]: I1121 14:36:17.074635 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 21 14:36:17 crc kubenswrapper[4897]: I1121 14:36:17.088045 4897 scope.go:117] "RemoveContainer" containerID="5c1ffe3f03b119bb9d7d244acd512729212c30aafbda3158b86b3b645cdc77e5" Nov 21 14:36:17 crc kubenswrapper[4897]: I1121 14:36:17.148936 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5lcxs\" (UniqueName: \"kubernetes.io/projected/76d57203-0f85-48aa-87b0-a56af184da2d-kube-api-access-5lcxs\") pod \"ceilometer-0\" (UID: \"76d57203-0f85-48aa-87b0-a56af184da2d\") " pod="openstack/ceilometer-0" Nov 21 14:36:17 crc kubenswrapper[4897]: I1121 14:36:17.149136 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76d57203-0f85-48aa-87b0-a56af184da2d-config-data\") pod \"ceilometer-0\" (UID: \"76d57203-0f85-48aa-87b0-a56af184da2d\") " pod="openstack/ceilometer-0" Nov 21 14:36:17 crc kubenswrapper[4897]: I1121 14:36:17.149190 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76d57203-0f85-48aa-87b0-a56af184da2d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"76d57203-0f85-48aa-87b0-a56af184da2d\") " pod="openstack/ceilometer-0" Nov 21 14:36:17 crc kubenswrapper[4897]: I1121 14:36:17.149245 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/76d57203-0f85-48aa-87b0-a56af184da2d-run-httpd\") pod \"ceilometer-0\" (UID: \"76d57203-0f85-48aa-87b0-a56af184da2d\") " pod="openstack/ceilometer-0" Nov 21 14:36:17 crc kubenswrapper[4897]: I1121 14:36:17.149366 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/76d57203-0f85-48aa-87b0-a56af184da2d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"76d57203-0f85-48aa-87b0-a56af184da2d\") " pod="openstack/ceilometer-0" Nov 21 14:36:17 crc kubenswrapper[4897]: I1121 14:36:17.149411 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/76d57203-0f85-48aa-87b0-a56af184da2d-scripts\") pod \"ceilometer-0\" (UID: \"76d57203-0f85-48aa-87b0-a56af184da2d\") " pod="openstack/ceilometer-0" Nov 21 14:36:17 crc kubenswrapper[4897]: I1121 14:36:17.149442 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/76d57203-0f85-48aa-87b0-a56af184da2d-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"76d57203-0f85-48aa-87b0-a56af184da2d\") " pod="openstack/ceilometer-0" Nov 21 14:36:17 crc kubenswrapper[4897]: I1121 14:36:17.149901 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/76d57203-0f85-48aa-87b0-a56af184da2d-log-httpd\") pod \"ceilometer-0\" (UID: \"76d57203-0f85-48aa-87b0-a56af184da2d\") " pod="openstack/ceilometer-0" Nov 21 14:36:17 crc kubenswrapper[4897]: I1121 14:36:17.257542 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5lcxs\" (UniqueName: \"kubernetes.io/projected/76d57203-0f85-48aa-87b0-a56af184da2d-kube-api-access-5lcxs\") pod \"ceilometer-0\" (UID: \"76d57203-0f85-48aa-87b0-a56af184da2d\") " pod="openstack/ceilometer-0" Nov 21 14:36:17 crc kubenswrapper[4897]: I1121 14:36:17.257741 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76d57203-0f85-48aa-87b0-a56af184da2d-config-data\") pod \"ceilometer-0\" (UID: \"76d57203-0f85-48aa-87b0-a56af184da2d\") " pod="openstack/ceilometer-0" Nov 21 14:36:17 crc kubenswrapper[4897]: I1121 14:36:17.257774 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76d57203-0f85-48aa-87b0-a56af184da2d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"76d57203-0f85-48aa-87b0-a56af184da2d\") " pod="openstack/ceilometer-0" Nov 21 14:36:17 crc kubenswrapper[4897]: I1121 14:36:17.257824 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/76d57203-0f85-48aa-87b0-a56af184da2d-run-httpd\") pod \"ceilometer-0\" (UID: \"76d57203-0f85-48aa-87b0-a56af184da2d\") " pod="openstack/ceilometer-0" Nov 21 14:36:17 crc kubenswrapper[4897]: I1121 14:36:17.257946 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/76d57203-0f85-48aa-87b0-a56af184da2d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"76d57203-0f85-48aa-87b0-a56af184da2d\") " pod="openstack/ceilometer-0" Nov 21 14:36:17 crc kubenswrapper[4897]: I1121 14:36:17.257989 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/76d57203-0f85-48aa-87b0-a56af184da2d-scripts\") pod \"ceilometer-0\" (UID: \"76d57203-0f85-48aa-87b0-a56af184da2d\") " pod="openstack/ceilometer-0" Nov 21 14:36:17 crc kubenswrapper[4897]: I1121 14:36:17.258010 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/76d57203-0f85-48aa-87b0-a56af184da2d-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"76d57203-0f85-48aa-87b0-a56af184da2d\") " pod="openstack/ceilometer-0" Nov 21 14:36:17 crc kubenswrapper[4897]: I1121 14:36:17.258093 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/76d57203-0f85-48aa-87b0-a56af184da2d-log-httpd\") pod \"ceilometer-0\" (UID: \"76d57203-0f85-48aa-87b0-a56af184da2d\") " pod="openstack/ceilometer-0" Nov 21 14:36:17 crc kubenswrapper[4897]: I1121 14:36:17.258616 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/76d57203-0f85-48aa-87b0-a56af184da2d-run-httpd\") pod \"ceilometer-0\" (UID: \"76d57203-0f85-48aa-87b0-a56af184da2d\") " pod="openstack/ceilometer-0" Nov 21 14:36:17 crc kubenswrapper[4897]: I1121 14:36:17.258714 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/76d57203-0f85-48aa-87b0-a56af184da2d-log-httpd\") pod \"ceilometer-0\" (UID: \"76d57203-0f85-48aa-87b0-a56af184da2d\") " pod="openstack/ceilometer-0" Nov 21 14:36:17 crc kubenswrapper[4897]: I1121 14:36:17.262525 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/76d57203-0f85-48aa-87b0-a56af184da2d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"76d57203-0f85-48aa-87b0-a56af184da2d\") " pod="openstack/ceilometer-0" Nov 21 14:36:17 crc kubenswrapper[4897]: I1121 14:36:17.263002 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/76d57203-0f85-48aa-87b0-a56af184da2d-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"76d57203-0f85-48aa-87b0-a56af184da2d\") " pod="openstack/ceilometer-0" Nov 21 14:36:17 crc kubenswrapper[4897]: I1121 14:36:17.263449 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76d57203-0f85-48aa-87b0-a56af184da2d-config-data\") pod \"ceilometer-0\" (UID: \"76d57203-0f85-48aa-87b0-a56af184da2d\") " pod="openstack/ceilometer-0" Nov 21 14:36:17 crc kubenswrapper[4897]: I1121 14:36:17.263472 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76d57203-0f85-48aa-87b0-a56af184da2d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"76d57203-0f85-48aa-87b0-a56af184da2d\") " pod="openstack/ceilometer-0" Nov 21 14:36:17 crc kubenswrapper[4897]: I1121 14:36:17.266431 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/76d57203-0f85-48aa-87b0-a56af184da2d-scripts\") pod \"ceilometer-0\" (UID: \"76d57203-0f85-48aa-87b0-a56af184da2d\") " pod="openstack/ceilometer-0" Nov 21 14:36:17 crc kubenswrapper[4897]: I1121 14:36:17.279373 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5lcxs\" (UniqueName: \"kubernetes.io/projected/76d57203-0f85-48aa-87b0-a56af184da2d-kube-api-access-5lcxs\") pod \"ceilometer-0\" (UID: \"76d57203-0f85-48aa-87b0-a56af184da2d\") " pod="openstack/ceilometer-0" Nov 21 14:36:17 crc kubenswrapper[4897]: I1121 14:36:17.353282 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 21 14:36:17 crc kubenswrapper[4897]: I1121 14:36:17.489996 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Nov 21 14:36:17 crc kubenswrapper[4897]: I1121 14:36:17.859720 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 21 14:36:18 crc kubenswrapper[4897]: I1121 14:36:18.102433 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efed1737-5de8-4313-9127-724bbd239f1a" path="/var/lib/kubelet/pods/efed1737-5de8-4313-9127-724bbd239f1a/volumes" Nov 21 14:36:18 crc kubenswrapper[4897]: I1121 14:36:18.871325 4897 generic.go:334] "Generic (PLEG): container finished" podID="d3a7078f-ec31-46d5-aadb-b119532328b8" containerID="785d3b3f0d81f47cf04eb76ef2e2d4232e551f90cd5d334cb72df9b89716f361" exitCode=0 Nov 21 14:36:18 crc kubenswrapper[4897]: I1121 14:36:18.871406 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d3a7078f-ec31-46d5-aadb-b119532328b8","Type":"ContainerDied","Data":"785d3b3f0d81f47cf04eb76ef2e2d4232e551f90cd5d334cb72df9b89716f361"} Nov 21 14:36:18 crc kubenswrapper[4897]: I1121 14:36:18.872809 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"76d57203-0f85-48aa-87b0-a56af184da2d","Type":"ContainerStarted","Data":"a47f15f73782016a708b146c32acd32a80a6e64ed1a06625faa63a9bb2c4c819"} Nov 21 14:36:19 crc kubenswrapper[4897]: E1121 14:36:19.783686 4897 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podafa69523_63c4_4a94_b94a_61c1343ef10c.slice/crio-conmon-6cb3731d3191a0096a5051532c43fd96973f7663ef9d3571618a165b3dd1afca.scope\": RecentStats: unable to find data in memory cache]" Nov 21 14:36:19 crc kubenswrapper[4897]: I1121 14:36:19.795135 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 21 14:36:19 crc kubenswrapper[4897]: I1121 14:36:19.888780 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d3a7078f-ec31-46d5-aadb-b119532328b8","Type":"ContainerDied","Data":"21066b7db277e30b514dd6581465dea216ce805e03982ae699fafa33de8c6c5f"} Nov 21 14:36:19 crc kubenswrapper[4897]: I1121 14:36:19.888855 4897 scope.go:117] "RemoveContainer" containerID="785d3b3f0d81f47cf04eb76ef2e2d4232e551f90cd5d334cb72df9b89716f361" Nov 21 14:36:19 crc kubenswrapper[4897]: I1121 14:36:19.889131 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 21 14:36:19 crc kubenswrapper[4897]: I1121 14:36:19.916599 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3a7078f-ec31-46d5-aadb-b119532328b8-config-data\") pod \"d3a7078f-ec31-46d5-aadb-b119532328b8\" (UID: \"d3a7078f-ec31-46d5-aadb-b119532328b8\") " Nov 21 14:36:19 crc kubenswrapper[4897]: I1121 14:36:19.916646 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hmfwh\" (UniqueName: \"kubernetes.io/projected/d3a7078f-ec31-46d5-aadb-b119532328b8-kube-api-access-hmfwh\") pod \"d3a7078f-ec31-46d5-aadb-b119532328b8\" (UID: \"d3a7078f-ec31-46d5-aadb-b119532328b8\") " Nov 21 14:36:19 crc kubenswrapper[4897]: I1121 14:36:19.916708 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d3a7078f-ec31-46d5-aadb-b119532328b8-logs\") pod \"d3a7078f-ec31-46d5-aadb-b119532328b8\" (UID: \"d3a7078f-ec31-46d5-aadb-b119532328b8\") " Nov 21 14:36:19 crc kubenswrapper[4897]: I1121 14:36:19.916742 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3a7078f-ec31-46d5-aadb-b119532328b8-combined-ca-bundle\") pod \"d3a7078f-ec31-46d5-aadb-b119532328b8\" (UID: \"d3a7078f-ec31-46d5-aadb-b119532328b8\") " Nov 21 14:36:19 crc kubenswrapper[4897]: I1121 14:36:19.917607 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d3a7078f-ec31-46d5-aadb-b119532328b8-logs" (OuterVolumeSpecName: "logs") pod "d3a7078f-ec31-46d5-aadb-b119532328b8" (UID: "d3a7078f-ec31-46d5-aadb-b119532328b8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:36:19 crc kubenswrapper[4897]: I1121 14:36:19.922402 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3a7078f-ec31-46d5-aadb-b119532328b8-kube-api-access-hmfwh" (OuterVolumeSpecName: "kube-api-access-hmfwh") pod "d3a7078f-ec31-46d5-aadb-b119532328b8" (UID: "d3a7078f-ec31-46d5-aadb-b119532328b8"). InnerVolumeSpecName "kube-api-access-hmfwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:36:19 crc kubenswrapper[4897]: I1121 14:36:19.958661 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3a7078f-ec31-46d5-aadb-b119532328b8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d3a7078f-ec31-46d5-aadb-b119532328b8" (UID: "d3a7078f-ec31-46d5-aadb-b119532328b8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:36:19 crc kubenswrapper[4897]: I1121 14:36:19.964663 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3a7078f-ec31-46d5-aadb-b119532328b8-config-data" (OuterVolumeSpecName: "config-data") pod "d3a7078f-ec31-46d5-aadb-b119532328b8" (UID: "d3a7078f-ec31-46d5-aadb-b119532328b8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:36:20 crc kubenswrapper[4897]: I1121 14:36:20.019702 4897 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3a7078f-ec31-46d5-aadb-b119532328b8-config-data\") on node \"crc\" DevicePath \"\"" Nov 21 14:36:20 crc kubenswrapper[4897]: I1121 14:36:20.019739 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hmfwh\" (UniqueName: \"kubernetes.io/projected/d3a7078f-ec31-46d5-aadb-b119532328b8-kube-api-access-hmfwh\") on node \"crc\" DevicePath \"\"" Nov 21 14:36:20 crc kubenswrapper[4897]: I1121 14:36:20.019749 4897 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d3a7078f-ec31-46d5-aadb-b119532328b8-logs\") on node \"crc\" DevicePath \"\"" Nov 21 14:36:20 crc kubenswrapper[4897]: I1121 14:36:20.019760 4897 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3a7078f-ec31-46d5-aadb-b119532328b8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 21 14:36:20 crc kubenswrapper[4897]: I1121 14:36:20.215927 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 21 14:36:20 crc kubenswrapper[4897]: I1121 14:36:20.227557 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 21 14:36:20 crc kubenswrapper[4897]: I1121 14:36:20.253467 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 21 14:36:20 crc kubenswrapper[4897]: E1121 14:36:20.254211 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3a7078f-ec31-46d5-aadb-b119532328b8" containerName="nova-api-api" Nov 21 14:36:20 crc kubenswrapper[4897]: I1121 14:36:20.254230 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3a7078f-ec31-46d5-aadb-b119532328b8" containerName="nova-api-api" Nov 21 14:36:20 crc kubenswrapper[4897]: E1121 14:36:20.254253 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3a7078f-ec31-46d5-aadb-b119532328b8" containerName="nova-api-log" Nov 21 14:36:20 crc kubenswrapper[4897]: I1121 14:36:20.254262 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3a7078f-ec31-46d5-aadb-b119532328b8" containerName="nova-api-log" Nov 21 14:36:20 crc kubenswrapper[4897]: I1121 14:36:20.254550 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3a7078f-ec31-46d5-aadb-b119532328b8" containerName="nova-api-log" Nov 21 14:36:20 crc kubenswrapper[4897]: I1121 14:36:20.254587 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3a7078f-ec31-46d5-aadb-b119532328b8" containerName="nova-api-api" Nov 21 14:36:20 crc kubenswrapper[4897]: I1121 14:36:20.256210 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 21 14:36:20 crc kubenswrapper[4897]: I1121 14:36:20.263868 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Nov 21 14:36:20 crc kubenswrapper[4897]: I1121 14:36:20.263942 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 21 14:36:20 crc kubenswrapper[4897]: I1121 14:36:20.264132 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Nov 21 14:36:20 crc kubenswrapper[4897]: I1121 14:36:20.269321 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 21 14:36:20 crc kubenswrapper[4897]: I1121 14:36:20.330597 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f688a5e9-9393-4dee-b49a-22b6c75ecfaf-internal-tls-certs\") pod \"nova-api-0\" (UID: \"f688a5e9-9393-4dee-b49a-22b6c75ecfaf\") " pod="openstack/nova-api-0" Nov 21 14:36:20 crc kubenswrapper[4897]: I1121 14:36:20.330673 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f688a5e9-9393-4dee-b49a-22b6c75ecfaf-public-tls-certs\") pod \"nova-api-0\" (UID: \"f688a5e9-9393-4dee-b49a-22b6c75ecfaf\") " pod="openstack/nova-api-0" Nov 21 14:36:20 crc kubenswrapper[4897]: I1121 14:36:20.330750 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f688a5e9-9393-4dee-b49a-22b6c75ecfaf-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"f688a5e9-9393-4dee-b49a-22b6c75ecfaf\") " pod="openstack/nova-api-0" Nov 21 14:36:20 crc kubenswrapper[4897]: I1121 14:36:20.330793 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f688a5e9-9393-4dee-b49a-22b6c75ecfaf-config-data\") pod \"nova-api-0\" (UID: \"f688a5e9-9393-4dee-b49a-22b6c75ecfaf\") " pod="openstack/nova-api-0" Nov 21 14:36:20 crc kubenswrapper[4897]: I1121 14:36:20.330875 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rzt2r\" (UniqueName: \"kubernetes.io/projected/f688a5e9-9393-4dee-b49a-22b6c75ecfaf-kube-api-access-rzt2r\") pod \"nova-api-0\" (UID: \"f688a5e9-9393-4dee-b49a-22b6c75ecfaf\") " pod="openstack/nova-api-0" Nov 21 14:36:20 crc kubenswrapper[4897]: I1121 14:36:20.332908 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f688a5e9-9393-4dee-b49a-22b6c75ecfaf-logs\") pod \"nova-api-0\" (UID: \"f688a5e9-9393-4dee-b49a-22b6c75ecfaf\") " pod="openstack/nova-api-0" Nov 21 14:36:20 crc kubenswrapper[4897]: I1121 14:36:20.347572 4897 scope.go:117] "RemoveContainer" containerID="0dceb413fddf48e1b6387a57eb97ad30211325e57d6120065cb6b2ff4baee214" Nov 21 14:36:20 crc kubenswrapper[4897]: I1121 14:36:20.438601 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f688a5e9-9393-4dee-b49a-22b6c75ecfaf-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"f688a5e9-9393-4dee-b49a-22b6c75ecfaf\") " pod="openstack/nova-api-0" Nov 21 14:36:20 crc kubenswrapper[4897]: I1121 14:36:20.439026 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f688a5e9-9393-4dee-b49a-22b6c75ecfaf-config-data\") pod \"nova-api-0\" (UID: \"f688a5e9-9393-4dee-b49a-22b6c75ecfaf\") " pod="openstack/nova-api-0" Nov 21 14:36:20 crc kubenswrapper[4897]: I1121 14:36:20.439110 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rzt2r\" (UniqueName: \"kubernetes.io/projected/f688a5e9-9393-4dee-b49a-22b6c75ecfaf-kube-api-access-rzt2r\") pod \"nova-api-0\" (UID: \"f688a5e9-9393-4dee-b49a-22b6c75ecfaf\") " pod="openstack/nova-api-0" Nov 21 14:36:20 crc kubenswrapper[4897]: I1121 14:36:20.439159 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f688a5e9-9393-4dee-b49a-22b6c75ecfaf-logs\") pod \"nova-api-0\" (UID: \"f688a5e9-9393-4dee-b49a-22b6c75ecfaf\") " pod="openstack/nova-api-0" Nov 21 14:36:20 crc kubenswrapper[4897]: I1121 14:36:20.439298 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f688a5e9-9393-4dee-b49a-22b6c75ecfaf-internal-tls-certs\") pod \"nova-api-0\" (UID: \"f688a5e9-9393-4dee-b49a-22b6c75ecfaf\") " pod="openstack/nova-api-0" Nov 21 14:36:20 crc kubenswrapper[4897]: I1121 14:36:20.439386 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f688a5e9-9393-4dee-b49a-22b6c75ecfaf-public-tls-certs\") pod \"nova-api-0\" (UID: \"f688a5e9-9393-4dee-b49a-22b6c75ecfaf\") " pod="openstack/nova-api-0" Nov 21 14:36:20 crc kubenswrapper[4897]: I1121 14:36:20.439868 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f688a5e9-9393-4dee-b49a-22b6c75ecfaf-logs\") pod \"nova-api-0\" (UID: \"f688a5e9-9393-4dee-b49a-22b6c75ecfaf\") " pod="openstack/nova-api-0" Nov 21 14:36:20 crc kubenswrapper[4897]: I1121 14:36:20.445266 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f688a5e9-9393-4dee-b49a-22b6c75ecfaf-internal-tls-certs\") pod \"nova-api-0\" (UID: \"f688a5e9-9393-4dee-b49a-22b6c75ecfaf\") " pod="openstack/nova-api-0" Nov 21 14:36:20 crc kubenswrapper[4897]: I1121 14:36:20.445769 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f688a5e9-9393-4dee-b49a-22b6c75ecfaf-public-tls-certs\") pod \"nova-api-0\" (UID: \"f688a5e9-9393-4dee-b49a-22b6c75ecfaf\") " pod="openstack/nova-api-0" Nov 21 14:36:20 crc kubenswrapper[4897]: I1121 14:36:20.451195 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f688a5e9-9393-4dee-b49a-22b6c75ecfaf-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"f688a5e9-9393-4dee-b49a-22b6c75ecfaf\") " pod="openstack/nova-api-0" Nov 21 14:36:20 crc kubenswrapper[4897]: I1121 14:36:20.454228 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f688a5e9-9393-4dee-b49a-22b6c75ecfaf-config-data\") pod \"nova-api-0\" (UID: \"f688a5e9-9393-4dee-b49a-22b6c75ecfaf\") " pod="openstack/nova-api-0" Nov 21 14:36:20 crc kubenswrapper[4897]: I1121 14:36:20.459283 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rzt2r\" (UniqueName: \"kubernetes.io/projected/f688a5e9-9393-4dee-b49a-22b6c75ecfaf-kube-api-access-rzt2r\") pod \"nova-api-0\" (UID: \"f688a5e9-9393-4dee-b49a-22b6c75ecfaf\") " pod="openstack/nova-api-0" Nov 21 14:36:20 crc kubenswrapper[4897]: I1121 14:36:20.579848 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 21 14:36:21 crc kubenswrapper[4897]: I1121 14:36:21.090335 4897 scope.go:117] "RemoveContainer" containerID="c68f1b003d12b4a2964169a85a04e3f0272d86c891f52e6bebd52f3781520d1a" Nov 21 14:36:21 crc kubenswrapper[4897]: E1121 14:36:21.090665 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 14:36:21 crc kubenswrapper[4897]: E1121 14:36:21.231712 4897 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podafa69523_63c4_4a94_b94a_61c1343ef10c.slice/crio-conmon-6cb3731d3191a0096a5051532c43fd96973f7663ef9d3571618a165b3dd1afca.scope\": RecentStats: unable to find data in memory cache]" Nov 21 14:36:21 crc kubenswrapper[4897]: I1121 14:36:21.280968 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 21 14:36:21 crc kubenswrapper[4897]: I1121 14:36:21.923071 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f688a5e9-9393-4dee-b49a-22b6c75ecfaf","Type":"ContainerStarted","Data":"5dc39c26fd2373cc55c8e7343e92edce69c4b0dde2cd59ca4fa105ccca9eae4c"} Nov 21 14:36:21 crc kubenswrapper[4897]: I1121 14:36:21.924081 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f688a5e9-9393-4dee-b49a-22b6c75ecfaf","Type":"ContainerStarted","Data":"b3a5275a5fd8512eb4e39e68ec4657cab31565fa3ef6b7295ef8a28ff8a4ed4f"} Nov 21 14:36:21 crc kubenswrapper[4897]: I1121 14:36:21.924731 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"76d57203-0f85-48aa-87b0-a56af184da2d","Type":"ContainerStarted","Data":"249e82c4c97ad9c748b6487ef169e6b51f1b63609fd68c0f73ad594cc5cf7f86"} Nov 21 14:36:22 crc kubenswrapper[4897]: I1121 14:36:22.135117 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3a7078f-ec31-46d5-aadb-b119532328b8" path="/var/lib/kubelet/pods/d3a7078f-ec31-46d5-aadb-b119532328b8/volumes" Nov 21 14:36:22 crc kubenswrapper[4897]: I1121 14:36:22.296700 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6d99f6bc7f-k7pkd" Nov 21 14:36:22 crc kubenswrapper[4897]: I1121 14:36:22.377850 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7877d89589-c28gz"] Nov 21 14:36:22 crc kubenswrapper[4897]: I1121 14:36:22.378423 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7877d89589-c28gz" podUID="190e6358-2083-4549-b5ac-5c35af169b06" containerName="dnsmasq-dns" containerID="cri-o://d2be4b278d182e43defddab52a77b6052ea8408529ad229d4d7dbe3734a7b4c4" gracePeriod=10 Nov 21 14:36:22 crc kubenswrapper[4897]: I1121 14:36:22.490780 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Nov 21 14:36:22 crc kubenswrapper[4897]: I1121 14:36:22.530794 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Nov 21 14:36:22 crc kubenswrapper[4897]: I1121 14:36:22.935120 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f688a5e9-9393-4dee-b49a-22b6c75ecfaf","Type":"ContainerStarted","Data":"d4a7106c42b29ced91fa9c22a9fab018391f685ce1bad91f2ea315fbc57e9bd9"} Nov 21 14:36:22 crc kubenswrapper[4897]: I1121 14:36:22.936478 4897 generic.go:334] "Generic (PLEG): container finished" podID="190e6358-2083-4549-b5ac-5c35af169b06" containerID="d2be4b278d182e43defddab52a77b6052ea8408529ad229d4d7dbe3734a7b4c4" exitCode=0 Nov 21 14:36:22 crc kubenswrapper[4897]: I1121 14:36:22.936497 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7877d89589-c28gz" event={"ID":"190e6358-2083-4549-b5ac-5c35af169b06","Type":"ContainerDied","Data":"d2be4b278d182e43defddab52a77b6052ea8408529ad229d4d7dbe3734a7b4c4"} Nov 21 14:36:22 crc kubenswrapper[4897]: I1121 14:36:22.936529 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7877d89589-c28gz" event={"ID":"190e6358-2083-4549-b5ac-5c35af169b06","Type":"ContainerDied","Data":"4866d456c7234eff14ff58668ec55c79fbc8f0dc56704b57566506437a25d37d"} Nov 21 14:36:22 crc kubenswrapper[4897]: I1121 14:36:22.936542 4897 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4866d456c7234eff14ff58668ec55c79fbc8f0dc56704b57566506437a25d37d" Nov 21 14:36:22 crc kubenswrapper[4897]: I1121 14:36:22.962841 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Nov 21 14:36:22 crc kubenswrapper[4897]: I1121 14:36:22.973355 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.97333421 podStartE2EDuration="2.97333421s" podCreationTimestamp="2025-11-21 14:36:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:36:22.960337411 +0000 UTC m=+1660.244930896" watchObservedRunningTime="2025-11-21 14:36:22.97333421 +0000 UTC m=+1660.257927685" Nov 21 14:36:23 crc kubenswrapper[4897]: I1121 14:36:23.076772 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7877d89589-c28gz" Nov 21 14:36:23 crc kubenswrapper[4897]: I1121 14:36:23.147071 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-dgr4b"] Nov 21 14:36:23 crc kubenswrapper[4897]: E1121 14:36:23.148088 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="190e6358-2083-4549-b5ac-5c35af169b06" containerName="dnsmasq-dns" Nov 21 14:36:23 crc kubenswrapper[4897]: I1121 14:36:23.148112 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="190e6358-2083-4549-b5ac-5c35af169b06" containerName="dnsmasq-dns" Nov 21 14:36:23 crc kubenswrapper[4897]: E1121 14:36:23.148160 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="190e6358-2083-4549-b5ac-5c35af169b06" containerName="init" Nov 21 14:36:23 crc kubenswrapper[4897]: I1121 14:36:23.148169 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="190e6358-2083-4549-b5ac-5c35af169b06" containerName="init" Nov 21 14:36:23 crc kubenswrapper[4897]: I1121 14:36:23.148632 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="190e6358-2083-4549-b5ac-5c35af169b06" containerName="dnsmasq-dns" Nov 21 14:36:23 crc kubenswrapper[4897]: I1121 14:36:23.149892 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-dgr4b" Nov 21 14:36:23 crc kubenswrapper[4897]: I1121 14:36:23.153471 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Nov 21 14:36:23 crc kubenswrapper[4897]: I1121 14:36:23.158533 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Nov 21 14:36:23 crc kubenswrapper[4897]: I1121 14:36:23.193243 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-dgr4b"] Nov 21 14:36:23 crc kubenswrapper[4897]: I1121 14:36:23.218594 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/190e6358-2083-4549-b5ac-5c35af169b06-dns-swift-storage-0\") pod \"190e6358-2083-4549-b5ac-5c35af169b06\" (UID: \"190e6358-2083-4549-b5ac-5c35af169b06\") " Nov 21 14:36:23 crc kubenswrapper[4897]: I1121 14:36:23.218680 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/190e6358-2083-4549-b5ac-5c35af169b06-ovsdbserver-nb\") pod \"190e6358-2083-4549-b5ac-5c35af169b06\" (UID: \"190e6358-2083-4549-b5ac-5c35af169b06\") " Nov 21 14:36:23 crc kubenswrapper[4897]: I1121 14:36:23.218802 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/190e6358-2083-4549-b5ac-5c35af169b06-dns-svc\") pod \"190e6358-2083-4549-b5ac-5c35af169b06\" (UID: \"190e6358-2083-4549-b5ac-5c35af169b06\") " Nov 21 14:36:23 crc kubenswrapper[4897]: I1121 14:36:23.218842 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4qcr2\" (UniqueName: \"kubernetes.io/projected/190e6358-2083-4549-b5ac-5c35af169b06-kube-api-access-4qcr2\") pod \"190e6358-2083-4549-b5ac-5c35af169b06\" (UID: \"190e6358-2083-4549-b5ac-5c35af169b06\") " Nov 21 14:36:23 crc kubenswrapper[4897]: I1121 14:36:23.218864 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/190e6358-2083-4549-b5ac-5c35af169b06-config\") pod \"190e6358-2083-4549-b5ac-5c35af169b06\" (UID: \"190e6358-2083-4549-b5ac-5c35af169b06\") " Nov 21 14:36:23 crc kubenswrapper[4897]: I1121 14:36:23.218923 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/190e6358-2083-4549-b5ac-5c35af169b06-ovsdbserver-sb\") pod \"190e6358-2083-4549-b5ac-5c35af169b06\" (UID: \"190e6358-2083-4549-b5ac-5c35af169b06\") " Nov 21 14:36:23 crc kubenswrapper[4897]: I1121 14:36:23.240177 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/190e6358-2083-4549-b5ac-5c35af169b06-kube-api-access-4qcr2" (OuterVolumeSpecName: "kube-api-access-4qcr2") pod "190e6358-2083-4549-b5ac-5c35af169b06" (UID: "190e6358-2083-4549-b5ac-5c35af169b06"). InnerVolumeSpecName "kube-api-access-4qcr2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:36:23 crc kubenswrapper[4897]: I1121 14:36:23.283972 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/190e6358-2083-4549-b5ac-5c35af169b06-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "190e6358-2083-4549-b5ac-5c35af169b06" (UID: "190e6358-2083-4549-b5ac-5c35af169b06"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:36:23 crc kubenswrapper[4897]: I1121 14:36:23.287149 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/190e6358-2083-4549-b5ac-5c35af169b06-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "190e6358-2083-4549-b5ac-5c35af169b06" (UID: "190e6358-2083-4549-b5ac-5c35af169b06"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:36:23 crc kubenswrapper[4897]: I1121 14:36:23.306111 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/190e6358-2083-4549-b5ac-5c35af169b06-config" (OuterVolumeSpecName: "config") pod "190e6358-2083-4549-b5ac-5c35af169b06" (UID: "190e6358-2083-4549-b5ac-5c35af169b06"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:36:23 crc kubenswrapper[4897]: I1121 14:36:23.307471 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/190e6358-2083-4549-b5ac-5c35af169b06-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "190e6358-2083-4549-b5ac-5c35af169b06" (UID: "190e6358-2083-4549-b5ac-5c35af169b06"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:36:23 crc kubenswrapper[4897]: I1121 14:36:23.322184 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z4bpn\" (UniqueName: \"kubernetes.io/projected/522a19f3-71a6-4609-89cd-6142f5f0d375-kube-api-access-z4bpn\") pod \"nova-cell1-cell-mapping-dgr4b\" (UID: \"522a19f3-71a6-4609-89cd-6142f5f0d375\") " pod="openstack/nova-cell1-cell-mapping-dgr4b" Nov 21 14:36:23 crc kubenswrapper[4897]: I1121 14:36:23.322283 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/522a19f3-71a6-4609-89cd-6142f5f0d375-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-dgr4b\" (UID: \"522a19f3-71a6-4609-89cd-6142f5f0d375\") " pod="openstack/nova-cell1-cell-mapping-dgr4b" Nov 21 14:36:23 crc kubenswrapper[4897]: I1121 14:36:23.322309 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/522a19f3-71a6-4609-89cd-6142f5f0d375-scripts\") pod \"nova-cell1-cell-mapping-dgr4b\" (UID: \"522a19f3-71a6-4609-89cd-6142f5f0d375\") " pod="openstack/nova-cell1-cell-mapping-dgr4b" Nov 21 14:36:23 crc kubenswrapper[4897]: I1121 14:36:23.322357 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/522a19f3-71a6-4609-89cd-6142f5f0d375-config-data\") pod \"nova-cell1-cell-mapping-dgr4b\" (UID: \"522a19f3-71a6-4609-89cd-6142f5f0d375\") " pod="openstack/nova-cell1-cell-mapping-dgr4b" Nov 21 14:36:23 crc kubenswrapper[4897]: I1121 14:36:23.322428 4897 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/190e6358-2083-4549-b5ac-5c35af169b06-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 21 14:36:23 crc kubenswrapper[4897]: I1121 14:36:23.322440 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4qcr2\" (UniqueName: \"kubernetes.io/projected/190e6358-2083-4549-b5ac-5c35af169b06-kube-api-access-4qcr2\") on node \"crc\" DevicePath \"\"" Nov 21 14:36:23 crc kubenswrapper[4897]: I1121 14:36:23.322450 4897 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/190e6358-2083-4549-b5ac-5c35af169b06-config\") on node \"crc\" DevicePath \"\"" Nov 21 14:36:23 crc kubenswrapper[4897]: I1121 14:36:23.322458 4897 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/190e6358-2083-4549-b5ac-5c35af169b06-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 21 14:36:23 crc kubenswrapper[4897]: I1121 14:36:23.322466 4897 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/190e6358-2083-4549-b5ac-5c35af169b06-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 21 14:36:23 crc kubenswrapper[4897]: I1121 14:36:23.325029 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/190e6358-2083-4549-b5ac-5c35af169b06-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "190e6358-2083-4549-b5ac-5c35af169b06" (UID: "190e6358-2083-4549-b5ac-5c35af169b06"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:36:23 crc kubenswrapper[4897]: I1121 14:36:23.427329 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/522a19f3-71a6-4609-89cd-6142f5f0d375-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-dgr4b\" (UID: \"522a19f3-71a6-4609-89cd-6142f5f0d375\") " pod="openstack/nova-cell1-cell-mapping-dgr4b" Nov 21 14:36:23 crc kubenswrapper[4897]: I1121 14:36:23.427401 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/522a19f3-71a6-4609-89cd-6142f5f0d375-scripts\") pod \"nova-cell1-cell-mapping-dgr4b\" (UID: \"522a19f3-71a6-4609-89cd-6142f5f0d375\") " pod="openstack/nova-cell1-cell-mapping-dgr4b" Nov 21 14:36:23 crc kubenswrapper[4897]: I1121 14:36:23.427484 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/522a19f3-71a6-4609-89cd-6142f5f0d375-config-data\") pod \"nova-cell1-cell-mapping-dgr4b\" (UID: \"522a19f3-71a6-4609-89cd-6142f5f0d375\") " pod="openstack/nova-cell1-cell-mapping-dgr4b" Nov 21 14:36:23 crc kubenswrapper[4897]: I1121 14:36:23.427797 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z4bpn\" (UniqueName: \"kubernetes.io/projected/522a19f3-71a6-4609-89cd-6142f5f0d375-kube-api-access-z4bpn\") pod \"nova-cell1-cell-mapping-dgr4b\" (UID: \"522a19f3-71a6-4609-89cd-6142f5f0d375\") " pod="openstack/nova-cell1-cell-mapping-dgr4b" Nov 21 14:36:23 crc kubenswrapper[4897]: I1121 14:36:23.427977 4897 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/190e6358-2083-4549-b5ac-5c35af169b06-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 21 14:36:23 crc kubenswrapper[4897]: I1121 14:36:23.436609 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/522a19f3-71a6-4609-89cd-6142f5f0d375-scripts\") pod \"nova-cell1-cell-mapping-dgr4b\" (UID: \"522a19f3-71a6-4609-89cd-6142f5f0d375\") " pod="openstack/nova-cell1-cell-mapping-dgr4b" Nov 21 14:36:23 crc kubenswrapper[4897]: I1121 14:36:23.440300 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/522a19f3-71a6-4609-89cd-6142f5f0d375-config-data\") pod \"nova-cell1-cell-mapping-dgr4b\" (UID: \"522a19f3-71a6-4609-89cd-6142f5f0d375\") " pod="openstack/nova-cell1-cell-mapping-dgr4b" Nov 21 14:36:23 crc kubenswrapper[4897]: I1121 14:36:23.441429 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/522a19f3-71a6-4609-89cd-6142f5f0d375-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-dgr4b\" (UID: \"522a19f3-71a6-4609-89cd-6142f5f0d375\") " pod="openstack/nova-cell1-cell-mapping-dgr4b" Nov 21 14:36:23 crc kubenswrapper[4897]: I1121 14:36:23.457283 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z4bpn\" (UniqueName: \"kubernetes.io/projected/522a19f3-71a6-4609-89cd-6142f5f0d375-kube-api-access-z4bpn\") pod \"nova-cell1-cell-mapping-dgr4b\" (UID: \"522a19f3-71a6-4609-89cd-6142f5f0d375\") " pod="openstack/nova-cell1-cell-mapping-dgr4b" Nov 21 14:36:23 crc kubenswrapper[4897]: I1121 14:36:23.504689 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-dgr4b" Nov 21 14:36:23 crc kubenswrapper[4897]: I1121 14:36:23.948105 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"76d57203-0f85-48aa-87b0-a56af184da2d","Type":"ContainerStarted","Data":"40b3f26ee7a574717df483a754491f7e0c2353dfd0859d58892267906ac0a7d2"} Nov 21 14:36:23 crc kubenswrapper[4897]: I1121 14:36:23.948156 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7877d89589-c28gz" Nov 21 14:36:23 crc kubenswrapper[4897]: I1121 14:36:23.987941 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7877d89589-c28gz"] Nov 21 14:36:23 crc kubenswrapper[4897]: I1121 14:36:23.999099 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7877d89589-c28gz"] Nov 21 14:36:24 crc kubenswrapper[4897]: I1121 14:36:24.014363 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-dgr4b"] Nov 21 14:36:24 crc kubenswrapper[4897]: I1121 14:36:24.106174 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="190e6358-2083-4549-b5ac-5c35af169b06" path="/var/lib/kubelet/pods/190e6358-2083-4549-b5ac-5c35af169b06/volumes" Nov 21 14:36:24 crc kubenswrapper[4897]: I1121 14:36:24.967746 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-dgr4b" event={"ID":"522a19f3-71a6-4609-89cd-6142f5f0d375","Type":"ContainerStarted","Data":"8f7788e6ea517fa5675fbd9e32ba8202c7b3ea53d2a83293a84c4235cad59e54"} Nov 21 14:36:24 crc kubenswrapper[4897]: I1121 14:36:24.968167 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-dgr4b" event={"ID":"522a19f3-71a6-4609-89cd-6142f5f0d375","Type":"ContainerStarted","Data":"b722d0d29f35a524443c07c64761f6fa719b334924c25016e5c245313081481f"} Nov 21 14:36:24 crc kubenswrapper[4897]: I1121 14:36:24.988413 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-dgr4b" podStartSLOduration=1.988393523 podStartE2EDuration="1.988393523s" podCreationTimestamp="2025-11-21 14:36:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:36:24.981253771 +0000 UTC m=+1662.265847236" watchObservedRunningTime="2025-11-21 14:36:24.988393523 +0000 UTC m=+1662.272986998" Nov 21 14:36:26 crc kubenswrapper[4897]: I1121 14:36:26.174818 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-dtwql" podUID="6d3cad20-fb17-4726-b8fa-9fd5303b5eea" containerName="registry-server" probeResult="failure" output=< Nov 21 14:36:26 crc kubenswrapper[4897]: timeout: failed to connect service ":50051" within 1s Nov 21 14:36:26 crc kubenswrapper[4897]: > Nov 21 14:36:27 crc kubenswrapper[4897]: I1121 14:36:27.008166 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"76d57203-0f85-48aa-87b0-a56af184da2d","Type":"ContainerStarted","Data":"46d3d3be2260d255fc002b9ac213e908e2c7c9482386cd865f4a9c2d267a6cb0"} Nov 21 14:36:30 crc kubenswrapper[4897]: I1121 14:36:30.045708 4897 generic.go:334] "Generic (PLEG): container finished" podID="522a19f3-71a6-4609-89cd-6142f5f0d375" containerID="8f7788e6ea517fa5675fbd9e32ba8202c7b3ea53d2a83293a84c4235cad59e54" exitCode=0 Nov 21 14:36:30 crc kubenswrapper[4897]: I1121 14:36:30.045787 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-dgr4b" event={"ID":"522a19f3-71a6-4609-89cd-6142f5f0d375","Type":"ContainerDied","Data":"8f7788e6ea517fa5675fbd9e32ba8202c7b3ea53d2a83293a84c4235cad59e54"} Nov 21 14:36:30 crc kubenswrapper[4897]: I1121 14:36:30.050468 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"76d57203-0f85-48aa-87b0-a56af184da2d","Type":"ContainerStarted","Data":"2e36427e415786092f0fb7a4cc3edfd365ffbaf6073ac34dfa737a9f97007dc1"} Nov 21 14:36:30 crc kubenswrapper[4897]: I1121 14:36:30.050760 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 21 14:36:30 crc kubenswrapper[4897]: I1121 14:36:30.100703 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.162951673 podStartE2EDuration="14.100681916s" podCreationTimestamp="2025-11-21 14:36:16 +0000 UTC" firstStartedPulling="2025-11-21 14:36:17.857143053 +0000 UTC m=+1655.141736538" lastFinishedPulling="2025-11-21 14:36:28.794873306 +0000 UTC m=+1666.079466781" observedRunningTime="2025-11-21 14:36:30.082410274 +0000 UTC m=+1667.367003759" watchObservedRunningTime="2025-11-21 14:36:30.100681916 +0000 UTC m=+1667.385275391" Nov 21 14:36:30 crc kubenswrapper[4897]: I1121 14:36:30.580566 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 21 14:36:30 crc kubenswrapper[4897]: I1121 14:36:30.580606 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 21 14:36:31 crc kubenswrapper[4897]: I1121 14:36:31.587777 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-dgr4b" Nov 21 14:36:31 crc kubenswrapper[4897]: I1121 14:36:31.598821 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="f688a5e9-9393-4dee-b49a-22b6c75ecfaf" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.255:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 21 14:36:31 crc kubenswrapper[4897]: I1121 14:36:31.599143 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="f688a5e9-9393-4dee-b49a-22b6c75ecfaf" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.255:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 21 14:36:31 crc kubenswrapper[4897]: E1121 14:36:31.732368 4897 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podafa69523_63c4_4a94_b94a_61c1343ef10c.slice/crio-conmon-6cb3731d3191a0096a5051532c43fd96973f7663ef9d3571618a165b3dd1afca.scope\": RecentStats: unable to find data in memory cache]" Nov 21 14:36:31 crc kubenswrapper[4897]: I1121 14:36:31.779337 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/522a19f3-71a6-4609-89cd-6142f5f0d375-combined-ca-bundle\") pod \"522a19f3-71a6-4609-89cd-6142f5f0d375\" (UID: \"522a19f3-71a6-4609-89cd-6142f5f0d375\") " Nov 21 14:36:31 crc kubenswrapper[4897]: I1121 14:36:31.779467 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/522a19f3-71a6-4609-89cd-6142f5f0d375-scripts\") pod \"522a19f3-71a6-4609-89cd-6142f5f0d375\" (UID: \"522a19f3-71a6-4609-89cd-6142f5f0d375\") " Nov 21 14:36:31 crc kubenswrapper[4897]: I1121 14:36:31.779556 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z4bpn\" (UniqueName: \"kubernetes.io/projected/522a19f3-71a6-4609-89cd-6142f5f0d375-kube-api-access-z4bpn\") pod \"522a19f3-71a6-4609-89cd-6142f5f0d375\" (UID: \"522a19f3-71a6-4609-89cd-6142f5f0d375\") " Nov 21 14:36:31 crc kubenswrapper[4897]: I1121 14:36:31.779590 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/522a19f3-71a6-4609-89cd-6142f5f0d375-config-data\") pod \"522a19f3-71a6-4609-89cd-6142f5f0d375\" (UID: \"522a19f3-71a6-4609-89cd-6142f5f0d375\") " Nov 21 14:36:31 crc kubenswrapper[4897]: I1121 14:36:31.793742 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/522a19f3-71a6-4609-89cd-6142f5f0d375-scripts" (OuterVolumeSpecName: "scripts") pod "522a19f3-71a6-4609-89cd-6142f5f0d375" (UID: "522a19f3-71a6-4609-89cd-6142f5f0d375"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:36:31 crc kubenswrapper[4897]: I1121 14:36:31.796561 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/522a19f3-71a6-4609-89cd-6142f5f0d375-kube-api-access-z4bpn" (OuterVolumeSpecName: "kube-api-access-z4bpn") pod "522a19f3-71a6-4609-89cd-6142f5f0d375" (UID: "522a19f3-71a6-4609-89cd-6142f5f0d375"). InnerVolumeSpecName "kube-api-access-z4bpn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:36:31 crc kubenswrapper[4897]: I1121 14:36:31.847825 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/522a19f3-71a6-4609-89cd-6142f5f0d375-config-data" (OuterVolumeSpecName: "config-data") pod "522a19f3-71a6-4609-89cd-6142f5f0d375" (UID: "522a19f3-71a6-4609-89cd-6142f5f0d375"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:36:31 crc kubenswrapper[4897]: I1121 14:36:31.869866 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/522a19f3-71a6-4609-89cd-6142f5f0d375-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "522a19f3-71a6-4609-89cd-6142f5f0d375" (UID: "522a19f3-71a6-4609-89cd-6142f5f0d375"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:36:31 crc kubenswrapper[4897]: I1121 14:36:31.884483 4897 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/522a19f3-71a6-4609-89cd-6142f5f0d375-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 21 14:36:31 crc kubenswrapper[4897]: I1121 14:36:31.884528 4897 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/522a19f3-71a6-4609-89cd-6142f5f0d375-scripts\") on node \"crc\" DevicePath \"\"" Nov 21 14:36:31 crc kubenswrapper[4897]: I1121 14:36:31.884538 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z4bpn\" (UniqueName: \"kubernetes.io/projected/522a19f3-71a6-4609-89cd-6142f5f0d375-kube-api-access-z4bpn\") on node \"crc\" DevicePath \"\"" Nov 21 14:36:31 crc kubenswrapper[4897]: I1121 14:36:31.884550 4897 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/522a19f3-71a6-4609-89cd-6142f5f0d375-config-data\") on node \"crc\" DevicePath \"\"" Nov 21 14:36:32 crc kubenswrapper[4897]: I1121 14:36:32.084849 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-dgr4b" event={"ID":"522a19f3-71a6-4609-89cd-6142f5f0d375","Type":"ContainerDied","Data":"b722d0d29f35a524443c07c64761f6fa719b334924c25016e5c245313081481f"} Nov 21 14:36:32 crc kubenswrapper[4897]: I1121 14:36:32.084894 4897 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b722d0d29f35a524443c07c64761f6fa719b334924c25016e5c245313081481f" Nov 21 14:36:32 crc kubenswrapper[4897]: I1121 14:36:32.085291 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-dgr4b" Nov 21 14:36:32 crc kubenswrapper[4897]: I1121 14:36:32.088923 4897 scope.go:117] "RemoveContainer" containerID="c68f1b003d12b4a2964169a85a04e3f0272d86c891f52e6bebd52f3781520d1a" Nov 21 14:36:32 crc kubenswrapper[4897]: E1121 14:36:32.089252 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 14:36:32 crc kubenswrapper[4897]: I1121 14:36:32.284603 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 21 14:36:32 crc kubenswrapper[4897]: I1121 14:36:32.285258 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="f688a5e9-9393-4dee-b49a-22b6c75ecfaf" containerName="nova-api-log" containerID="cri-o://5dc39c26fd2373cc55c8e7343e92edce69c4b0dde2cd59ca4fa105ccca9eae4c" gracePeriod=30 Nov 21 14:36:32 crc kubenswrapper[4897]: I1121 14:36:32.285354 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="f688a5e9-9393-4dee-b49a-22b6c75ecfaf" containerName="nova-api-api" containerID="cri-o://d4a7106c42b29ced91fa9c22a9fab018391f685ce1bad91f2ea315fbc57e9bd9" gracePeriod=30 Nov 21 14:36:32 crc kubenswrapper[4897]: I1121 14:36:32.303825 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 21 14:36:32 crc kubenswrapper[4897]: I1121 14:36:32.304155 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="7ee39fa1-44f9-46e8-9206-39d5ad1e2e6c" containerName="nova-scheduler-scheduler" containerID="cri-o://fdaf635e274efe1f2c24cf94ed630f0e027ba9d8b0078089b84a1f4dfb4949ba" gracePeriod=30 Nov 21 14:36:32 crc kubenswrapper[4897]: I1121 14:36:32.364672 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 21 14:36:32 crc kubenswrapper[4897]: I1121 14:36:32.364920 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="6810bdc5-077d-44fc-a616-3582d0929d42" containerName="nova-metadata-log" containerID="cri-o://8fb72facf97f0167bacd8eb084a6f636d166cfb4100b89a35018d53ad964a621" gracePeriod=30 Nov 21 14:36:32 crc kubenswrapper[4897]: I1121 14:36:32.365052 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="6810bdc5-077d-44fc-a616-3582d0929d42" containerName="nova-metadata-metadata" containerID="cri-o://f4856114b27cbe08f59a71c5c4ba4eb464c61da14a4b1fbfd1dc2d65a22619ae" gracePeriod=30 Nov 21 14:36:33 crc kubenswrapper[4897]: I1121 14:36:33.104765 4897 generic.go:334] "Generic (PLEG): container finished" podID="6810bdc5-077d-44fc-a616-3582d0929d42" containerID="8fb72facf97f0167bacd8eb084a6f636d166cfb4100b89a35018d53ad964a621" exitCode=143 Nov 21 14:36:33 crc kubenswrapper[4897]: I1121 14:36:33.104861 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6810bdc5-077d-44fc-a616-3582d0929d42","Type":"ContainerDied","Data":"8fb72facf97f0167bacd8eb084a6f636d166cfb4100b89a35018d53ad964a621"} Nov 21 14:36:33 crc kubenswrapper[4897]: I1121 14:36:33.109473 4897 generic.go:334] "Generic (PLEG): container finished" podID="f688a5e9-9393-4dee-b49a-22b6c75ecfaf" containerID="5dc39c26fd2373cc55c8e7343e92edce69c4b0dde2cd59ca4fa105ccca9eae4c" exitCode=143 Nov 21 14:36:33 crc kubenswrapper[4897]: I1121 14:36:33.109594 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f688a5e9-9393-4dee-b49a-22b6c75ecfaf","Type":"ContainerDied","Data":"5dc39c26fd2373cc55c8e7343e92edce69c4b0dde2cd59ca4fa105ccca9eae4c"} Nov 21 14:36:34 crc kubenswrapper[4897]: E1121 14:36:34.464973 4897 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podafa69523_63c4_4a94_b94a_61c1343ef10c.slice/crio-conmon-6cb3731d3191a0096a5051532c43fd96973f7663ef9d3571618a165b3dd1afca.scope\": RecentStats: unable to find data in memory cache]" Nov 21 14:36:34 crc kubenswrapper[4897]: E1121 14:36:34.876573 4897 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of fdaf635e274efe1f2c24cf94ed630f0e027ba9d8b0078089b84a1f4dfb4949ba is running failed: container process not found" containerID="fdaf635e274efe1f2c24cf94ed630f0e027ba9d8b0078089b84a1f4dfb4949ba" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 21 14:36:34 crc kubenswrapper[4897]: E1121 14:36:34.877493 4897 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of fdaf635e274efe1f2c24cf94ed630f0e027ba9d8b0078089b84a1f4dfb4949ba is running failed: container process not found" containerID="fdaf635e274efe1f2c24cf94ed630f0e027ba9d8b0078089b84a1f4dfb4949ba" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 21 14:36:34 crc kubenswrapper[4897]: E1121 14:36:34.878024 4897 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of fdaf635e274efe1f2c24cf94ed630f0e027ba9d8b0078089b84a1f4dfb4949ba is running failed: container process not found" containerID="fdaf635e274efe1f2c24cf94ed630f0e027ba9d8b0078089b84a1f4dfb4949ba" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 21 14:36:34 crc kubenswrapper[4897]: E1121 14:36:34.878059 4897 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of fdaf635e274efe1f2c24cf94ed630f0e027ba9d8b0078089b84a1f4dfb4949ba is running failed: container process not found" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="7ee39fa1-44f9-46e8-9206-39d5ad1e2e6c" containerName="nova-scheduler-scheduler" Nov 21 14:36:35 crc kubenswrapper[4897]: I1121 14:36:35.186756 4897 generic.go:334] "Generic (PLEG): container finished" podID="7ee39fa1-44f9-46e8-9206-39d5ad1e2e6c" containerID="fdaf635e274efe1f2c24cf94ed630f0e027ba9d8b0078089b84a1f4dfb4949ba" exitCode=0 Nov 21 14:36:35 crc kubenswrapper[4897]: I1121 14:36:35.186805 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"7ee39fa1-44f9-46e8-9206-39d5ad1e2e6c","Type":"ContainerDied","Data":"fdaf635e274efe1f2c24cf94ed630f0e027ba9d8b0078089b84a1f4dfb4949ba"} Nov 21 14:36:35 crc kubenswrapper[4897]: I1121 14:36:35.318101 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-dtwql" Nov 21 14:36:35 crc kubenswrapper[4897]: I1121 14:36:35.370853 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 21 14:36:35 crc kubenswrapper[4897]: I1121 14:36:35.388520 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l4mm4\" (UniqueName: \"kubernetes.io/projected/7ee39fa1-44f9-46e8-9206-39d5ad1e2e6c-kube-api-access-l4mm4\") pod \"7ee39fa1-44f9-46e8-9206-39d5ad1e2e6c\" (UID: \"7ee39fa1-44f9-46e8-9206-39d5ad1e2e6c\") " Nov 21 14:36:35 crc kubenswrapper[4897]: I1121 14:36:35.388609 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ee39fa1-44f9-46e8-9206-39d5ad1e2e6c-config-data\") pod \"7ee39fa1-44f9-46e8-9206-39d5ad1e2e6c\" (UID: \"7ee39fa1-44f9-46e8-9206-39d5ad1e2e6c\") " Nov 21 14:36:35 crc kubenswrapper[4897]: I1121 14:36:35.388685 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ee39fa1-44f9-46e8-9206-39d5ad1e2e6c-combined-ca-bundle\") pod \"7ee39fa1-44f9-46e8-9206-39d5ad1e2e6c\" (UID: \"7ee39fa1-44f9-46e8-9206-39d5ad1e2e6c\") " Nov 21 14:36:35 crc kubenswrapper[4897]: I1121 14:36:35.395208 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-dtwql" Nov 21 14:36:35 crc kubenswrapper[4897]: I1121 14:36:35.396048 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ee39fa1-44f9-46e8-9206-39d5ad1e2e6c-kube-api-access-l4mm4" (OuterVolumeSpecName: "kube-api-access-l4mm4") pod "7ee39fa1-44f9-46e8-9206-39d5ad1e2e6c" (UID: "7ee39fa1-44f9-46e8-9206-39d5ad1e2e6c"). InnerVolumeSpecName "kube-api-access-l4mm4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:36:35 crc kubenswrapper[4897]: I1121 14:36:35.445385 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ee39fa1-44f9-46e8-9206-39d5ad1e2e6c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7ee39fa1-44f9-46e8-9206-39d5ad1e2e6c" (UID: "7ee39fa1-44f9-46e8-9206-39d5ad1e2e6c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:36:35 crc kubenswrapper[4897]: I1121 14:36:35.461325 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ee39fa1-44f9-46e8-9206-39d5ad1e2e6c-config-data" (OuterVolumeSpecName: "config-data") pod "7ee39fa1-44f9-46e8-9206-39d5ad1e2e6c" (UID: "7ee39fa1-44f9-46e8-9206-39d5ad1e2e6c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:36:35 crc kubenswrapper[4897]: I1121 14:36:35.495373 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l4mm4\" (UniqueName: \"kubernetes.io/projected/7ee39fa1-44f9-46e8-9206-39d5ad1e2e6c-kube-api-access-l4mm4\") on node \"crc\" DevicePath \"\"" Nov 21 14:36:35 crc kubenswrapper[4897]: I1121 14:36:35.496234 4897 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ee39fa1-44f9-46e8-9206-39d5ad1e2e6c-config-data\") on node \"crc\" DevicePath \"\"" Nov 21 14:36:35 crc kubenswrapper[4897]: I1121 14:36:35.496307 4897 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ee39fa1-44f9-46e8-9206-39d5ad1e2e6c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 21 14:36:35 crc kubenswrapper[4897]: I1121 14:36:35.511128 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="6810bdc5-077d-44fc-a616-3582d0929d42" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.247:8775/\": read tcp 10.217.0.2:36300->10.217.0.247:8775: read: connection reset by peer" Nov 21 14:36:35 crc kubenswrapper[4897]: I1121 14:36:35.511177 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="6810bdc5-077d-44fc-a616-3582d0929d42" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.247:8775/\": read tcp 10.217.0.2:36302->10.217.0.247:8775: read: connection reset by peer" Nov 21 14:36:35 crc kubenswrapper[4897]: I1121 14:36:35.572013 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dtwql"] Nov 21 14:36:36 crc kubenswrapper[4897]: I1121 14:36:36.202794 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"7ee39fa1-44f9-46e8-9206-39d5ad1e2e6c","Type":"ContainerDied","Data":"ba416c5a2341719fb456ec3197d8e3416e6a5195c4e1c7cc12249e52f9daa04a"} Nov 21 14:36:36 crc kubenswrapper[4897]: I1121 14:36:36.203148 4897 scope.go:117] "RemoveContainer" containerID="fdaf635e274efe1f2c24cf94ed630f0e027ba9d8b0078089b84a1f4dfb4949ba" Nov 21 14:36:36 crc kubenswrapper[4897]: I1121 14:36:36.202812 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 21 14:36:36 crc kubenswrapper[4897]: I1121 14:36:36.205029 4897 generic.go:334] "Generic (PLEG): container finished" podID="6810bdc5-077d-44fc-a616-3582d0929d42" containerID="f4856114b27cbe08f59a71c5c4ba4eb464c61da14a4b1fbfd1dc2d65a22619ae" exitCode=0 Nov 21 14:36:36 crc kubenswrapper[4897]: I1121 14:36:36.205088 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6810bdc5-077d-44fc-a616-3582d0929d42","Type":"ContainerDied","Data":"f4856114b27cbe08f59a71c5c4ba4eb464c61da14a4b1fbfd1dc2d65a22619ae"} Nov 21 14:36:36 crc kubenswrapper[4897]: I1121 14:36:36.237742 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 21 14:36:36 crc kubenswrapper[4897]: I1121 14:36:36.256242 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Nov 21 14:36:36 crc kubenswrapper[4897]: I1121 14:36:36.275737 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 21 14:36:36 crc kubenswrapper[4897]: E1121 14:36:36.276728 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="522a19f3-71a6-4609-89cd-6142f5f0d375" containerName="nova-manage" Nov 21 14:36:36 crc kubenswrapper[4897]: I1121 14:36:36.276757 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="522a19f3-71a6-4609-89cd-6142f5f0d375" containerName="nova-manage" Nov 21 14:36:36 crc kubenswrapper[4897]: E1121 14:36:36.277029 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ee39fa1-44f9-46e8-9206-39d5ad1e2e6c" containerName="nova-scheduler-scheduler" Nov 21 14:36:36 crc kubenswrapper[4897]: I1121 14:36:36.279725 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ee39fa1-44f9-46e8-9206-39d5ad1e2e6c" containerName="nova-scheduler-scheduler" Nov 21 14:36:36 crc kubenswrapper[4897]: I1121 14:36:36.280048 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ee39fa1-44f9-46e8-9206-39d5ad1e2e6c" containerName="nova-scheduler-scheduler" Nov 21 14:36:36 crc kubenswrapper[4897]: I1121 14:36:36.280092 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="522a19f3-71a6-4609-89cd-6142f5f0d375" containerName="nova-manage" Nov 21 14:36:36 crc kubenswrapper[4897]: I1121 14:36:36.281346 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 21 14:36:36 crc kubenswrapper[4897]: I1121 14:36:36.285588 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 21 14:36:36 crc kubenswrapper[4897]: I1121 14:36:36.290799 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 21 14:36:36 crc kubenswrapper[4897]: I1121 14:36:36.314642 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6784aaa9-6208-4b98-be6b-0ce286e9b860-config-data\") pod \"nova-scheduler-0\" (UID: \"6784aaa9-6208-4b98-be6b-0ce286e9b860\") " pod="openstack/nova-scheduler-0" Nov 21 14:36:36 crc kubenswrapper[4897]: I1121 14:36:36.314713 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vkpss\" (UniqueName: \"kubernetes.io/projected/6784aaa9-6208-4b98-be6b-0ce286e9b860-kube-api-access-vkpss\") pod \"nova-scheduler-0\" (UID: \"6784aaa9-6208-4b98-be6b-0ce286e9b860\") " pod="openstack/nova-scheduler-0" Nov 21 14:36:36 crc kubenswrapper[4897]: I1121 14:36:36.314914 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6784aaa9-6208-4b98-be6b-0ce286e9b860-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"6784aaa9-6208-4b98-be6b-0ce286e9b860\") " pod="openstack/nova-scheduler-0" Nov 21 14:36:36 crc kubenswrapper[4897]: I1121 14:36:36.417526 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6784aaa9-6208-4b98-be6b-0ce286e9b860-config-data\") pod \"nova-scheduler-0\" (UID: \"6784aaa9-6208-4b98-be6b-0ce286e9b860\") " pod="openstack/nova-scheduler-0" Nov 21 14:36:36 crc kubenswrapper[4897]: I1121 14:36:36.417607 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vkpss\" (UniqueName: \"kubernetes.io/projected/6784aaa9-6208-4b98-be6b-0ce286e9b860-kube-api-access-vkpss\") pod \"nova-scheduler-0\" (UID: \"6784aaa9-6208-4b98-be6b-0ce286e9b860\") " pod="openstack/nova-scheduler-0" Nov 21 14:36:36 crc kubenswrapper[4897]: I1121 14:36:36.417672 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6784aaa9-6208-4b98-be6b-0ce286e9b860-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"6784aaa9-6208-4b98-be6b-0ce286e9b860\") " pod="openstack/nova-scheduler-0" Nov 21 14:36:36 crc kubenswrapper[4897]: I1121 14:36:36.422774 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6784aaa9-6208-4b98-be6b-0ce286e9b860-config-data\") pod \"nova-scheduler-0\" (UID: \"6784aaa9-6208-4b98-be6b-0ce286e9b860\") " pod="openstack/nova-scheduler-0" Nov 21 14:36:36 crc kubenswrapper[4897]: I1121 14:36:36.425225 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6784aaa9-6208-4b98-be6b-0ce286e9b860-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"6784aaa9-6208-4b98-be6b-0ce286e9b860\") " pod="openstack/nova-scheduler-0" Nov 21 14:36:36 crc kubenswrapper[4897]: I1121 14:36:36.468352 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vkpss\" (UniqueName: \"kubernetes.io/projected/6784aaa9-6208-4b98-be6b-0ce286e9b860-kube-api-access-vkpss\") pod \"nova-scheduler-0\" (UID: \"6784aaa9-6208-4b98-be6b-0ce286e9b860\") " pod="openstack/nova-scheduler-0" Nov 21 14:36:36 crc kubenswrapper[4897]: I1121 14:36:36.617787 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 21 14:36:36 crc kubenswrapper[4897]: I1121 14:36:36.749905 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 21 14:36:36 crc kubenswrapper[4897]: I1121 14:36:36.838143 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6wvhk\" (UniqueName: \"kubernetes.io/projected/6810bdc5-077d-44fc-a616-3582d0929d42-kube-api-access-6wvhk\") pod \"6810bdc5-077d-44fc-a616-3582d0929d42\" (UID: \"6810bdc5-077d-44fc-a616-3582d0929d42\") " Nov 21 14:36:36 crc kubenswrapper[4897]: I1121 14:36:36.838221 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6810bdc5-077d-44fc-a616-3582d0929d42-nova-metadata-tls-certs\") pod \"6810bdc5-077d-44fc-a616-3582d0929d42\" (UID: \"6810bdc5-077d-44fc-a616-3582d0929d42\") " Nov 21 14:36:36 crc kubenswrapper[4897]: I1121 14:36:36.838282 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6810bdc5-077d-44fc-a616-3582d0929d42-config-data\") pod \"6810bdc5-077d-44fc-a616-3582d0929d42\" (UID: \"6810bdc5-077d-44fc-a616-3582d0929d42\") " Nov 21 14:36:36 crc kubenswrapper[4897]: I1121 14:36:36.838619 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6810bdc5-077d-44fc-a616-3582d0929d42-logs\") pod \"6810bdc5-077d-44fc-a616-3582d0929d42\" (UID: \"6810bdc5-077d-44fc-a616-3582d0929d42\") " Nov 21 14:36:36 crc kubenswrapper[4897]: I1121 14:36:36.838731 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6810bdc5-077d-44fc-a616-3582d0929d42-combined-ca-bundle\") pod \"6810bdc5-077d-44fc-a616-3582d0929d42\" (UID: \"6810bdc5-077d-44fc-a616-3582d0929d42\") " Nov 21 14:36:36 crc kubenswrapper[4897]: I1121 14:36:36.840884 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6810bdc5-077d-44fc-a616-3582d0929d42-logs" (OuterVolumeSpecName: "logs") pod "6810bdc5-077d-44fc-a616-3582d0929d42" (UID: "6810bdc5-077d-44fc-a616-3582d0929d42"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:36:36 crc kubenswrapper[4897]: I1121 14:36:36.848025 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6810bdc5-077d-44fc-a616-3582d0929d42-kube-api-access-6wvhk" (OuterVolumeSpecName: "kube-api-access-6wvhk") pod "6810bdc5-077d-44fc-a616-3582d0929d42" (UID: "6810bdc5-077d-44fc-a616-3582d0929d42"). InnerVolumeSpecName "kube-api-access-6wvhk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:36:36 crc kubenswrapper[4897]: I1121 14:36:36.890963 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6810bdc5-077d-44fc-a616-3582d0929d42-config-data" (OuterVolumeSpecName: "config-data") pod "6810bdc5-077d-44fc-a616-3582d0929d42" (UID: "6810bdc5-077d-44fc-a616-3582d0929d42"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:36:36 crc kubenswrapper[4897]: I1121 14:36:36.893715 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6810bdc5-077d-44fc-a616-3582d0929d42-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6810bdc5-077d-44fc-a616-3582d0929d42" (UID: "6810bdc5-077d-44fc-a616-3582d0929d42"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:36:36 crc kubenswrapper[4897]: I1121 14:36:36.911025 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6810bdc5-077d-44fc-a616-3582d0929d42-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "6810bdc5-077d-44fc-a616-3582d0929d42" (UID: "6810bdc5-077d-44fc-a616-3582d0929d42"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:36:36 crc kubenswrapper[4897]: I1121 14:36:36.942769 4897 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6810bdc5-077d-44fc-a616-3582d0929d42-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 21 14:36:36 crc kubenswrapper[4897]: I1121 14:36:36.942829 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6wvhk\" (UniqueName: \"kubernetes.io/projected/6810bdc5-077d-44fc-a616-3582d0929d42-kube-api-access-6wvhk\") on node \"crc\" DevicePath \"\"" Nov 21 14:36:36 crc kubenswrapper[4897]: I1121 14:36:36.942843 4897 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6810bdc5-077d-44fc-a616-3582d0929d42-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 21 14:36:36 crc kubenswrapper[4897]: I1121 14:36:36.942852 4897 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6810bdc5-077d-44fc-a616-3582d0929d42-config-data\") on node \"crc\" DevicePath \"\"" Nov 21 14:36:36 crc kubenswrapper[4897]: I1121 14:36:36.942861 4897 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6810bdc5-077d-44fc-a616-3582d0929d42-logs\") on node \"crc\" DevicePath \"\"" Nov 21 14:36:37 crc kubenswrapper[4897]: I1121 14:36:37.105745 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 21 14:36:37 crc kubenswrapper[4897]: W1121 14:36:37.106335 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6784aaa9_6208_4b98_be6b_0ce286e9b860.slice/crio-00f028b169e16e07e1e93a758ff90689a71159029804b0e6d6f0dd3e6c782862 WatchSource:0}: Error finding container 00f028b169e16e07e1e93a758ff90689a71159029804b0e6d6f0dd3e6c782862: Status 404 returned error can't find the container with id 00f028b169e16e07e1e93a758ff90689a71159029804b0e6d6f0dd3e6c782862 Nov 21 14:36:37 crc kubenswrapper[4897]: I1121 14:36:37.245645 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"6784aaa9-6208-4b98-be6b-0ce286e9b860","Type":"ContainerStarted","Data":"00f028b169e16e07e1e93a758ff90689a71159029804b0e6d6f0dd3e6c782862"} Nov 21 14:36:37 crc kubenswrapper[4897]: I1121 14:36:37.252348 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6810bdc5-077d-44fc-a616-3582d0929d42","Type":"ContainerDied","Data":"aad9611cad00d40b8f5680fbd9d96866afb7427c68d5f8b904be2beee52320bf"} Nov 21 14:36:37 crc kubenswrapper[4897]: I1121 14:36:37.252373 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-dtwql" podUID="6d3cad20-fb17-4726-b8fa-9fd5303b5eea" containerName="registry-server" containerID="cri-o://5ee18c9557dc1781cb41a1a2df7bdb7b7161886e8a11d31888240d1445a4be95" gracePeriod=2 Nov 21 14:36:37 crc kubenswrapper[4897]: I1121 14:36:37.252414 4897 scope.go:117] "RemoveContainer" containerID="f4856114b27cbe08f59a71c5c4ba4eb464c61da14a4b1fbfd1dc2d65a22619ae" Nov 21 14:36:37 crc kubenswrapper[4897]: I1121 14:36:37.252679 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 21 14:36:37 crc kubenswrapper[4897]: I1121 14:36:37.337205 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 21 14:36:37 crc kubenswrapper[4897]: I1121 14:36:37.353343 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 21 14:36:37 crc kubenswrapper[4897]: I1121 14:36:37.369689 4897 scope.go:117] "RemoveContainer" containerID="8fb72facf97f0167bacd8eb084a6f636d166cfb4100b89a35018d53ad964a621" Nov 21 14:36:37 crc kubenswrapper[4897]: I1121 14:36:37.374406 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 21 14:36:37 crc kubenswrapper[4897]: E1121 14:36:37.374967 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6810bdc5-077d-44fc-a616-3582d0929d42" containerName="nova-metadata-metadata" Nov 21 14:36:37 crc kubenswrapper[4897]: I1121 14:36:37.374992 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="6810bdc5-077d-44fc-a616-3582d0929d42" containerName="nova-metadata-metadata" Nov 21 14:36:37 crc kubenswrapper[4897]: E1121 14:36:37.375015 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6810bdc5-077d-44fc-a616-3582d0929d42" containerName="nova-metadata-log" Nov 21 14:36:37 crc kubenswrapper[4897]: I1121 14:36:37.375023 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="6810bdc5-077d-44fc-a616-3582d0929d42" containerName="nova-metadata-log" Nov 21 14:36:37 crc kubenswrapper[4897]: I1121 14:36:37.375263 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="6810bdc5-077d-44fc-a616-3582d0929d42" containerName="nova-metadata-log" Nov 21 14:36:37 crc kubenswrapper[4897]: I1121 14:36:37.375294 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="6810bdc5-077d-44fc-a616-3582d0929d42" containerName="nova-metadata-metadata" Nov 21 14:36:37 crc kubenswrapper[4897]: I1121 14:36:37.376750 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 21 14:36:37 crc kubenswrapper[4897]: I1121 14:36:37.385379 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 21 14:36:37 crc kubenswrapper[4897]: I1121 14:36:37.386095 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 21 14:36:37 crc kubenswrapper[4897]: I1121 14:36:37.386397 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 21 14:36:37 crc kubenswrapper[4897]: I1121 14:36:37.454178 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e7201326-967f-4b69-b80c-be6a5ecfdc80-logs\") pod \"nova-metadata-0\" (UID: \"e7201326-967f-4b69-b80c-be6a5ecfdc80\") " pod="openstack/nova-metadata-0" Nov 21 14:36:37 crc kubenswrapper[4897]: I1121 14:36:37.454271 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/e7201326-967f-4b69-b80c-be6a5ecfdc80-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"e7201326-967f-4b69-b80c-be6a5ecfdc80\") " pod="openstack/nova-metadata-0" Nov 21 14:36:37 crc kubenswrapper[4897]: I1121 14:36:37.454317 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7201326-967f-4b69-b80c-be6a5ecfdc80-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e7201326-967f-4b69-b80c-be6a5ecfdc80\") " pod="openstack/nova-metadata-0" Nov 21 14:36:37 crc kubenswrapper[4897]: I1121 14:36:37.454332 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4l7pj\" (UniqueName: \"kubernetes.io/projected/e7201326-967f-4b69-b80c-be6a5ecfdc80-kube-api-access-4l7pj\") pod \"nova-metadata-0\" (UID: \"e7201326-967f-4b69-b80c-be6a5ecfdc80\") " pod="openstack/nova-metadata-0" Nov 21 14:36:37 crc kubenswrapper[4897]: I1121 14:36:37.454435 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e7201326-967f-4b69-b80c-be6a5ecfdc80-config-data\") pod \"nova-metadata-0\" (UID: \"e7201326-967f-4b69-b80c-be6a5ecfdc80\") " pod="openstack/nova-metadata-0" Nov 21 14:36:37 crc kubenswrapper[4897]: I1121 14:36:37.556965 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/e7201326-967f-4b69-b80c-be6a5ecfdc80-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"e7201326-967f-4b69-b80c-be6a5ecfdc80\") " pod="openstack/nova-metadata-0" Nov 21 14:36:37 crc kubenswrapper[4897]: I1121 14:36:37.557035 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7201326-967f-4b69-b80c-be6a5ecfdc80-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e7201326-967f-4b69-b80c-be6a5ecfdc80\") " pod="openstack/nova-metadata-0" Nov 21 14:36:37 crc kubenswrapper[4897]: I1121 14:36:37.557055 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4l7pj\" (UniqueName: \"kubernetes.io/projected/e7201326-967f-4b69-b80c-be6a5ecfdc80-kube-api-access-4l7pj\") pod \"nova-metadata-0\" (UID: \"e7201326-967f-4b69-b80c-be6a5ecfdc80\") " pod="openstack/nova-metadata-0" Nov 21 14:36:37 crc kubenswrapper[4897]: I1121 14:36:37.557164 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e7201326-967f-4b69-b80c-be6a5ecfdc80-config-data\") pod \"nova-metadata-0\" (UID: \"e7201326-967f-4b69-b80c-be6a5ecfdc80\") " pod="openstack/nova-metadata-0" Nov 21 14:36:37 crc kubenswrapper[4897]: I1121 14:36:37.557221 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e7201326-967f-4b69-b80c-be6a5ecfdc80-logs\") pod \"nova-metadata-0\" (UID: \"e7201326-967f-4b69-b80c-be6a5ecfdc80\") " pod="openstack/nova-metadata-0" Nov 21 14:36:37 crc kubenswrapper[4897]: I1121 14:36:37.557683 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e7201326-967f-4b69-b80c-be6a5ecfdc80-logs\") pod \"nova-metadata-0\" (UID: \"e7201326-967f-4b69-b80c-be6a5ecfdc80\") " pod="openstack/nova-metadata-0" Nov 21 14:36:37 crc kubenswrapper[4897]: I1121 14:36:37.563026 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/e7201326-967f-4b69-b80c-be6a5ecfdc80-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"e7201326-967f-4b69-b80c-be6a5ecfdc80\") " pod="openstack/nova-metadata-0" Nov 21 14:36:37 crc kubenswrapper[4897]: I1121 14:36:37.563318 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7201326-967f-4b69-b80c-be6a5ecfdc80-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e7201326-967f-4b69-b80c-be6a5ecfdc80\") " pod="openstack/nova-metadata-0" Nov 21 14:36:37 crc kubenswrapper[4897]: I1121 14:36:37.563400 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e7201326-967f-4b69-b80c-be6a5ecfdc80-config-data\") pod \"nova-metadata-0\" (UID: \"e7201326-967f-4b69-b80c-be6a5ecfdc80\") " pod="openstack/nova-metadata-0" Nov 21 14:36:37 crc kubenswrapper[4897]: I1121 14:36:37.575148 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4l7pj\" (UniqueName: \"kubernetes.io/projected/e7201326-967f-4b69-b80c-be6a5ecfdc80-kube-api-access-4l7pj\") pod \"nova-metadata-0\" (UID: \"e7201326-967f-4b69-b80c-be6a5ecfdc80\") " pod="openstack/nova-metadata-0" Nov 21 14:36:37 crc kubenswrapper[4897]: I1121 14:36:37.756128 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 21 14:36:38 crc kubenswrapper[4897]: I1121 14:36:38.123050 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6810bdc5-077d-44fc-a616-3582d0929d42" path="/var/lib/kubelet/pods/6810bdc5-077d-44fc-a616-3582d0929d42/volumes" Nov 21 14:36:38 crc kubenswrapper[4897]: I1121 14:36:38.124091 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ee39fa1-44f9-46e8-9206-39d5ad1e2e6c" path="/var/lib/kubelet/pods/7ee39fa1-44f9-46e8-9206-39d5ad1e2e6c/volumes" Nov 21 14:36:38 crc kubenswrapper[4897]: I1121 14:36:38.238682 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 21 14:36:38 crc kubenswrapper[4897]: I1121 14:36:38.276255 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"6784aaa9-6208-4b98-be6b-0ce286e9b860","Type":"ContainerStarted","Data":"4274616b6f0748ad6f3674fe34545cfd58600ebf1330a78c2633300776fbaa5c"} Nov 21 14:36:38 crc kubenswrapper[4897]: I1121 14:36:38.294481 4897 generic.go:334] "Generic (PLEG): container finished" podID="f688a5e9-9393-4dee-b49a-22b6c75ecfaf" containerID="d4a7106c42b29ced91fa9c22a9fab018391f685ce1bad91f2ea315fbc57e9bd9" exitCode=0 Nov 21 14:36:38 crc kubenswrapper[4897]: I1121 14:36:38.294773 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f688a5e9-9393-4dee-b49a-22b6c75ecfaf","Type":"ContainerDied","Data":"d4a7106c42b29ced91fa9c22a9fab018391f685ce1bad91f2ea315fbc57e9bd9"} Nov 21 14:36:38 crc kubenswrapper[4897]: I1121 14:36:38.311622 4897 generic.go:334] "Generic (PLEG): container finished" podID="6d3cad20-fb17-4726-b8fa-9fd5303b5eea" containerID="5ee18c9557dc1781cb41a1a2df7bdb7b7161886e8a11d31888240d1445a4be95" exitCode=0 Nov 21 14:36:38 crc kubenswrapper[4897]: I1121 14:36:38.311764 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dtwql" event={"ID":"6d3cad20-fb17-4726-b8fa-9fd5303b5eea","Type":"ContainerDied","Data":"5ee18c9557dc1781cb41a1a2df7bdb7b7161886e8a11d31888240d1445a4be95"} Nov 21 14:36:38 crc kubenswrapper[4897]: I1121 14:36:38.311809 4897 scope.go:117] "RemoveContainer" containerID="36d1ba740cdac40a73ac8d07a9d49188d3f951e9e6e9797089e9005483687829" Nov 21 14:36:38 crc kubenswrapper[4897]: I1121 14:36:38.313770 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e7201326-967f-4b69-b80c-be6a5ecfdc80","Type":"ContainerStarted","Data":"ced5a5c9884320555a8fa878e81d4ae892d169fd5c5d75e0460571c6c0625e12"} Nov 21 14:36:38 crc kubenswrapper[4897]: I1121 14:36:38.316870 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.316841887 podStartE2EDuration="2.316841887s" podCreationTimestamp="2025-11-21 14:36:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:36:38.300497377 +0000 UTC m=+1675.585090872" watchObservedRunningTime="2025-11-21 14:36:38.316841887 +0000 UTC m=+1675.601435362" Nov 21 14:36:38 crc kubenswrapper[4897]: I1121 14:36:38.943906 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dtwql" Nov 21 14:36:39 crc kubenswrapper[4897]: I1121 14:36:39.006586 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d3cad20-fb17-4726-b8fa-9fd5303b5eea-catalog-content\") pod \"6d3cad20-fb17-4726-b8fa-9fd5303b5eea\" (UID: \"6d3cad20-fb17-4726-b8fa-9fd5303b5eea\") " Nov 21 14:36:39 crc kubenswrapper[4897]: I1121 14:36:39.006875 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d3cad20-fb17-4726-b8fa-9fd5303b5eea-utilities\") pod \"6d3cad20-fb17-4726-b8fa-9fd5303b5eea\" (UID: \"6d3cad20-fb17-4726-b8fa-9fd5303b5eea\") " Nov 21 14:36:39 crc kubenswrapper[4897]: I1121 14:36:39.007064 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7l7sr\" (UniqueName: \"kubernetes.io/projected/6d3cad20-fb17-4726-b8fa-9fd5303b5eea-kube-api-access-7l7sr\") pod \"6d3cad20-fb17-4726-b8fa-9fd5303b5eea\" (UID: \"6d3cad20-fb17-4726-b8fa-9fd5303b5eea\") " Nov 21 14:36:39 crc kubenswrapper[4897]: I1121 14:36:39.007438 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6d3cad20-fb17-4726-b8fa-9fd5303b5eea-utilities" (OuterVolumeSpecName: "utilities") pod "6d3cad20-fb17-4726-b8fa-9fd5303b5eea" (UID: "6d3cad20-fb17-4726-b8fa-9fd5303b5eea"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:36:39 crc kubenswrapper[4897]: I1121 14:36:39.007781 4897 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d3cad20-fb17-4726-b8fa-9fd5303b5eea-utilities\") on node \"crc\" DevicePath \"\"" Nov 21 14:36:39 crc kubenswrapper[4897]: I1121 14:36:39.011094 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d3cad20-fb17-4726-b8fa-9fd5303b5eea-kube-api-access-7l7sr" (OuterVolumeSpecName: "kube-api-access-7l7sr") pod "6d3cad20-fb17-4726-b8fa-9fd5303b5eea" (UID: "6d3cad20-fb17-4726-b8fa-9fd5303b5eea"). InnerVolumeSpecName "kube-api-access-7l7sr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:36:39 crc kubenswrapper[4897]: I1121 14:36:39.016173 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 21 14:36:39 crc kubenswrapper[4897]: I1121 14:36:39.109825 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f688a5e9-9393-4dee-b49a-22b6c75ecfaf-config-data\") pod \"f688a5e9-9393-4dee-b49a-22b6c75ecfaf\" (UID: \"f688a5e9-9393-4dee-b49a-22b6c75ecfaf\") " Nov 21 14:36:39 crc kubenswrapper[4897]: I1121 14:36:39.109881 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rzt2r\" (UniqueName: \"kubernetes.io/projected/f688a5e9-9393-4dee-b49a-22b6c75ecfaf-kube-api-access-rzt2r\") pod \"f688a5e9-9393-4dee-b49a-22b6c75ecfaf\" (UID: \"f688a5e9-9393-4dee-b49a-22b6c75ecfaf\") " Nov 21 14:36:39 crc kubenswrapper[4897]: I1121 14:36:39.109938 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f688a5e9-9393-4dee-b49a-22b6c75ecfaf-combined-ca-bundle\") pod \"f688a5e9-9393-4dee-b49a-22b6c75ecfaf\" (UID: \"f688a5e9-9393-4dee-b49a-22b6c75ecfaf\") " Nov 21 14:36:39 crc kubenswrapper[4897]: I1121 14:36:39.110141 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f688a5e9-9393-4dee-b49a-22b6c75ecfaf-internal-tls-certs\") pod \"f688a5e9-9393-4dee-b49a-22b6c75ecfaf\" (UID: \"f688a5e9-9393-4dee-b49a-22b6c75ecfaf\") " Nov 21 14:36:39 crc kubenswrapper[4897]: I1121 14:36:39.110190 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f688a5e9-9393-4dee-b49a-22b6c75ecfaf-logs\") pod \"f688a5e9-9393-4dee-b49a-22b6c75ecfaf\" (UID: \"f688a5e9-9393-4dee-b49a-22b6c75ecfaf\") " Nov 21 14:36:39 crc kubenswrapper[4897]: I1121 14:36:39.110434 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f688a5e9-9393-4dee-b49a-22b6c75ecfaf-public-tls-certs\") pod \"f688a5e9-9393-4dee-b49a-22b6c75ecfaf\" (UID: \"f688a5e9-9393-4dee-b49a-22b6c75ecfaf\") " Nov 21 14:36:39 crc kubenswrapper[4897]: I1121 14:36:39.111954 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7l7sr\" (UniqueName: \"kubernetes.io/projected/6d3cad20-fb17-4726-b8fa-9fd5303b5eea-kube-api-access-7l7sr\") on node \"crc\" DevicePath \"\"" Nov 21 14:36:39 crc kubenswrapper[4897]: I1121 14:36:39.112260 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f688a5e9-9393-4dee-b49a-22b6c75ecfaf-logs" (OuterVolumeSpecName: "logs") pod "f688a5e9-9393-4dee-b49a-22b6c75ecfaf" (UID: "f688a5e9-9393-4dee-b49a-22b6c75ecfaf"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:36:39 crc kubenswrapper[4897]: I1121 14:36:39.115872 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f688a5e9-9393-4dee-b49a-22b6c75ecfaf-kube-api-access-rzt2r" (OuterVolumeSpecName: "kube-api-access-rzt2r") pod "f688a5e9-9393-4dee-b49a-22b6c75ecfaf" (UID: "f688a5e9-9393-4dee-b49a-22b6c75ecfaf"). InnerVolumeSpecName "kube-api-access-rzt2r". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:36:39 crc kubenswrapper[4897]: I1121 14:36:39.119164 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6d3cad20-fb17-4726-b8fa-9fd5303b5eea-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6d3cad20-fb17-4726-b8fa-9fd5303b5eea" (UID: "6d3cad20-fb17-4726-b8fa-9fd5303b5eea"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:36:39 crc kubenswrapper[4897]: I1121 14:36:39.160588 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f688a5e9-9393-4dee-b49a-22b6c75ecfaf-config-data" (OuterVolumeSpecName: "config-data") pod "f688a5e9-9393-4dee-b49a-22b6c75ecfaf" (UID: "f688a5e9-9393-4dee-b49a-22b6c75ecfaf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:36:39 crc kubenswrapper[4897]: I1121 14:36:39.162725 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f688a5e9-9393-4dee-b49a-22b6c75ecfaf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f688a5e9-9393-4dee-b49a-22b6c75ecfaf" (UID: "f688a5e9-9393-4dee-b49a-22b6c75ecfaf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:36:39 crc kubenswrapper[4897]: I1121 14:36:39.208204 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f688a5e9-9393-4dee-b49a-22b6c75ecfaf-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "f688a5e9-9393-4dee-b49a-22b6c75ecfaf" (UID: "f688a5e9-9393-4dee-b49a-22b6c75ecfaf"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:36:39 crc kubenswrapper[4897]: I1121 14:36:39.214779 4897 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f688a5e9-9393-4dee-b49a-22b6c75ecfaf-config-data\") on node \"crc\" DevicePath \"\"" Nov 21 14:36:39 crc kubenswrapper[4897]: I1121 14:36:39.214802 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rzt2r\" (UniqueName: \"kubernetes.io/projected/f688a5e9-9393-4dee-b49a-22b6c75ecfaf-kube-api-access-rzt2r\") on node \"crc\" DevicePath \"\"" Nov 21 14:36:39 crc kubenswrapper[4897]: I1121 14:36:39.214811 4897 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f688a5e9-9393-4dee-b49a-22b6c75ecfaf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 21 14:36:39 crc kubenswrapper[4897]: I1121 14:36:39.214822 4897 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d3cad20-fb17-4726-b8fa-9fd5303b5eea-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 21 14:36:39 crc kubenswrapper[4897]: I1121 14:36:39.214830 4897 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f688a5e9-9393-4dee-b49a-22b6c75ecfaf-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 21 14:36:39 crc kubenswrapper[4897]: I1121 14:36:39.214839 4897 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f688a5e9-9393-4dee-b49a-22b6c75ecfaf-logs\") on node \"crc\" DevicePath \"\"" Nov 21 14:36:39 crc kubenswrapper[4897]: I1121 14:36:39.214995 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f688a5e9-9393-4dee-b49a-22b6c75ecfaf-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "f688a5e9-9393-4dee-b49a-22b6c75ecfaf" (UID: "f688a5e9-9393-4dee-b49a-22b6c75ecfaf"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:36:39 crc kubenswrapper[4897]: I1121 14:36:39.317901 4897 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f688a5e9-9393-4dee-b49a-22b6c75ecfaf-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 21 14:36:39 crc kubenswrapper[4897]: I1121 14:36:39.338243 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f688a5e9-9393-4dee-b49a-22b6c75ecfaf","Type":"ContainerDied","Data":"b3a5275a5fd8512eb4e39e68ec4657cab31565fa3ef6b7295ef8a28ff8a4ed4f"} Nov 21 14:36:39 crc kubenswrapper[4897]: I1121 14:36:39.338305 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 21 14:36:39 crc kubenswrapper[4897]: I1121 14:36:39.338327 4897 scope.go:117] "RemoveContainer" containerID="d4a7106c42b29ced91fa9c22a9fab018391f685ce1bad91f2ea315fbc57e9bd9" Nov 21 14:36:39 crc kubenswrapper[4897]: I1121 14:36:39.349923 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dtwql" event={"ID":"6d3cad20-fb17-4726-b8fa-9fd5303b5eea","Type":"ContainerDied","Data":"e7308deb1678f8dd48dc93b1c1d499f86812740488e7d5e6f5dba15c1e1f3f3b"} Nov 21 14:36:39 crc kubenswrapper[4897]: I1121 14:36:39.350054 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dtwql" Nov 21 14:36:39 crc kubenswrapper[4897]: I1121 14:36:39.358148 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e7201326-967f-4b69-b80c-be6a5ecfdc80","Type":"ContainerStarted","Data":"0e91a1a930bba36c07351d03da2585b40e6d524093b2261a632cffdff984befb"} Nov 21 14:36:39 crc kubenswrapper[4897]: I1121 14:36:39.382888 4897 scope.go:117] "RemoveContainer" containerID="5dc39c26fd2373cc55c8e7343e92edce69c4b0dde2cd59ca4fa105ccca9eae4c" Nov 21 14:36:39 crc kubenswrapper[4897]: I1121 14:36:39.413865 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 21 14:36:39 crc kubenswrapper[4897]: I1121 14:36:39.417554 4897 scope.go:117] "RemoveContainer" containerID="5ee18c9557dc1781cb41a1a2df7bdb7b7161886e8a11d31888240d1445a4be95" Nov 21 14:36:39 crc kubenswrapper[4897]: I1121 14:36:39.428294 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 21 14:36:39 crc kubenswrapper[4897]: I1121 14:36:39.444868 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 21 14:36:39 crc kubenswrapper[4897]: E1121 14:36:39.445523 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f688a5e9-9393-4dee-b49a-22b6c75ecfaf" containerName="nova-api-api" Nov 21 14:36:39 crc kubenswrapper[4897]: I1121 14:36:39.445549 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="f688a5e9-9393-4dee-b49a-22b6c75ecfaf" containerName="nova-api-api" Nov 21 14:36:39 crc kubenswrapper[4897]: E1121 14:36:39.445594 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d3cad20-fb17-4726-b8fa-9fd5303b5eea" containerName="extract-content" Nov 21 14:36:39 crc kubenswrapper[4897]: I1121 14:36:39.445604 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d3cad20-fb17-4726-b8fa-9fd5303b5eea" containerName="extract-content" Nov 21 14:36:39 crc kubenswrapper[4897]: E1121 14:36:39.445620 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f688a5e9-9393-4dee-b49a-22b6c75ecfaf" containerName="nova-api-log" Nov 21 14:36:39 crc kubenswrapper[4897]: I1121 14:36:39.445627 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="f688a5e9-9393-4dee-b49a-22b6c75ecfaf" containerName="nova-api-log" Nov 21 14:36:39 crc kubenswrapper[4897]: E1121 14:36:39.445658 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d3cad20-fb17-4726-b8fa-9fd5303b5eea" containerName="extract-utilities" Nov 21 14:36:39 crc kubenswrapper[4897]: I1121 14:36:39.445668 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d3cad20-fb17-4726-b8fa-9fd5303b5eea" containerName="extract-utilities" Nov 21 14:36:39 crc kubenswrapper[4897]: E1121 14:36:39.445689 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d3cad20-fb17-4726-b8fa-9fd5303b5eea" containerName="registry-server" Nov 21 14:36:39 crc kubenswrapper[4897]: I1121 14:36:39.445697 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d3cad20-fb17-4726-b8fa-9fd5303b5eea" containerName="registry-server" Nov 21 14:36:39 crc kubenswrapper[4897]: E1121 14:36:39.445717 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d3cad20-fb17-4726-b8fa-9fd5303b5eea" containerName="registry-server" Nov 21 14:36:39 crc kubenswrapper[4897]: I1121 14:36:39.445724 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d3cad20-fb17-4726-b8fa-9fd5303b5eea" containerName="registry-server" Nov 21 14:36:39 crc kubenswrapper[4897]: I1121 14:36:39.445972 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d3cad20-fb17-4726-b8fa-9fd5303b5eea" containerName="registry-server" Nov 21 14:36:39 crc kubenswrapper[4897]: I1121 14:36:39.445999 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d3cad20-fb17-4726-b8fa-9fd5303b5eea" containerName="registry-server" Nov 21 14:36:39 crc kubenswrapper[4897]: I1121 14:36:39.446020 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="f688a5e9-9393-4dee-b49a-22b6c75ecfaf" containerName="nova-api-log" Nov 21 14:36:39 crc kubenswrapper[4897]: I1121 14:36:39.446055 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="f688a5e9-9393-4dee-b49a-22b6c75ecfaf" containerName="nova-api-api" Nov 21 14:36:39 crc kubenswrapper[4897]: I1121 14:36:39.448063 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 21 14:36:39 crc kubenswrapper[4897]: I1121 14:36:39.453167 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 21 14:36:39 crc kubenswrapper[4897]: I1121 14:36:39.453395 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Nov 21 14:36:39 crc kubenswrapper[4897]: I1121 14:36:39.454813 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Nov 21 14:36:39 crc kubenswrapper[4897]: I1121 14:36:39.467839 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dtwql"] Nov 21 14:36:39 crc kubenswrapper[4897]: I1121 14:36:39.472765 4897 scope.go:117] "RemoveContainer" containerID="684b374923ebdf191516f8820f70c1d389b071397175c74c286552f78600a36a" Nov 21 14:36:39 crc kubenswrapper[4897]: I1121 14:36:39.483173 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-dtwql"] Nov 21 14:36:39 crc kubenswrapper[4897]: I1121 14:36:39.500782 4897 scope.go:117] "RemoveContainer" containerID="a1569dddfd2a291372fa4878a85c9efe401ee175a061a9b22c686f0354f98c05" Nov 21 14:36:39 crc kubenswrapper[4897]: I1121 14:36:39.524341 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/30327523-0b35-459f-be8c-ee423cc5a46c-public-tls-certs\") pod \"nova-api-0\" (UID: \"30327523-0b35-459f-be8c-ee423cc5a46c\") " pod="openstack/nova-api-0" Nov 21 14:36:39 crc kubenswrapper[4897]: I1121 14:36:39.524422 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/30327523-0b35-459f-be8c-ee423cc5a46c-internal-tls-certs\") pod \"nova-api-0\" (UID: \"30327523-0b35-459f-be8c-ee423cc5a46c\") " pod="openstack/nova-api-0" Nov 21 14:36:39 crc kubenswrapper[4897]: I1121 14:36:39.524521 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30327523-0b35-459f-be8c-ee423cc5a46c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"30327523-0b35-459f-be8c-ee423cc5a46c\") " pod="openstack/nova-api-0" Nov 21 14:36:39 crc kubenswrapper[4897]: I1121 14:36:39.524684 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-67f64\" (UniqueName: \"kubernetes.io/projected/30327523-0b35-459f-be8c-ee423cc5a46c-kube-api-access-67f64\") pod \"nova-api-0\" (UID: \"30327523-0b35-459f-be8c-ee423cc5a46c\") " pod="openstack/nova-api-0" Nov 21 14:36:39 crc kubenswrapper[4897]: I1121 14:36:39.524807 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30327523-0b35-459f-be8c-ee423cc5a46c-config-data\") pod \"nova-api-0\" (UID: \"30327523-0b35-459f-be8c-ee423cc5a46c\") " pod="openstack/nova-api-0" Nov 21 14:36:39 crc kubenswrapper[4897]: I1121 14:36:39.525055 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/30327523-0b35-459f-be8c-ee423cc5a46c-logs\") pod \"nova-api-0\" (UID: \"30327523-0b35-459f-be8c-ee423cc5a46c\") " pod="openstack/nova-api-0" Nov 21 14:36:39 crc kubenswrapper[4897]: I1121 14:36:39.529634 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 21 14:36:39 crc kubenswrapper[4897]: I1121 14:36:39.628522 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/30327523-0b35-459f-be8c-ee423cc5a46c-internal-tls-certs\") pod \"nova-api-0\" (UID: \"30327523-0b35-459f-be8c-ee423cc5a46c\") " pod="openstack/nova-api-0" Nov 21 14:36:39 crc kubenswrapper[4897]: I1121 14:36:39.628683 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30327523-0b35-459f-be8c-ee423cc5a46c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"30327523-0b35-459f-be8c-ee423cc5a46c\") " pod="openstack/nova-api-0" Nov 21 14:36:39 crc kubenswrapper[4897]: I1121 14:36:39.628721 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-67f64\" (UniqueName: \"kubernetes.io/projected/30327523-0b35-459f-be8c-ee423cc5a46c-kube-api-access-67f64\") pod \"nova-api-0\" (UID: \"30327523-0b35-459f-be8c-ee423cc5a46c\") " pod="openstack/nova-api-0" Nov 21 14:36:39 crc kubenswrapper[4897]: I1121 14:36:39.628794 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30327523-0b35-459f-be8c-ee423cc5a46c-config-data\") pod \"nova-api-0\" (UID: \"30327523-0b35-459f-be8c-ee423cc5a46c\") " pod="openstack/nova-api-0" Nov 21 14:36:39 crc kubenswrapper[4897]: I1121 14:36:39.628947 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/30327523-0b35-459f-be8c-ee423cc5a46c-logs\") pod \"nova-api-0\" (UID: \"30327523-0b35-459f-be8c-ee423cc5a46c\") " pod="openstack/nova-api-0" Nov 21 14:36:39 crc kubenswrapper[4897]: I1121 14:36:39.629117 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/30327523-0b35-459f-be8c-ee423cc5a46c-public-tls-certs\") pod \"nova-api-0\" (UID: \"30327523-0b35-459f-be8c-ee423cc5a46c\") " pod="openstack/nova-api-0" Nov 21 14:36:39 crc kubenswrapper[4897]: I1121 14:36:39.630809 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/30327523-0b35-459f-be8c-ee423cc5a46c-logs\") pod \"nova-api-0\" (UID: \"30327523-0b35-459f-be8c-ee423cc5a46c\") " pod="openstack/nova-api-0" Nov 21 14:36:39 crc kubenswrapper[4897]: I1121 14:36:39.633936 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/30327523-0b35-459f-be8c-ee423cc5a46c-internal-tls-certs\") pod \"nova-api-0\" (UID: \"30327523-0b35-459f-be8c-ee423cc5a46c\") " pod="openstack/nova-api-0" Nov 21 14:36:39 crc kubenswrapper[4897]: I1121 14:36:39.635139 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/30327523-0b35-459f-be8c-ee423cc5a46c-public-tls-certs\") pod \"nova-api-0\" (UID: \"30327523-0b35-459f-be8c-ee423cc5a46c\") " pod="openstack/nova-api-0" Nov 21 14:36:39 crc kubenswrapper[4897]: I1121 14:36:39.636139 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30327523-0b35-459f-be8c-ee423cc5a46c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"30327523-0b35-459f-be8c-ee423cc5a46c\") " pod="openstack/nova-api-0" Nov 21 14:36:39 crc kubenswrapper[4897]: I1121 14:36:39.637662 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30327523-0b35-459f-be8c-ee423cc5a46c-config-data\") pod \"nova-api-0\" (UID: \"30327523-0b35-459f-be8c-ee423cc5a46c\") " pod="openstack/nova-api-0" Nov 21 14:36:39 crc kubenswrapper[4897]: I1121 14:36:39.655757 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-67f64\" (UniqueName: \"kubernetes.io/projected/30327523-0b35-459f-be8c-ee423cc5a46c-kube-api-access-67f64\") pod \"nova-api-0\" (UID: \"30327523-0b35-459f-be8c-ee423cc5a46c\") " pod="openstack/nova-api-0" Nov 21 14:36:39 crc kubenswrapper[4897]: I1121 14:36:39.769632 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 21 14:36:40 crc kubenswrapper[4897]: I1121 14:36:40.100858 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6d3cad20-fb17-4726-b8fa-9fd5303b5eea" path="/var/lib/kubelet/pods/6d3cad20-fb17-4726-b8fa-9fd5303b5eea/volumes" Nov 21 14:36:40 crc kubenswrapper[4897]: I1121 14:36:40.101633 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f688a5e9-9393-4dee-b49a-22b6c75ecfaf" path="/var/lib/kubelet/pods/f688a5e9-9393-4dee-b49a-22b6c75ecfaf/volumes" Nov 21 14:36:40 crc kubenswrapper[4897]: I1121 14:36:40.239599 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 21 14:36:40 crc kubenswrapper[4897]: I1121 14:36:40.368811 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"30327523-0b35-459f-be8c-ee423cc5a46c","Type":"ContainerStarted","Data":"2cc99ee84a4b666ce8408e35f47d360242436c3b2126eaf9f881f8478d3a1da2"} Nov 21 14:36:40 crc kubenswrapper[4897]: I1121 14:36:40.385522 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e7201326-967f-4b69-b80c-be6a5ecfdc80","Type":"ContainerStarted","Data":"722b7ba5336239a4366c1b893ffa8297bf2de5e756ed0daf6be55eb3dcc802ac"} Nov 21 14:36:40 crc kubenswrapper[4897]: I1121 14:36:40.406021 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.406003287 podStartE2EDuration="3.406003287s" podCreationTimestamp="2025-11-21 14:36:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:36:40.405956514 +0000 UTC m=+1677.690549989" watchObservedRunningTime="2025-11-21 14:36:40.406003287 +0000 UTC m=+1677.690596762" Nov 21 14:36:41 crc kubenswrapper[4897]: I1121 14:36:41.399460 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"30327523-0b35-459f-be8c-ee423cc5a46c","Type":"ContainerStarted","Data":"7376bad15a494623e7903f0ee5e9900b1856999210452371cb714a789a42ff07"} Nov 21 14:36:41 crc kubenswrapper[4897]: I1121 14:36:41.618894 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 21 14:36:42 crc kubenswrapper[4897]: E1121 14:36:42.107961 4897 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podafa69523_63c4_4a94_b94a_61c1343ef10c.slice/crio-conmon-6cb3731d3191a0096a5051532c43fd96973f7663ef9d3571618a165b3dd1afca.scope\": RecentStats: unable to find data in memory cache]" Nov 21 14:36:42 crc kubenswrapper[4897]: I1121 14:36:42.415684 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"30327523-0b35-459f-be8c-ee423cc5a46c","Type":"ContainerStarted","Data":"2c44a183c078dcacfb44d8503f23ba970fffe39117955abb34dae37146773ae5"} Nov 21 14:36:42 crc kubenswrapper[4897]: I1121 14:36:42.756587 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 21 14:36:42 crc kubenswrapper[4897]: I1121 14:36:42.756656 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 21 14:36:45 crc kubenswrapper[4897]: I1121 14:36:45.088811 4897 scope.go:117] "RemoveContainer" containerID="c68f1b003d12b4a2964169a85a04e3f0272d86c891f52e6bebd52f3781520d1a" Nov 21 14:36:45 crc kubenswrapper[4897]: E1121 14:36:45.089534 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 14:36:46 crc kubenswrapper[4897]: I1121 14:36:46.618423 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 21 14:36:46 crc kubenswrapper[4897]: I1121 14:36:46.655793 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 21 14:36:46 crc kubenswrapper[4897]: I1121 14:36:46.670752 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=7.670733659 podStartE2EDuration="7.670733659s" podCreationTimestamp="2025-11-21 14:36:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:36:42.441952842 +0000 UTC m=+1679.726546347" watchObservedRunningTime="2025-11-21 14:36:46.670733659 +0000 UTC m=+1683.955327134" Nov 21 14:36:47 crc kubenswrapper[4897]: I1121 14:36:47.511381 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 21 14:36:47 crc kubenswrapper[4897]: I1121 14:36:47.541932 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 21 14:36:47 crc kubenswrapper[4897]: I1121 14:36:47.756258 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 21 14:36:47 crc kubenswrapper[4897]: I1121 14:36:47.756825 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 21 14:36:48 crc kubenswrapper[4897]: I1121 14:36:48.770627 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="e7201326-967f-4b69-b80c-be6a5ecfdc80" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.1.2:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 21 14:36:48 crc kubenswrapper[4897]: I1121 14:36:48.770676 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="e7201326-967f-4b69-b80c-be6a5ecfdc80" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.1.2:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 21 14:36:49 crc kubenswrapper[4897]: I1121 14:36:49.769988 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 21 14:36:49 crc kubenswrapper[4897]: I1121 14:36:49.770370 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 21 14:36:50 crc kubenswrapper[4897]: I1121 14:36:50.781727 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="30327523-0b35-459f-be8c-ee423cc5a46c" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.1.3:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 21 14:36:50 crc kubenswrapper[4897]: I1121 14:36:50.781735 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="30327523-0b35-459f-be8c-ee423cc5a46c" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.1.3:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 21 14:36:56 crc kubenswrapper[4897]: I1121 14:36:56.090849 4897 scope.go:117] "RemoveContainer" containerID="c68f1b003d12b4a2964169a85a04e3f0272d86c891f52e6bebd52f3781520d1a" Nov 21 14:36:56 crc kubenswrapper[4897]: E1121 14:36:56.091765 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 14:36:57 crc kubenswrapper[4897]: I1121 14:36:57.765695 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 21 14:36:57 crc kubenswrapper[4897]: I1121 14:36:57.766695 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 21 14:36:57 crc kubenswrapper[4897]: I1121 14:36:57.772901 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 21 14:36:57 crc kubenswrapper[4897]: I1121 14:36:57.774267 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 21 14:36:59 crc kubenswrapper[4897]: I1121 14:36:59.776852 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 21 14:36:59 crc kubenswrapper[4897]: I1121 14:36:59.777681 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 21 14:36:59 crc kubenswrapper[4897]: I1121 14:36:59.778157 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 21 14:36:59 crc kubenswrapper[4897]: I1121 14:36:59.781859 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 21 14:37:00 crc kubenswrapper[4897]: I1121 14:37:00.625666 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 21 14:37:00 crc kubenswrapper[4897]: I1121 14:37:00.631423 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 21 14:37:02 crc kubenswrapper[4897]: I1121 14:37:02.498742 4897 scope.go:117] "RemoveContainer" containerID="968d88e3eb6d022e132ec663aa8194b45efff34819cb40fb992e07daba2fd847" Nov 21 14:37:11 crc kubenswrapper[4897]: I1121 14:37:11.089335 4897 scope.go:117] "RemoveContainer" containerID="c68f1b003d12b4a2964169a85a04e3f0272d86c891f52e6bebd52f3781520d1a" Nov 21 14:37:11 crc kubenswrapper[4897]: E1121 14:37:11.090361 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 14:37:12 crc kubenswrapper[4897]: I1121 14:37:12.534466 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-sync-tbtcb"] Nov 21 14:37:12 crc kubenswrapper[4897]: I1121 14:37:12.547652 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-sync-tbtcb"] Nov 21 14:37:12 crc kubenswrapper[4897]: I1121 14:37:12.645517 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-sync-m55sf"] Nov 21 14:37:12 crc kubenswrapper[4897]: I1121 14:37:12.647159 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-m55sf" Nov 21 14:37:12 crc kubenswrapper[4897]: I1121 14:37:12.685443 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-m55sf"] Nov 21 14:37:12 crc kubenswrapper[4897]: I1121 14:37:12.757893 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4c479c7-7728-41fc-bf98-ce2dcf800840-combined-ca-bundle\") pod \"heat-db-sync-m55sf\" (UID: \"c4c479c7-7728-41fc-bf98-ce2dcf800840\") " pod="openstack/heat-db-sync-m55sf" Nov 21 14:37:12 crc kubenswrapper[4897]: I1121 14:37:12.758409 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c4c479c7-7728-41fc-bf98-ce2dcf800840-config-data\") pod \"heat-db-sync-m55sf\" (UID: \"c4c479c7-7728-41fc-bf98-ce2dcf800840\") " pod="openstack/heat-db-sync-m55sf" Nov 21 14:37:12 crc kubenswrapper[4897]: I1121 14:37:12.758623 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-th9pr\" (UniqueName: \"kubernetes.io/projected/c4c479c7-7728-41fc-bf98-ce2dcf800840-kube-api-access-th9pr\") pod \"heat-db-sync-m55sf\" (UID: \"c4c479c7-7728-41fc-bf98-ce2dcf800840\") " pod="openstack/heat-db-sync-m55sf" Nov 21 14:37:12 crc kubenswrapper[4897]: I1121 14:37:12.862024 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4c479c7-7728-41fc-bf98-ce2dcf800840-combined-ca-bundle\") pod \"heat-db-sync-m55sf\" (UID: \"c4c479c7-7728-41fc-bf98-ce2dcf800840\") " pod="openstack/heat-db-sync-m55sf" Nov 21 14:37:12 crc kubenswrapper[4897]: I1121 14:37:12.862140 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c4c479c7-7728-41fc-bf98-ce2dcf800840-config-data\") pod \"heat-db-sync-m55sf\" (UID: \"c4c479c7-7728-41fc-bf98-ce2dcf800840\") " pod="openstack/heat-db-sync-m55sf" Nov 21 14:37:12 crc kubenswrapper[4897]: I1121 14:37:12.862229 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-th9pr\" (UniqueName: \"kubernetes.io/projected/c4c479c7-7728-41fc-bf98-ce2dcf800840-kube-api-access-th9pr\") pod \"heat-db-sync-m55sf\" (UID: \"c4c479c7-7728-41fc-bf98-ce2dcf800840\") " pod="openstack/heat-db-sync-m55sf" Nov 21 14:37:12 crc kubenswrapper[4897]: I1121 14:37:12.870861 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4c479c7-7728-41fc-bf98-ce2dcf800840-combined-ca-bundle\") pod \"heat-db-sync-m55sf\" (UID: \"c4c479c7-7728-41fc-bf98-ce2dcf800840\") " pod="openstack/heat-db-sync-m55sf" Nov 21 14:37:12 crc kubenswrapper[4897]: I1121 14:37:12.870970 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c4c479c7-7728-41fc-bf98-ce2dcf800840-config-data\") pod \"heat-db-sync-m55sf\" (UID: \"c4c479c7-7728-41fc-bf98-ce2dcf800840\") " pod="openstack/heat-db-sync-m55sf" Nov 21 14:37:12 crc kubenswrapper[4897]: I1121 14:37:12.887328 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-th9pr\" (UniqueName: \"kubernetes.io/projected/c4c479c7-7728-41fc-bf98-ce2dcf800840-kube-api-access-th9pr\") pod \"heat-db-sync-m55sf\" (UID: \"c4c479c7-7728-41fc-bf98-ce2dcf800840\") " pod="openstack/heat-db-sync-m55sf" Nov 21 14:37:12 crc kubenswrapper[4897]: I1121 14:37:12.978391 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-m55sf" Nov 21 14:37:13 crc kubenswrapper[4897]: I1121 14:37:13.559168 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-m55sf"] Nov 21 14:37:13 crc kubenswrapper[4897]: I1121 14:37:13.796234 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-m55sf" event={"ID":"c4c479c7-7728-41fc-bf98-ce2dcf800840","Type":"ContainerStarted","Data":"68dfec735f2c3af1ffe088b4d082a02a473347f95c141f1c45d36b2a00fb59ce"} Nov 21 14:37:14 crc kubenswrapper[4897]: I1121 14:37:14.105008 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="59fd52b7-f25b-4a81-a961-41e6cb526c89" path="/var/lib/kubelet/pods/59fd52b7-f25b-4a81-a961-41e6cb526c89/volumes" Nov 21 14:37:14 crc kubenswrapper[4897]: I1121 14:37:14.927996 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 21 14:37:14 crc kubenswrapper[4897]: I1121 14:37:14.929286 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="76d57203-0f85-48aa-87b0-a56af184da2d" containerName="ceilometer-central-agent" containerID="cri-o://249e82c4c97ad9c748b6487ef169e6b51f1b63609fd68c0f73ad594cc5cf7f86" gracePeriod=30 Nov 21 14:37:14 crc kubenswrapper[4897]: I1121 14:37:14.929431 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="76d57203-0f85-48aa-87b0-a56af184da2d" containerName="sg-core" containerID="cri-o://46d3d3be2260d255fc002b9ac213e908e2c7c9482386cd865f4a9c2d267a6cb0" gracePeriod=30 Nov 21 14:37:14 crc kubenswrapper[4897]: I1121 14:37:14.929620 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="76d57203-0f85-48aa-87b0-a56af184da2d" containerName="proxy-httpd" containerID="cri-o://2e36427e415786092f0fb7a4cc3edfd365ffbaf6073ac34dfa737a9f97007dc1" gracePeriod=30 Nov 21 14:37:14 crc kubenswrapper[4897]: I1121 14:37:14.929629 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="76d57203-0f85-48aa-87b0-a56af184da2d" containerName="ceilometer-notification-agent" containerID="cri-o://40b3f26ee7a574717df483a754491f7e0c2353dfd0859d58892267906ac0a7d2" gracePeriod=30 Nov 21 14:37:15 crc kubenswrapper[4897]: I1121 14:37:15.343577 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 21 14:37:15 crc kubenswrapper[4897]: I1121 14:37:15.824341 4897 generic.go:334] "Generic (PLEG): container finished" podID="76d57203-0f85-48aa-87b0-a56af184da2d" containerID="46d3d3be2260d255fc002b9ac213e908e2c7c9482386cd865f4a9c2d267a6cb0" exitCode=2 Nov 21 14:37:15 crc kubenswrapper[4897]: I1121 14:37:15.824419 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"76d57203-0f85-48aa-87b0-a56af184da2d","Type":"ContainerDied","Data":"46d3d3be2260d255fc002b9ac213e908e2c7c9482386cd865f4a9c2d267a6cb0"} Nov 21 14:37:16 crc kubenswrapper[4897]: I1121 14:37:16.651853 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 21 14:37:16 crc kubenswrapper[4897]: I1121 14:37:16.843385 4897 generic.go:334] "Generic (PLEG): container finished" podID="76d57203-0f85-48aa-87b0-a56af184da2d" containerID="2e36427e415786092f0fb7a4cc3edfd365ffbaf6073ac34dfa737a9f97007dc1" exitCode=0 Nov 21 14:37:16 crc kubenswrapper[4897]: I1121 14:37:16.843419 4897 generic.go:334] "Generic (PLEG): container finished" podID="76d57203-0f85-48aa-87b0-a56af184da2d" containerID="249e82c4c97ad9c748b6487ef169e6b51f1b63609fd68c0f73ad594cc5cf7f86" exitCode=0 Nov 21 14:37:16 crc kubenswrapper[4897]: I1121 14:37:16.843432 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"76d57203-0f85-48aa-87b0-a56af184da2d","Type":"ContainerDied","Data":"2e36427e415786092f0fb7a4cc3edfd365ffbaf6073ac34dfa737a9f97007dc1"} Nov 21 14:37:16 crc kubenswrapper[4897]: I1121 14:37:16.843493 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"76d57203-0f85-48aa-87b0-a56af184da2d","Type":"ContainerDied","Data":"249e82c4c97ad9c748b6487ef169e6b51f1b63609fd68c0f73ad594cc5cf7f86"} Nov 21 14:37:17 crc kubenswrapper[4897]: I1121 14:37:17.357260 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="76d57203-0f85-48aa-87b0-a56af184da2d" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.254:3000/\": dial tcp 10.217.0.254:3000: connect: connection refused" Nov 21 14:37:20 crc kubenswrapper[4897]: I1121 14:37:20.899019 4897 generic.go:334] "Generic (PLEG): container finished" podID="76d57203-0f85-48aa-87b0-a56af184da2d" containerID="40b3f26ee7a574717df483a754491f7e0c2353dfd0859d58892267906ac0a7d2" exitCode=0 Nov 21 14:37:20 crc kubenswrapper[4897]: I1121 14:37:20.899088 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"76d57203-0f85-48aa-87b0-a56af184da2d","Type":"ContainerDied","Data":"40b3f26ee7a574717df483a754491f7e0c2353dfd0859d58892267906ac0a7d2"} Nov 21 14:37:22 crc kubenswrapper[4897]: I1121 14:37:22.432270 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="bd692188-6da6-4387-a46d-003a2da0d0c8" containerName="rabbitmq" containerID="cri-o://99656890d61e67a0d062b6637a03db823a2e86706363f8d471ea687b544eb8ea" gracePeriod=604795 Nov 21 14:37:22 crc kubenswrapper[4897]: I1121 14:37:22.888889 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="b513e4ee-7c30-4942-b997-4d52414e1ec0" containerName="rabbitmq" containerID="cri-o://4ded3f218b2e102f2165dac36295117473edb574550d19babb7b246c41c8316f" gracePeriod=604793 Nov 21 14:37:23 crc kubenswrapper[4897]: I1121 14:37:23.269391 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 21 14:37:23 crc kubenswrapper[4897]: I1121 14:37:23.431074 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/76d57203-0f85-48aa-87b0-a56af184da2d-ceilometer-tls-certs\") pod \"76d57203-0f85-48aa-87b0-a56af184da2d\" (UID: \"76d57203-0f85-48aa-87b0-a56af184da2d\") " Nov 21 14:37:23 crc kubenswrapper[4897]: I1121 14:37:23.431601 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/76d57203-0f85-48aa-87b0-a56af184da2d-scripts\") pod \"76d57203-0f85-48aa-87b0-a56af184da2d\" (UID: \"76d57203-0f85-48aa-87b0-a56af184da2d\") " Nov 21 14:37:23 crc kubenswrapper[4897]: I1121 14:37:23.431673 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/76d57203-0f85-48aa-87b0-a56af184da2d-run-httpd\") pod \"76d57203-0f85-48aa-87b0-a56af184da2d\" (UID: \"76d57203-0f85-48aa-87b0-a56af184da2d\") " Nov 21 14:37:23 crc kubenswrapper[4897]: I1121 14:37:23.431723 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/76d57203-0f85-48aa-87b0-a56af184da2d-sg-core-conf-yaml\") pod \"76d57203-0f85-48aa-87b0-a56af184da2d\" (UID: \"76d57203-0f85-48aa-87b0-a56af184da2d\") " Nov 21 14:37:23 crc kubenswrapper[4897]: I1121 14:37:23.431905 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/76d57203-0f85-48aa-87b0-a56af184da2d-log-httpd\") pod \"76d57203-0f85-48aa-87b0-a56af184da2d\" (UID: \"76d57203-0f85-48aa-87b0-a56af184da2d\") " Nov 21 14:37:23 crc kubenswrapper[4897]: I1121 14:37:23.431935 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76d57203-0f85-48aa-87b0-a56af184da2d-combined-ca-bundle\") pod \"76d57203-0f85-48aa-87b0-a56af184da2d\" (UID: \"76d57203-0f85-48aa-87b0-a56af184da2d\") " Nov 21 14:37:23 crc kubenswrapper[4897]: I1121 14:37:23.432214 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76d57203-0f85-48aa-87b0-a56af184da2d-config-data\") pod \"76d57203-0f85-48aa-87b0-a56af184da2d\" (UID: \"76d57203-0f85-48aa-87b0-a56af184da2d\") " Nov 21 14:37:23 crc kubenswrapper[4897]: I1121 14:37:23.432239 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/76d57203-0f85-48aa-87b0-a56af184da2d-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "76d57203-0f85-48aa-87b0-a56af184da2d" (UID: "76d57203-0f85-48aa-87b0-a56af184da2d"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:37:23 crc kubenswrapper[4897]: I1121 14:37:23.432270 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/76d57203-0f85-48aa-87b0-a56af184da2d-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "76d57203-0f85-48aa-87b0-a56af184da2d" (UID: "76d57203-0f85-48aa-87b0-a56af184da2d"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:37:23 crc kubenswrapper[4897]: I1121 14:37:23.432314 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5lcxs\" (UniqueName: \"kubernetes.io/projected/76d57203-0f85-48aa-87b0-a56af184da2d-kube-api-access-5lcxs\") pod \"76d57203-0f85-48aa-87b0-a56af184da2d\" (UID: \"76d57203-0f85-48aa-87b0-a56af184da2d\") " Nov 21 14:37:23 crc kubenswrapper[4897]: I1121 14:37:23.433099 4897 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/76d57203-0f85-48aa-87b0-a56af184da2d-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 21 14:37:23 crc kubenswrapper[4897]: I1121 14:37:23.433114 4897 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/76d57203-0f85-48aa-87b0-a56af184da2d-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 21 14:37:23 crc kubenswrapper[4897]: I1121 14:37:23.438759 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76d57203-0f85-48aa-87b0-a56af184da2d-scripts" (OuterVolumeSpecName: "scripts") pod "76d57203-0f85-48aa-87b0-a56af184da2d" (UID: "76d57203-0f85-48aa-87b0-a56af184da2d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:37:23 crc kubenswrapper[4897]: I1121 14:37:23.438916 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76d57203-0f85-48aa-87b0-a56af184da2d-kube-api-access-5lcxs" (OuterVolumeSpecName: "kube-api-access-5lcxs") pod "76d57203-0f85-48aa-87b0-a56af184da2d" (UID: "76d57203-0f85-48aa-87b0-a56af184da2d"). InnerVolumeSpecName "kube-api-access-5lcxs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:37:23 crc kubenswrapper[4897]: I1121 14:37:23.483326 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76d57203-0f85-48aa-87b0-a56af184da2d-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "76d57203-0f85-48aa-87b0-a56af184da2d" (UID: "76d57203-0f85-48aa-87b0-a56af184da2d"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:37:23 crc kubenswrapper[4897]: I1121 14:37:23.493567 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76d57203-0f85-48aa-87b0-a56af184da2d-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "76d57203-0f85-48aa-87b0-a56af184da2d" (UID: "76d57203-0f85-48aa-87b0-a56af184da2d"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:37:23 crc kubenswrapper[4897]: I1121 14:37:23.527622 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76d57203-0f85-48aa-87b0-a56af184da2d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "76d57203-0f85-48aa-87b0-a56af184da2d" (UID: "76d57203-0f85-48aa-87b0-a56af184da2d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:37:23 crc kubenswrapper[4897]: I1121 14:37:23.535753 4897 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76d57203-0f85-48aa-87b0-a56af184da2d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 21 14:37:23 crc kubenswrapper[4897]: I1121 14:37:23.535790 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5lcxs\" (UniqueName: \"kubernetes.io/projected/76d57203-0f85-48aa-87b0-a56af184da2d-kube-api-access-5lcxs\") on node \"crc\" DevicePath \"\"" Nov 21 14:37:23 crc kubenswrapper[4897]: I1121 14:37:23.535805 4897 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/76d57203-0f85-48aa-87b0-a56af184da2d-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 21 14:37:23 crc kubenswrapper[4897]: I1121 14:37:23.535819 4897 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/76d57203-0f85-48aa-87b0-a56af184da2d-scripts\") on node \"crc\" DevicePath \"\"" Nov 21 14:37:23 crc kubenswrapper[4897]: I1121 14:37:23.535830 4897 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/76d57203-0f85-48aa-87b0-a56af184da2d-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 21 14:37:23 crc kubenswrapper[4897]: I1121 14:37:23.585605 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76d57203-0f85-48aa-87b0-a56af184da2d-config-data" (OuterVolumeSpecName: "config-data") pod "76d57203-0f85-48aa-87b0-a56af184da2d" (UID: "76d57203-0f85-48aa-87b0-a56af184da2d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:37:23 crc kubenswrapper[4897]: I1121 14:37:23.640566 4897 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76d57203-0f85-48aa-87b0-a56af184da2d-config-data\") on node \"crc\" DevicePath \"\"" Nov 21 14:37:23 crc kubenswrapper[4897]: I1121 14:37:23.957700 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"76d57203-0f85-48aa-87b0-a56af184da2d","Type":"ContainerDied","Data":"a47f15f73782016a708b146c32acd32a80a6e64ed1a06625faa63a9bb2c4c819"} Nov 21 14:37:23 crc kubenswrapper[4897]: I1121 14:37:23.957848 4897 scope.go:117] "RemoveContainer" containerID="2e36427e415786092f0fb7a4cc3edfd365ffbaf6073ac34dfa737a9f97007dc1" Nov 21 14:37:23 crc kubenswrapper[4897]: I1121 14:37:23.958070 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 21 14:37:24 crc kubenswrapper[4897]: I1121 14:37:24.009511 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 21 14:37:24 crc kubenswrapper[4897]: I1121 14:37:24.023560 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 21 14:37:24 crc kubenswrapper[4897]: I1121 14:37:24.030786 4897 scope.go:117] "RemoveContainer" containerID="46d3d3be2260d255fc002b9ac213e908e2c7c9482386cd865f4a9c2d267a6cb0" Nov 21 14:37:24 crc kubenswrapper[4897]: I1121 14:37:24.041240 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 21 14:37:24 crc kubenswrapper[4897]: E1121 14:37:24.041910 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76d57203-0f85-48aa-87b0-a56af184da2d" containerName="ceilometer-central-agent" Nov 21 14:37:24 crc kubenswrapper[4897]: I1121 14:37:24.041976 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="76d57203-0f85-48aa-87b0-a56af184da2d" containerName="ceilometer-central-agent" Nov 21 14:37:24 crc kubenswrapper[4897]: E1121 14:37:24.042004 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76d57203-0f85-48aa-87b0-a56af184da2d" containerName="ceilometer-notification-agent" Nov 21 14:37:24 crc kubenswrapper[4897]: I1121 14:37:24.042014 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="76d57203-0f85-48aa-87b0-a56af184da2d" containerName="ceilometer-notification-agent" Nov 21 14:37:24 crc kubenswrapper[4897]: E1121 14:37:24.042027 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76d57203-0f85-48aa-87b0-a56af184da2d" containerName="sg-core" Nov 21 14:37:24 crc kubenswrapper[4897]: I1121 14:37:24.042035 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="76d57203-0f85-48aa-87b0-a56af184da2d" containerName="sg-core" Nov 21 14:37:24 crc kubenswrapper[4897]: E1121 14:37:24.042060 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76d57203-0f85-48aa-87b0-a56af184da2d" containerName="proxy-httpd" Nov 21 14:37:24 crc kubenswrapper[4897]: I1121 14:37:24.042068 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="76d57203-0f85-48aa-87b0-a56af184da2d" containerName="proxy-httpd" Nov 21 14:37:24 crc kubenswrapper[4897]: I1121 14:37:24.042414 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="76d57203-0f85-48aa-87b0-a56af184da2d" containerName="ceilometer-central-agent" Nov 21 14:37:24 crc kubenswrapper[4897]: I1121 14:37:24.042443 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="76d57203-0f85-48aa-87b0-a56af184da2d" containerName="sg-core" Nov 21 14:37:24 crc kubenswrapper[4897]: I1121 14:37:24.042471 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="76d57203-0f85-48aa-87b0-a56af184da2d" containerName="proxy-httpd" Nov 21 14:37:24 crc kubenswrapper[4897]: I1121 14:37:24.042484 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="76d57203-0f85-48aa-87b0-a56af184da2d" containerName="ceilometer-notification-agent" Nov 21 14:37:24 crc kubenswrapper[4897]: I1121 14:37:24.045585 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 21 14:37:24 crc kubenswrapper[4897]: I1121 14:37:24.051950 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 21 14:37:24 crc kubenswrapper[4897]: I1121 14:37:24.052286 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 21 14:37:24 crc kubenswrapper[4897]: I1121 14:37:24.052437 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 21 14:37:24 crc kubenswrapper[4897]: I1121 14:37:24.052716 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 21 14:37:24 crc kubenswrapper[4897]: I1121 14:37:24.062465 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c628be92-b88f-4e02-a981-4b8fe6fd7c13-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c628be92-b88f-4e02-a981-4b8fe6fd7c13\") " pod="openstack/ceilometer-0" Nov 21 14:37:24 crc kubenswrapper[4897]: I1121 14:37:24.062559 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c628be92-b88f-4e02-a981-4b8fe6fd7c13-scripts\") pod \"ceilometer-0\" (UID: \"c628be92-b88f-4e02-a981-4b8fe6fd7c13\") " pod="openstack/ceilometer-0" Nov 21 14:37:24 crc kubenswrapper[4897]: I1121 14:37:24.062607 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c628be92-b88f-4e02-a981-4b8fe6fd7c13-log-httpd\") pod \"ceilometer-0\" (UID: \"c628be92-b88f-4e02-a981-4b8fe6fd7c13\") " pod="openstack/ceilometer-0" Nov 21 14:37:24 crc kubenswrapper[4897]: I1121 14:37:24.062658 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c628be92-b88f-4e02-a981-4b8fe6fd7c13-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"c628be92-b88f-4e02-a981-4b8fe6fd7c13\") " pod="openstack/ceilometer-0" Nov 21 14:37:24 crc kubenswrapper[4897]: I1121 14:37:24.062793 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c628be92-b88f-4e02-a981-4b8fe6fd7c13-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c628be92-b88f-4e02-a981-4b8fe6fd7c13\") " pod="openstack/ceilometer-0" Nov 21 14:37:24 crc kubenswrapper[4897]: I1121 14:37:24.062819 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c628be92-b88f-4e02-a981-4b8fe6fd7c13-config-data\") pod \"ceilometer-0\" (UID: \"c628be92-b88f-4e02-a981-4b8fe6fd7c13\") " pod="openstack/ceilometer-0" Nov 21 14:37:24 crc kubenswrapper[4897]: I1121 14:37:24.062958 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k2qzq\" (UniqueName: \"kubernetes.io/projected/c628be92-b88f-4e02-a981-4b8fe6fd7c13-kube-api-access-k2qzq\") pod \"ceilometer-0\" (UID: \"c628be92-b88f-4e02-a981-4b8fe6fd7c13\") " pod="openstack/ceilometer-0" Nov 21 14:37:24 crc kubenswrapper[4897]: I1121 14:37:24.063089 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c628be92-b88f-4e02-a981-4b8fe6fd7c13-run-httpd\") pod \"ceilometer-0\" (UID: \"c628be92-b88f-4e02-a981-4b8fe6fd7c13\") " pod="openstack/ceilometer-0" Nov 21 14:37:24 crc kubenswrapper[4897]: I1121 14:37:24.088661 4897 scope.go:117] "RemoveContainer" containerID="40b3f26ee7a574717df483a754491f7e0c2353dfd0859d58892267906ac0a7d2" Nov 21 14:37:24 crc kubenswrapper[4897]: I1121 14:37:24.102989 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="76d57203-0f85-48aa-87b0-a56af184da2d" path="/var/lib/kubelet/pods/76d57203-0f85-48aa-87b0-a56af184da2d/volumes" Nov 21 14:37:24 crc kubenswrapper[4897]: I1121 14:37:24.110776 4897 scope.go:117] "RemoveContainer" containerID="249e82c4c97ad9c748b6487ef169e6b51f1b63609fd68c0f73ad594cc5cf7f86" Nov 21 14:37:24 crc kubenswrapper[4897]: I1121 14:37:24.166698 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c628be92-b88f-4e02-a981-4b8fe6fd7c13-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c628be92-b88f-4e02-a981-4b8fe6fd7c13\") " pod="openstack/ceilometer-0" Nov 21 14:37:24 crc kubenswrapper[4897]: I1121 14:37:24.166758 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c628be92-b88f-4e02-a981-4b8fe6fd7c13-config-data\") pod \"ceilometer-0\" (UID: \"c628be92-b88f-4e02-a981-4b8fe6fd7c13\") " pod="openstack/ceilometer-0" Nov 21 14:37:24 crc kubenswrapper[4897]: I1121 14:37:24.166919 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k2qzq\" (UniqueName: \"kubernetes.io/projected/c628be92-b88f-4e02-a981-4b8fe6fd7c13-kube-api-access-k2qzq\") pod \"ceilometer-0\" (UID: \"c628be92-b88f-4e02-a981-4b8fe6fd7c13\") " pod="openstack/ceilometer-0" Nov 21 14:37:24 crc kubenswrapper[4897]: I1121 14:37:24.166980 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c628be92-b88f-4e02-a981-4b8fe6fd7c13-run-httpd\") pod \"ceilometer-0\" (UID: \"c628be92-b88f-4e02-a981-4b8fe6fd7c13\") " pod="openstack/ceilometer-0" Nov 21 14:37:24 crc kubenswrapper[4897]: I1121 14:37:24.167199 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c628be92-b88f-4e02-a981-4b8fe6fd7c13-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c628be92-b88f-4e02-a981-4b8fe6fd7c13\") " pod="openstack/ceilometer-0" Nov 21 14:37:24 crc kubenswrapper[4897]: I1121 14:37:24.167237 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c628be92-b88f-4e02-a981-4b8fe6fd7c13-scripts\") pod \"ceilometer-0\" (UID: \"c628be92-b88f-4e02-a981-4b8fe6fd7c13\") " pod="openstack/ceilometer-0" Nov 21 14:37:24 crc kubenswrapper[4897]: I1121 14:37:24.167267 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c628be92-b88f-4e02-a981-4b8fe6fd7c13-log-httpd\") pod \"ceilometer-0\" (UID: \"c628be92-b88f-4e02-a981-4b8fe6fd7c13\") " pod="openstack/ceilometer-0" Nov 21 14:37:24 crc kubenswrapper[4897]: I1121 14:37:24.167306 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c628be92-b88f-4e02-a981-4b8fe6fd7c13-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"c628be92-b88f-4e02-a981-4b8fe6fd7c13\") " pod="openstack/ceilometer-0" Nov 21 14:37:24 crc kubenswrapper[4897]: I1121 14:37:24.167961 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c628be92-b88f-4e02-a981-4b8fe6fd7c13-run-httpd\") pod \"ceilometer-0\" (UID: \"c628be92-b88f-4e02-a981-4b8fe6fd7c13\") " pod="openstack/ceilometer-0" Nov 21 14:37:24 crc kubenswrapper[4897]: I1121 14:37:24.168255 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c628be92-b88f-4e02-a981-4b8fe6fd7c13-log-httpd\") pod \"ceilometer-0\" (UID: \"c628be92-b88f-4e02-a981-4b8fe6fd7c13\") " pod="openstack/ceilometer-0" Nov 21 14:37:24 crc kubenswrapper[4897]: I1121 14:37:24.221111 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c628be92-b88f-4e02-a981-4b8fe6fd7c13-scripts\") pod \"ceilometer-0\" (UID: \"c628be92-b88f-4e02-a981-4b8fe6fd7c13\") " pod="openstack/ceilometer-0" Nov 21 14:37:24 crc kubenswrapper[4897]: I1121 14:37:24.221526 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c628be92-b88f-4e02-a981-4b8fe6fd7c13-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"c628be92-b88f-4e02-a981-4b8fe6fd7c13\") " pod="openstack/ceilometer-0" Nov 21 14:37:24 crc kubenswrapper[4897]: I1121 14:37:24.221882 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k2qzq\" (UniqueName: \"kubernetes.io/projected/c628be92-b88f-4e02-a981-4b8fe6fd7c13-kube-api-access-k2qzq\") pod \"ceilometer-0\" (UID: \"c628be92-b88f-4e02-a981-4b8fe6fd7c13\") " pod="openstack/ceilometer-0" Nov 21 14:37:24 crc kubenswrapper[4897]: I1121 14:37:24.222680 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c628be92-b88f-4e02-a981-4b8fe6fd7c13-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c628be92-b88f-4e02-a981-4b8fe6fd7c13\") " pod="openstack/ceilometer-0" Nov 21 14:37:24 crc kubenswrapper[4897]: I1121 14:37:24.224560 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c628be92-b88f-4e02-a981-4b8fe6fd7c13-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c628be92-b88f-4e02-a981-4b8fe6fd7c13\") " pod="openstack/ceilometer-0" Nov 21 14:37:24 crc kubenswrapper[4897]: I1121 14:37:24.234843 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c628be92-b88f-4e02-a981-4b8fe6fd7c13-config-data\") pod \"ceilometer-0\" (UID: \"c628be92-b88f-4e02-a981-4b8fe6fd7c13\") " pod="openstack/ceilometer-0" Nov 21 14:37:24 crc kubenswrapper[4897]: I1121 14:37:24.378610 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 21 14:37:24 crc kubenswrapper[4897]: I1121 14:37:24.856776 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 21 14:37:24 crc kubenswrapper[4897]: I1121 14:37:24.954378 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="b513e4ee-7c30-4942-b997-4d52414e1ec0" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.130:5671: connect: connection refused" Nov 21 14:37:24 crc kubenswrapper[4897]: I1121 14:37:24.972587 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c628be92-b88f-4e02-a981-4b8fe6fd7c13","Type":"ContainerStarted","Data":"45f92c465862ebeda7aa4e62d2413df6884dfcb614e062ad6c42da1b275c7528"} Nov 21 14:37:25 crc kubenswrapper[4897]: I1121 14:37:25.390282 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="bd692188-6da6-4387-a46d-003a2da0d0c8" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.131:5671: connect: connection refused" Nov 21 14:37:26 crc kubenswrapper[4897]: I1121 14:37:26.091000 4897 scope.go:117] "RemoveContainer" containerID="c68f1b003d12b4a2964169a85a04e3f0272d86c891f52e6bebd52f3781520d1a" Nov 21 14:37:26 crc kubenswrapper[4897]: E1121 14:37:26.091675 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 14:37:34 crc kubenswrapper[4897]: I1121 14:37:34.955167 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="b513e4ee-7c30-4942-b997-4d52414e1ec0" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.130:5671: connect: connection refused" Nov 21 14:37:35 crc kubenswrapper[4897]: I1121 14:37:35.124901 4897 generic.go:334] "Generic (PLEG): container finished" podID="bd692188-6da6-4387-a46d-003a2da0d0c8" containerID="99656890d61e67a0d062b6637a03db823a2e86706363f8d471ea687b544eb8ea" exitCode=0 Nov 21 14:37:35 crc kubenswrapper[4897]: I1121 14:37:35.124956 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"bd692188-6da6-4387-a46d-003a2da0d0c8","Type":"ContainerDied","Data":"99656890d61e67a0d062b6637a03db823a2e86706363f8d471ea687b544eb8ea"} Nov 21 14:37:35 crc kubenswrapper[4897]: I1121 14:37:35.390825 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="bd692188-6da6-4387-a46d-003a2da0d0c8" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.131:5671: connect: connection refused" Nov 21 14:37:36 crc kubenswrapper[4897]: I1121 14:37:36.147025 4897 generic.go:334] "Generic (PLEG): container finished" podID="b513e4ee-7c30-4942-b997-4d52414e1ec0" containerID="4ded3f218b2e102f2165dac36295117473edb574550d19babb7b246c41c8316f" exitCode=0 Nov 21 14:37:36 crc kubenswrapper[4897]: I1121 14:37:36.147100 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"b513e4ee-7c30-4942-b997-4d52414e1ec0","Type":"ContainerDied","Data":"4ded3f218b2e102f2165dac36295117473edb574550d19babb7b246c41c8316f"} Nov 21 14:37:40 crc kubenswrapper[4897]: I1121 14:37:40.089214 4897 scope.go:117] "RemoveContainer" containerID="c68f1b003d12b4a2964169a85a04e3f0272d86c891f52e6bebd52f3781520d1a" Nov 21 14:37:40 crc kubenswrapper[4897]: E1121 14:37:40.090294 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 14:37:44 crc kubenswrapper[4897]: I1121 14:37:44.954167 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="b513e4ee-7c30-4942-b997-4d52414e1ec0" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.130:5671: connect: connection refused" Nov 21 14:37:44 crc kubenswrapper[4897]: I1121 14:37:44.954878 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Nov 21 14:37:45 crc kubenswrapper[4897]: I1121 14:37:45.390918 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="bd692188-6da6-4387-a46d-003a2da0d0c8" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.131:5671: connect: connection refused" Nov 21 14:37:45 crc kubenswrapper[4897]: I1121 14:37:45.391332 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Nov 21 14:37:51 crc kubenswrapper[4897]: I1121 14:37:51.089316 4897 scope.go:117] "RemoveContainer" containerID="c68f1b003d12b4a2964169a85a04e3f0272d86c891f52e6bebd52f3781520d1a" Nov 21 14:37:51 crc kubenswrapper[4897]: E1121 14:37:51.090351 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 14:37:59 crc kubenswrapper[4897]: I1121 14:37:59.954572 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="b513e4ee-7c30-4942-b997-4d52414e1ec0" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.130:5671: i/o timeout" Nov 21 14:38:00 crc kubenswrapper[4897]: I1121 14:38:00.391271 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="bd692188-6da6-4387-a46d-003a2da0d0c8" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.131:5671: i/o timeout" Nov 21 14:38:02 crc kubenswrapper[4897]: I1121 14:38:02.080728 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 21 14:38:02 crc kubenswrapper[4897]: I1121 14:38:02.087093 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 21 14:38:02 crc kubenswrapper[4897]: I1121 14:38:02.154325 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/bd692188-6da6-4387-a46d-003a2da0d0c8-plugins-conf\") pod \"bd692188-6da6-4387-a46d-003a2da0d0c8\" (UID: \"bd692188-6da6-4387-a46d-003a2da0d0c8\") " Nov 21 14:38:02 crc kubenswrapper[4897]: I1121 14:38:02.154394 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b513e4ee-7c30-4942-b997-4d52414e1ec0-rabbitmq-confd\") pod \"b513e4ee-7c30-4942-b997-4d52414e1ec0\" (UID: \"b513e4ee-7c30-4942-b997-4d52414e1ec0\") " Nov 21 14:38:02 crc kubenswrapper[4897]: I1121 14:38:02.154446 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"bd692188-6da6-4387-a46d-003a2da0d0c8\" (UID: \"bd692188-6da6-4387-a46d-003a2da0d0c8\") " Nov 21 14:38:02 crc kubenswrapper[4897]: I1121 14:38:02.154481 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/bd692188-6da6-4387-a46d-003a2da0d0c8-rabbitmq-tls\") pod \"bd692188-6da6-4387-a46d-003a2da0d0c8\" (UID: \"bd692188-6da6-4387-a46d-003a2da0d0c8\") " Nov 21 14:38:02 crc kubenswrapper[4897]: I1121 14:38:02.154533 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b513e4ee-7c30-4942-b997-4d52414e1ec0-server-conf\") pod \"b513e4ee-7c30-4942-b997-4d52414e1ec0\" (UID: \"b513e4ee-7c30-4942-b997-4d52414e1ec0\") " Nov 21 14:38:02 crc kubenswrapper[4897]: I1121 14:38:02.154571 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b513e4ee-7c30-4942-b997-4d52414e1ec0-rabbitmq-plugins\") pod \"b513e4ee-7c30-4942-b997-4d52414e1ec0\" (UID: \"b513e4ee-7c30-4942-b997-4d52414e1ec0\") " Nov 21 14:38:02 crc kubenswrapper[4897]: I1121 14:38:02.154624 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/bd692188-6da6-4387-a46d-003a2da0d0c8-rabbitmq-plugins\") pod \"bd692188-6da6-4387-a46d-003a2da0d0c8\" (UID: \"bd692188-6da6-4387-a46d-003a2da0d0c8\") " Nov 21 14:38:02 crc kubenswrapper[4897]: I1121 14:38:02.154663 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"b513e4ee-7c30-4942-b997-4d52414e1ec0\" (UID: \"b513e4ee-7c30-4942-b997-4d52414e1ec0\") " Nov 21 14:38:02 crc kubenswrapper[4897]: I1121 14:38:02.154701 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kbnvn\" (UniqueName: \"kubernetes.io/projected/bd692188-6da6-4387-a46d-003a2da0d0c8-kube-api-access-kbnvn\") pod \"bd692188-6da6-4387-a46d-003a2da0d0c8\" (UID: \"bd692188-6da6-4387-a46d-003a2da0d0c8\") " Nov 21 14:38:02 crc kubenswrapper[4897]: I1121 14:38:02.154734 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s2gt5\" (UniqueName: \"kubernetes.io/projected/b513e4ee-7c30-4942-b997-4d52414e1ec0-kube-api-access-s2gt5\") pod \"b513e4ee-7c30-4942-b997-4d52414e1ec0\" (UID: \"b513e4ee-7c30-4942-b997-4d52414e1ec0\") " Nov 21 14:38:02 crc kubenswrapper[4897]: I1121 14:38:02.154768 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/bd692188-6da6-4387-a46d-003a2da0d0c8-server-conf\") pod \"bd692188-6da6-4387-a46d-003a2da0d0c8\" (UID: \"bd692188-6da6-4387-a46d-003a2da0d0c8\") " Nov 21 14:38:02 crc kubenswrapper[4897]: I1121 14:38:02.154823 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b513e4ee-7c30-4942-b997-4d52414e1ec0-erlang-cookie-secret\") pod \"b513e4ee-7c30-4942-b997-4d52414e1ec0\" (UID: \"b513e4ee-7c30-4942-b997-4d52414e1ec0\") " Nov 21 14:38:02 crc kubenswrapper[4897]: I1121 14:38:02.154851 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b513e4ee-7c30-4942-b997-4d52414e1ec0-rabbitmq-erlang-cookie\") pod \"b513e4ee-7c30-4942-b997-4d52414e1ec0\" (UID: \"b513e4ee-7c30-4942-b997-4d52414e1ec0\") " Nov 21 14:38:02 crc kubenswrapper[4897]: I1121 14:38:02.154869 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bd692188-6da6-4387-a46d-003a2da0d0c8-config-data\") pod \"bd692188-6da6-4387-a46d-003a2da0d0c8\" (UID: \"bd692188-6da6-4387-a46d-003a2da0d0c8\") " Nov 21 14:38:02 crc kubenswrapper[4897]: I1121 14:38:02.154943 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b513e4ee-7c30-4942-b997-4d52414e1ec0-plugins-conf\") pod \"b513e4ee-7c30-4942-b997-4d52414e1ec0\" (UID: \"b513e4ee-7c30-4942-b997-4d52414e1ec0\") " Nov 21 14:38:02 crc kubenswrapper[4897]: I1121 14:38:02.154959 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/bd692188-6da6-4387-a46d-003a2da0d0c8-rabbitmq-erlang-cookie\") pod \"bd692188-6da6-4387-a46d-003a2da0d0c8\" (UID: \"bd692188-6da6-4387-a46d-003a2da0d0c8\") " Nov 21 14:38:02 crc kubenswrapper[4897]: I1121 14:38:02.154990 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/bd692188-6da6-4387-a46d-003a2da0d0c8-pod-info\") pod \"bd692188-6da6-4387-a46d-003a2da0d0c8\" (UID: \"bd692188-6da6-4387-a46d-003a2da0d0c8\") " Nov 21 14:38:02 crc kubenswrapper[4897]: I1121 14:38:02.154995 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bd692188-6da6-4387-a46d-003a2da0d0c8-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "bd692188-6da6-4387-a46d-003a2da0d0c8" (UID: "bd692188-6da6-4387-a46d-003a2da0d0c8"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:38:02 crc kubenswrapper[4897]: I1121 14:38:02.155013 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b513e4ee-7c30-4942-b997-4d52414e1ec0-rabbitmq-tls\") pod \"b513e4ee-7c30-4942-b997-4d52414e1ec0\" (UID: \"b513e4ee-7c30-4942-b997-4d52414e1ec0\") " Nov 21 14:38:02 crc kubenswrapper[4897]: I1121 14:38:02.155077 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/bd692188-6da6-4387-a46d-003a2da0d0c8-erlang-cookie-secret\") pod \"bd692188-6da6-4387-a46d-003a2da0d0c8\" (UID: \"bd692188-6da6-4387-a46d-003a2da0d0c8\") " Nov 21 14:38:02 crc kubenswrapper[4897]: I1121 14:38:02.155126 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b513e4ee-7c30-4942-b997-4d52414e1ec0-pod-info\") pod \"b513e4ee-7c30-4942-b997-4d52414e1ec0\" (UID: \"b513e4ee-7c30-4942-b997-4d52414e1ec0\") " Nov 21 14:38:02 crc kubenswrapper[4897]: I1121 14:38:02.155177 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b513e4ee-7c30-4942-b997-4d52414e1ec0-config-data\") pod \"b513e4ee-7c30-4942-b997-4d52414e1ec0\" (UID: \"b513e4ee-7c30-4942-b997-4d52414e1ec0\") " Nov 21 14:38:02 crc kubenswrapper[4897]: I1121 14:38:02.155206 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/bd692188-6da6-4387-a46d-003a2da0d0c8-rabbitmq-confd\") pod \"bd692188-6da6-4387-a46d-003a2da0d0c8\" (UID: \"bd692188-6da6-4387-a46d-003a2da0d0c8\") " Nov 21 14:38:02 crc kubenswrapper[4897]: I1121 14:38:02.156423 4897 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/bd692188-6da6-4387-a46d-003a2da0d0c8-plugins-conf\") on node \"crc\" DevicePath \"\"" Nov 21 14:38:02 crc kubenswrapper[4897]: I1121 14:38:02.174281 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bd692188-6da6-4387-a46d-003a2da0d0c8-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "bd692188-6da6-4387-a46d-003a2da0d0c8" (UID: "bd692188-6da6-4387-a46d-003a2da0d0c8"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:38:02 crc kubenswrapper[4897]: I1121 14:38:02.211060 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd692188-6da6-4387-a46d-003a2da0d0c8-kube-api-access-kbnvn" (OuterVolumeSpecName: "kube-api-access-kbnvn") pod "bd692188-6da6-4387-a46d-003a2da0d0c8" (UID: "bd692188-6da6-4387-a46d-003a2da0d0c8"). InnerVolumeSpecName "kube-api-access-kbnvn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:38:02 crc kubenswrapper[4897]: I1121 14:38:02.212488 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/bd692188-6da6-4387-a46d-003a2da0d0c8-pod-info" (OuterVolumeSpecName: "pod-info") pod "bd692188-6da6-4387-a46d-003a2da0d0c8" (UID: "bd692188-6da6-4387-a46d-003a2da0d0c8"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Nov 21 14:38:02 crc kubenswrapper[4897]: I1121 14:38:02.215653 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "persistence") pod "b513e4ee-7c30-4942-b997-4d52414e1ec0" (UID: "b513e4ee-7c30-4942-b997-4d52414e1ec0"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 21 14:38:02 crc kubenswrapper[4897]: I1121 14:38:02.220875 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd692188-6da6-4387-a46d-003a2da0d0c8-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "bd692188-6da6-4387-a46d-003a2da0d0c8" (UID: "bd692188-6da6-4387-a46d-003a2da0d0c8"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:38:02 crc kubenswrapper[4897]: I1121 14:38:02.226212 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "persistence") pod "bd692188-6da6-4387-a46d-003a2da0d0c8" (UID: "bd692188-6da6-4387-a46d-003a2da0d0c8"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 21 14:38:02 crc kubenswrapper[4897]: I1121 14:38:02.227502 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b513e4ee-7c30-4942-b997-4d52414e1ec0-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "b513e4ee-7c30-4942-b997-4d52414e1ec0" (UID: "b513e4ee-7c30-4942-b997-4d52414e1ec0"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:38:02 crc kubenswrapper[4897]: I1121 14:38:02.230235 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b513e4ee-7c30-4942-b997-4d52414e1ec0-kube-api-access-s2gt5" (OuterVolumeSpecName: "kube-api-access-s2gt5") pod "b513e4ee-7c30-4942-b997-4d52414e1ec0" (UID: "b513e4ee-7c30-4942-b997-4d52414e1ec0"). InnerVolumeSpecName "kube-api-access-s2gt5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:38:02 crc kubenswrapper[4897]: I1121 14:38:02.235819 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/b513e4ee-7c30-4942-b997-4d52414e1ec0-pod-info" (OuterVolumeSpecName: "pod-info") pod "b513e4ee-7c30-4942-b997-4d52414e1ec0" (UID: "b513e4ee-7c30-4942-b997-4d52414e1ec0"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Nov 21 14:38:02 crc kubenswrapper[4897]: I1121 14:38:02.243665 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b513e4ee-7c30-4942-b997-4d52414e1ec0-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "b513e4ee-7c30-4942-b997-4d52414e1ec0" (UID: "b513e4ee-7c30-4942-b997-4d52414e1ec0"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:38:02 crc kubenswrapper[4897]: I1121 14:38:02.243726 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b513e4ee-7c30-4942-b997-4d52414e1ec0-config-data" (OuterVolumeSpecName: "config-data") pod "b513e4ee-7c30-4942-b997-4d52414e1ec0" (UID: "b513e4ee-7c30-4942-b997-4d52414e1ec0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:38:02 crc kubenswrapper[4897]: I1121 14:38:02.258051 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s2gt5\" (UniqueName: \"kubernetes.io/projected/b513e4ee-7c30-4942-b997-4d52414e1ec0-kube-api-access-s2gt5\") on node \"crc\" DevicePath \"\"" Nov 21 14:38:02 crc kubenswrapper[4897]: I1121 14:38:02.258363 4897 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b513e4ee-7c30-4942-b997-4d52414e1ec0-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Nov 21 14:38:02 crc kubenswrapper[4897]: I1121 14:38:02.258373 4897 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/bd692188-6da6-4387-a46d-003a2da0d0c8-pod-info\") on node \"crc\" DevicePath \"\"" Nov 21 14:38:02 crc kubenswrapper[4897]: I1121 14:38:02.258383 4897 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b513e4ee-7c30-4942-b997-4d52414e1ec0-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Nov 21 14:38:02 crc kubenswrapper[4897]: I1121 14:38:02.258391 4897 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/bd692188-6da6-4387-a46d-003a2da0d0c8-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Nov 21 14:38:02 crc kubenswrapper[4897]: I1121 14:38:02.258399 4897 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b513e4ee-7c30-4942-b997-4d52414e1ec0-pod-info\") on node \"crc\" DevicePath \"\"" Nov 21 14:38:02 crc kubenswrapper[4897]: I1121 14:38:02.258407 4897 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b513e4ee-7c30-4942-b997-4d52414e1ec0-config-data\") on node \"crc\" DevicePath \"\"" Nov 21 14:38:02 crc kubenswrapper[4897]: I1121 14:38:02.258436 4897 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Nov 21 14:38:02 crc kubenswrapper[4897]: I1121 14:38:02.258445 4897 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/bd692188-6da6-4387-a46d-003a2da0d0c8-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Nov 21 14:38:02 crc kubenswrapper[4897]: I1121 14:38:02.258458 4897 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Nov 21 14:38:02 crc kubenswrapper[4897]: I1121 14:38:02.258476 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kbnvn\" (UniqueName: \"kubernetes.io/projected/bd692188-6da6-4387-a46d-003a2da0d0c8-kube-api-access-kbnvn\") on node \"crc\" DevicePath \"\"" Nov 21 14:38:02 crc kubenswrapper[4897]: I1121 14:38:02.285041 4897 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Nov 21 14:38:02 crc kubenswrapper[4897]: I1121 14:38:02.291521 4897 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Nov 21 14:38:02 crc kubenswrapper[4897]: I1121 14:38:02.361428 4897 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Nov 21 14:38:02 crc kubenswrapper[4897]: I1121 14:38:02.361471 4897 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Nov 21 14:38:02 crc kubenswrapper[4897]: I1121 14:38:02.477081 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"bd692188-6da6-4387-a46d-003a2da0d0c8","Type":"ContainerDied","Data":"a8d57c480eb257dc71e9bdee5d9c69bc7b91ca6a3572a2555ff0fd0a19f8543e"} Nov 21 14:38:02 crc kubenswrapper[4897]: I1121 14:38:02.477146 4897 scope.go:117] "RemoveContainer" containerID="99656890d61e67a0d062b6637a03db823a2e86706363f8d471ea687b544eb8ea" Nov 21 14:38:02 crc kubenswrapper[4897]: I1121 14:38:02.477105 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 21 14:38:02 crc kubenswrapper[4897]: I1121 14:38:02.481038 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"b513e4ee-7c30-4942-b997-4d52414e1ec0","Type":"ContainerDied","Data":"d57051f4a2d9e61d4d7988ce611e9c0574d0451a505583e0be5558754171da2b"} Nov 21 14:38:02 crc kubenswrapper[4897]: I1121 14:38:02.481109 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 21 14:38:02 crc kubenswrapper[4897]: I1121 14:38:02.665912 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd692188-6da6-4387-a46d-003a2da0d0c8-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "bd692188-6da6-4387-a46d-003a2da0d0c8" (UID: "bd692188-6da6-4387-a46d-003a2da0d0c8"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:38:02 crc kubenswrapper[4897]: I1121 14:38:02.665945 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bd692188-6da6-4387-a46d-003a2da0d0c8-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "bd692188-6da6-4387-a46d-003a2da0d0c8" (UID: "bd692188-6da6-4387-a46d-003a2da0d0c8"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:38:02 crc kubenswrapper[4897]: I1121 14:38:02.666853 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b513e4ee-7c30-4942-b997-4d52414e1ec0-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "b513e4ee-7c30-4942-b997-4d52414e1ec0" (UID: "b513e4ee-7c30-4942-b997-4d52414e1ec0"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:38:02 crc kubenswrapper[4897]: I1121 14:38:02.667270 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bd692188-6da6-4387-a46d-003a2da0d0c8-server-conf" (OuterVolumeSpecName: "server-conf") pod "bd692188-6da6-4387-a46d-003a2da0d0c8" (UID: "bd692188-6da6-4387-a46d-003a2da0d0c8"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:38:02 crc kubenswrapper[4897]: I1121 14:38:02.667430 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bd692188-6da6-4387-a46d-003a2da0d0c8-config-data" (OuterVolumeSpecName: "config-data") pod "bd692188-6da6-4387-a46d-003a2da0d0c8" (UID: "bd692188-6da6-4387-a46d-003a2da0d0c8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:38:02 crc kubenswrapper[4897]: I1121 14:38:02.667458 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/bd692188-6da6-4387-a46d-003a2da0d0c8-server-conf\") pod \"bd692188-6da6-4387-a46d-003a2da0d0c8\" (UID: \"bd692188-6da6-4387-a46d-003a2da0d0c8\") " Nov 21 14:38:02 crc kubenswrapper[4897]: I1121 14:38:02.667566 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bd692188-6da6-4387-a46d-003a2da0d0c8-config-data\") pod \"bd692188-6da6-4387-a46d-003a2da0d0c8\" (UID: \"bd692188-6da6-4387-a46d-003a2da0d0c8\") " Nov 21 14:38:02 crc kubenswrapper[4897]: I1121 14:38:02.667613 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b513e4ee-7c30-4942-b997-4d52414e1ec0-plugins-conf\") pod \"b513e4ee-7c30-4942-b997-4d52414e1ec0\" (UID: \"b513e4ee-7c30-4942-b997-4d52414e1ec0\") " Nov 21 14:38:02 crc kubenswrapper[4897]: W1121 14:38:02.667636 4897 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/bd692188-6da6-4387-a46d-003a2da0d0c8/volumes/kubernetes.io~configmap/server-conf Nov 21 14:38:02 crc kubenswrapper[4897]: I1121 14:38:02.667670 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bd692188-6da6-4387-a46d-003a2da0d0c8-server-conf" (OuterVolumeSpecName: "server-conf") pod "bd692188-6da6-4387-a46d-003a2da0d0c8" (UID: "bd692188-6da6-4387-a46d-003a2da0d0c8"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:38:02 crc kubenswrapper[4897]: W1121 14:38:02.667761 4897 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/bd692188-6da6-4387-a46d-003a2da0d0c8/volumes/kubernetes.io~configmap/config-data Nov 21 14:38:02 crc kubenswrapper[4897]: I1121 14:38:02.667785 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bd692188-6da6-4387-a46d-003a2da0d0c8-config-data" (OuterVolumeSpecName: "config-data") pod "bd692188-6da6-4387-a46d-003a2da0d0c8" (UID: "bd692188-6da6-4387-a46d-003a2da0d0c8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:38:02 crc kubenswrapper[4897]: W1121 14:38:02.668007 4897 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/b513e4ee-7c30-4942-b997-4d52414e1ec0/volumes/kubernetes.io~configmap/plugins-conf Nov 21 14:38:02 crc kubenswrapper[4897]: I1121 14:38:02.668035 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b513e4ee-7c30-4942-b997-4d52414e1ec0-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "b513e4ee-7c30-4942-b997-4d52414e1ec0" (UID: "b513e4ee-7c30-4942-b997-4d52414e1ec0"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:38:02 crc kubenswrapper[4897]: I1121 14:38:02.668269 4897 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bd692188-6da6-4387-a46d-003a2da0d0c8-config-data\") on node \"crc\" DevicePath \"\"" Nov 21 14:38:02 crc kubenswrapper[4897]: I1121 14:38:02.668291 4897 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b513e4ee-7c30-4942-b997-4d52414e1ec0-plugins-conf\") on node \"crc\" DevicePath \"\"" Nov 21 14:38:02 crc kubenswrapper[4897]: I1121 14:38:02.668303 4897 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/bd692188-6da6-4387-a46d-003a2da0d0c8-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Nov 21 14:38:02 crc kubenswrapper[4897]: I1121 14:38:02.668312 4897 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/bd692188-6da6-4387-a46d-003a2da0d0c8-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Nov 21 14:38:02 crc kubenswrapper[4897]: I1121 14:38:02.668320 4897 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/bd692188-6da6-4387-a46d-003a2da0d0c8-server-conf\") on node \"crc\" DevicePath \"\"" Nov 21 14:38:02 crc kubenswrapper[4897]: I1121 14:38:02.702428 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b513e4ee-7c30-4942-b997-4d52414e1ec0-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "b513e4ee-7c30-4942-b997-4d52414e1ec0" (UID: "b513e4ee-7c30-4942-b997-4d52414e1ec0"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:38:02 crc kubenswrapper[4897]: I1121 14:38:02.702837 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b513e4ee-7c30-4942-b997-4d52414e1ec0-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "b513e4ee-7c30-4942-b997-4d52414e1ec0" (UID: "b513e4ee-7c30-4942-b997-4d52414e1ec0"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:38:02 crc kubenswrapper[4897]: I1121 14:38:02.706019 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b513e4ee-7c30-4942-b997-4d52414e1ec0-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "b513e4ee-7c30-4942-b997-4d52414e1ec0" (UID: "b513e4ee-7c30-4942-b997-4d52414e1ec0"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:38:02 crc kubenswrapper[4897]: I1121 14:38:02.709346 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b513e4ee-7c30-4942-b997-4d52414e1ec0-server-conf" (OuterVolumeSpecName: "server-conf") pod "b513e4ee-7c30-4942-b997-4d52414e1ec0" (UID: "b513e4ee-7c30-4942-b997-4d52414e1ec0"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:38:02 crc kubenswrapper[4897]: I1121 14:38:02.736632 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd692188-6da6-4387-a46d-003a2da0d0c8-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "bd692188-6da6-4387-a46d-003a2da0d0c8" (UID: "bd692188-6da6-4387-a46d-003a2da0d0c8"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:38:02 crc kubenswrapper[4897]: I1121 14:38:02.770327 4897 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b513e4ee-7c30-4942-b997-4d52414e1ec0-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Nov 21 14:38:02 crc kubenswrapper[4897]: I1121 14:38:02.770368 4897 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b513e4ee-7c30-4942-b997-4d52414e1ec0-server-conf\") on node \"crc\" DevicePath \"\"" Nov 21 14:38:02 crc kubenswrapper[4897]: I1121 14:38:02.770381 4897 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b513e4ee-7c30-4942-b997-4d52414e1ec0-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Nov 21 14:38:02 crc kubenswrapper[4897]: I1121 14:38:02.770396 4897 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b513e4ee-7c30-4942-b997-4d52414e1ec0-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Nov 21 14:38:02 crc kubenswrapper[4897]: I1121 14:38:02.770411 4897 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/bd692188-6da6-4387-a46d-003a2da0d0c8-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Nov 21 14:38:02 crc kubenswrapper[4897]: I1121 14:38:02.999358 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.009816 4897 scope.go:117] "RemoveContainer" containerID="c3dd08ed558f92e886e50c86f9564fa7a7936c723ad747d6a561c7c2c73adc0b" Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.017057 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.040819 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.083092 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.103920 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Nov 21 14:38:03 crc kubenswrapper[4897]: E1121 14:38:03.104976 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b513e4ee-7c30-4942-b997-4d52414e1ec0" containerName="setup-container" Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.105004 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="b513e4ee-7c30-4942-b997-4d52414e1ec0" containerName="setup-container" Nov 21 14:38:03 crc kubenswrapper[4897]: E1121 14:38:03.105042 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd692188-6da6-4387-a46d-003a2da0d0c8" containerName="rabbitmq" Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.105051 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd692188-6da6-4387-a46d-003a2da0d0c8" containerName="rabbitmq" Nov 21 14:38:03 crc kubenswrapper[4897]: E1121 14:38:03.105132 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd692188-6da6-4387-a46d-003a2da0d0c8" containerName="setup-container" Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.105143 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd692188-6da6-4387-a46d-003a2da0d0c8" containerName="setup-container" Nov 21 14:38:03 crc kubenswrapper[4897]: E1121 14:38:03.105163 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b513e4ee-7c30-4942-b997-4d52414e1ec0" containerName="rabbitmq" Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.105170 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="b513e4ee-7c30-4942-b997-4d52414e1ec0" containerName="rabbitmq" Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.105829 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd692188-6da6-4387-a46d-003a2da0d0c8" containerName="rabbitmq" Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.105872 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="b513e4ee-7c30-4942-b997-4d52414e1ec0" containerName="rabbitmq" Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.108830 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.111042 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.114624 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.114880 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.114911 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.115261 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-kksgw" Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.115364 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.115475 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.121327 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.125194 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.129941 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.130280 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-w65rf" Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.130560 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.130920 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.131051 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.130960 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.134673 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.154685 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.179948 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.198800 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9fb2dc84-c75b-45e9-9873-081851fd8ea8-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"9fb2dc84-c75b-45e9-9873-081851fd8ea8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.198877 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/7a0fd0c8-15d2-45c0-9126-bf7e9e7926e8-server-conf\") pod \"rabbitmq-server-0\" (UID: \"7a0fd0c8-15d2-45c0-9126-bf7e9e7926e8\") " pod="openstack/rabbitmq-server-0" Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.198939 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d28k8\" (UniqueName: \"kubernetes.io/projected/9fb2dc84-c75b-45e9-9873-081851fd8ea8-kube-api-access-d28k8\") pod \"rabbitmq-cell1-server-0\" (UID: \"9fb2dc84-c75b-45e9-9873-081851fd8ea8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.199000 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9fb2dc84-c75b-45e9-9873-081851fd8ea8-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"9fb2dc84-c75b-45e9-9873-081851fd8ea8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.199017 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/7a0fd0c8-15d2-45c0-9126-bf7e9e7926e8-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"7a0fd0c8-15d2-45c0-9126-bf7e9e7926e8\") " pod="openstack/rabbitmq-server-0" Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.199034 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7a0fd0c8-15d2-45c0-9126-bf7e9e7926e8-config-data\") pod \"rabbitmq-server-0\" (UID: \"7a0fd0c8-15d2-45c0-9126-bf7e9e7926e8\") " pod="openstack/rabbitmq-server-0" Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.199058 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/7a0fd0c8-15d2-45c0-9126-bf7e9e7926e8-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"7a0fd0c8-15d2-45c0-9126-bf7e9e7926e8\") " pod="openstack/rabbitmq-server-0" Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.199081 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/7a0fd0c8-15d2-45c0-9126-bf7e9e7926e8-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"7a0fd0c8-15d2-45c0-9126-bf7e9e7926e8\") " pod="openstack/rabbitmq-server-0" Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.199123 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/7a0fd0c8-15d2-45c0-9126-bf7e9e7926e8-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"7a0fd0c8-15d2-45c0-9126-bf7e9e7926e8\") " pod="openstack/rabbitmq-server-0" Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.199148 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9fb2dc84-c75b-45e9-9873-081851fd8ea8-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"9fb2dc84-c75b-45e9-9873-081851fd8ea8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.199193 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"7a0fd0c8-15d2-45c0-9126-bf7e9e7926e8\") " pod="openstack/rabbitmq-server-0" Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.199416 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9fb2dc84-c75b-45e9-9873-081851fd8ea8-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"9fb2dc84-c75b-45e9-9873-081851fd8ea8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.199471 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/7a0fd0c8-15d2-45c0-9126-bf7e9e7926e8-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"7a0fd0c8-15d2-45c0-9126-bf7e9e7926e8\") " pod="openstack/rabbitmq-server-0" Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.199529 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/7a0fd0c8-15d2-45c0-9126-bf7e9e7926e8-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"7a0fd0c8-15d2-45c0-9126-bf7e9e7926e8\") " pod="openstack/rabbitmq-server-0" Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.199618 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9fb2dc84-c75b-45e9-9873-081851fd8ea8-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"9fb2dc84-c75b-45e9-9873-081851fd8ea8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.199857 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"9fb2dc84-c75b-45e9-9873-081851fd8ea8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.199940 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9fb2dc84-c75b-45e9-9873-081851fd8ea8-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"9fb2dc84-c75b-45e9-9873-081851fd8ea8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.200064 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9fb2dc84-c75b-45e9-9873-081851fd8ea8-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"9fb2dc84-c75b-45e9-9873-081851fd8ea8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.200100 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9fb2dc84-c75b-45e9-9873-081851fd8ea8-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"9fb2dc84-c75b-45e9-9873-081851fd8ea8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.200136 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/7a0fd0c8-15d2-45c0-9126-bf7e9e7926e8-pod-info\") pod \"rabbitmq-server-0\" (UID: \"7a0fd0c8-15d2-45c0-9126-bf7e9e7926e8\") " pod="openstack/rabbitmq-server-0" Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.200167 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-42wwm\" (UniqueName: \"kubernetes.io/projected/7a0fd0c8-15d2-45c0-9126-bf7e9e7926e8-kube-api-access-42wwm\") pod \"rabbitmq-server-0\" (UID: \"7a0fd0c8-15d2-45c0-9126-bf7e9e7926e8\") " pod="openstack/rabbitmq-server-0" Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.200209 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9fb2dc84-c75b-45e9-9873-081851fd8ea8-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"9fb2dc84-c75b-45e9-9873-081851fd8ea8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.302055 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"9fb2dc84-c75b-45e9-9873-081851fd8ea8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.302123 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9fb2dc84-c75b-45e9-9873-081851fd8ea8-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"9fb2dc84-c75b-45e9-9873-081851fd8ea8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.302159 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9fb2dc84-c75b-45e9-9873-081851fd8ea8-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"9fb2dc84-c75b-45e9-9873-081851fd8ea8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.302186 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9fb2dc84-c75b-45e9-9873-081851fd8ea8-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"9fb2dc84-c75b-45e9-9873-081851fd8ea8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.302217 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/7a0fd0c8-15d2-45c0-9126-bf7e9e7926e8-pod-info\") pod \"rabbitmq-server-0\" (UID: \"7a0fd0c8-15d2-45c0-9126-bf7e9e7926e8\") " pod="openstack/rabbitmq-server-0" Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.302243 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-42wwm\" (UniqueName: \"kubernetes.io/projected/7a0fd0c8-15d2-45c0-9126-bf7e9e7926e8-kube-api-access-42wwm\") pod \"rabbitmq-server-0\" (UID: \"7a0fd0c8-15d2-45c0-9126-bf7e9e7926e8\") " pod="openstack/rabbitmq-server-0" Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.302267 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9fb2dc84-c75b-45e9-9873-081851fd8ea8-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"9fb2dc84-c75b-45e9-9873-081851fd8ea8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.302306 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9fb2dc84-c75b-45e9-9873-081851fd8ea8-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"9fb2dc84-c75b-45e9-9873-081851fd8ea8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.302338 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/7a0fd0c8-15d2-45c0-9126-bf7e9e7926e8-server-conf\") pod \"rabbitmq-server-0\" (UID: \"7a0fd0c8-15d2-45c0-9126-bf7e9e7926e8\") " pod="openstack/rabbitmq-server-0" Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.302378 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d28k8\" (UniqueName: \"kubernetes.io/projected/9fb2dc84-c75b-45e9-9873-081851fd8ea8-kube-api-access-d28k8\") pod \"rabbitmq-cell1-server-0\" (UID: \"9fb2dc84-c75b-45e9-9873-081851fd8ea8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.302409 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9fb2dc84-c75b-45e9-9873-081851fd8ea8-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"9fb2dc84-c75b-45e9-9873-081851fd8ea8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.302434 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/7a0fd0c8-15d2-45c0-9126-bf7e9e7926e8-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"7a0fd0c8-15d2-45c0-9126-bf7e9e7926e8\") " pod="openstack/rabbitmq-server-0" Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.302456 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7a0fd0c8-15d2-45c0-9126-bf7e9e7926e8-config-data\") pod \"rabbitmq-server-0\" (UID: \"7a0fd0c8-15d2-45c0-9126-bf7e9e7926e8\") " pod="openstack/rabbitmq-server-0" Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.302485 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/7a0fd0c8-15d2-45c0-9126-bf7e9e7926e8-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"7a0fd0c8-15d2-45c0-9126-bf7e9e7926e8\") " pod="openstack/rabbitmq-server-0" Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.302626 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/7a0fd0c8-15d2-45c0-9126-bf7e9e7926e8-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"7a0fd0c8-15d2-45c0-9126-bf7e9e7926e8\") " pod="openstack/rabbitmq-server-0" Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.302677 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/7a0fd0c8-15d2-45c0-9126-bf7e9e7926e8-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"7a0fd0c8-15d2-45c0-9126-bf7e9e7926e8\") " pod="openstack/rabbitmq-server-0" Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.302744 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9fb2dc84-c75b-45e9-9873-081851fd8ea8-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"9fb2dc84-c75b-45e9-9873-081851fd8ea8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.302782 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"7a0fd0c8-15d2-45c0-9126-bf7e9e7926e8\") " pod="openstack/rabbitmq-server-0" Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.302833 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9fb2dc84-c75b-45e9-9873-081851fd8ea8-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"9fb2dc84-c75b-45e9-9873-081851fd8ea8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.302856 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/7a0fd0c8-15d2-45c0-9126-bf7e9e7926e8-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"7a0fd0c8-15d2-45c0-9126-bf7e9e7926e8\") " pod="openstack/rabbitmq-server-0" Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.302882 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/7a0fd0c8-15d2-45c0-9126-bf7e9e7926e8-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"7a0fd0c8-15d2-45c0-9126-bf7e9e7926e8\") " pod="openstack/rabbitmq-server-0" Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.302929 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9fb2dc84-c75b-45e9-9873-081851fd8ea8-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"9fb2dc84-c75b-45e9-9873-081851fd8ea8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.305101 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7a0fd0c8-15d2-45c0-9126-bf7e9e7926e8-config-data\") pod \"rabbitmq-server-0\" (UID: \"7a0fd0c8-15d2-45c0-9126-bf7e9e7926e8\") " pod="openstack/rabbitmq-server-0" Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.306192 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/7a0fd0c8-15d2-45c0-9126-bf7e9e7926e8-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"7a0fd0c8-15d2-45c0-9126-bf7e9e7926e8\") " pod="openstack/rabbitmq-server-0" Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.306432 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9fb2dc84-c75b-45e9-9873-081851fd8ea8-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"9fb2dc84-c75b-45e9-9873-081851fd8ea8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.306437 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9fb2dc84-c75b-45e9-9873-081851fd8ea8-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"9fb2dc84-c75b-45e9-9873-081851fd8ea8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.306757 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/7a0fd0c8-15d2-45c0-9126-bf7e9e7926e8-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"7a0fd0c8-15d2-45c0-9126-bf7e9e7926e8\") " pod="openstack/rabbitmq-server-0" Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.306912 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9fb2dc84-c75b-45e9-9873-081851fd8ea8-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"9fb2dc84-c75b-45e9-9873-081851fd8ea8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.307123 4897 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"7a0fd0c8-15d2-45c0-9126-bf7e9e7926e8\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/rabbitmq-server-0" Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.307149 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9fb2dc84-c75b-45e9-9873-081851fd8ea8-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"9fb2dc84-c75b-45e9-9873-081851fd8ea8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.307243 4897 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"9fb2dc84-c75b-45e9-9873-081851fd8ea8\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/rabbitmq-cell1-server-0" Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.309663 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9fb2dc84-c75b-45e9-9873-081851fd8ea8-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"9fb2dc84-c75b-45e9-9873-081851fd8ea8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.310022 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/7a0fd0c8-15d2-45c0-9126-bf7e9e7926e8-pod-info\") pod \"rabbitmq-server-0\" (UID: \"7a0fd0c8-15d2-45c0-9126-bf7e9e7926e8\") " pod="openstack/rabbitmq-server-0" Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.310336 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/7a0fd0c8-15d2-45c0-9126-bf7e9e7926e8-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"7a0fd0c8-15d2-45c0-9126-bf7e9e7926e8\") " pod="openstack/rabbitmq-server-0" Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.310770 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9fb2dc84-c75b-45e9-9873-081851fd8ea8-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"9fb2dc84-c75b-45e9-9873-081851fd8ea8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.310885 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/7a0fd0c8-15d2-45c0-9126-bf7e9e7926e8-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"7a0fd0c8-15d2-45c0-9126-bf7e9e7926e8\") " pod="openstack/rabbitmq-server-0" Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.311312 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9fb2dc84-c75b-45e9-9873-081851fd8ea8-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"9fb2dc84-c75b-45e9-9873-081851fd8ea8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.311351 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9fb2dc84-c75b-45e9-9873-081851fd8ea8-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"9fb2dc84-c75b-45e9-9873-081851fd8ea8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.312162 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9fb2dc84-c75b-45e9-9873-081851fd8ea8-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"9fb2dc84-c75b-45e9-9873-081851fd8ea8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.312312 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/7a0fd0c8-15d2-45c0-9126-bf7e9e7926e8-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"7a0fd0c8-15d2-45c0-9126-bf7e9e7926e8\") " pod="openstack/rabbitmq-server-0" Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.314121 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/7a0fd0c8-15d2-45c0-9126-bf7e9e7926e8-server-conf\") pod \"rabbitmq-server-0\" (UID: \"7a0fd0c8-15d2-45c0-9126-bf7e9e7926e8\") " pod="openstack/rabbitmq-server-0" Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.315718 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/7a0fd0c8-15d2-45c0-9126-bf7e9e7926e8-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"7a0fd0c8-15d2-45c0-9126-bf7e9e7926e8\") " pod="openstack/rabbitmq-server-0" Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.325649 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-42wwm\" (UniqueName: \"kubernetes.io/projected/7a0fd0c8-15d2-45c0-9126-bf7e9e7926e8-kube-api-access-42wwm\") pod \"rabbitmq-server-0\" (UID: \"7a0fd0c8-15d2-45c0-9126-bf7e9e7926e8\") " pod="openstack/rabbitmq-server-0" Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.331911 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d28k8\" (UniqueName: \"kubernetes.io/projected/9fb2dc84-c75b-45e9-9873-081851fd8ea8-kube-api-access-d28k8\") pod \"rabbitmq-cell1-server-0\" (UID: \"9fb2dc84-c75b-45e9-9873-081851fd8ea8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.440744 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"9fb2dc84-c75b-45e9-9873-081851fd8ea8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.456158 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.715991 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"7a0fd0c8-15d2-45c0-9126-bf7e9e7926e8\") " pod="openstack/rabbitmq-server-0" Nov 21 14:38:03 crc kubenswrapper[4897]: I1121 14:38:03.730021 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 21 14:38:04 crc kubenswrapper[4897]: I1121 14:38:04.108093 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b513e4ee-7c30-4942-b997-4d52414e1ec0" path="/var/lib/kubelet/pods/b513e4ee-7c30-4942-b997-4d52414e1ec0/volumes" Nov 21 14:38:04 crc kubenswrapper[4897]: I1121 14:38:04.157992 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd692188-6da6-4387-a46d-003a2da0d0c8" path="/var/lib/kubelet/pods/bd692188-6da6-4387-a46d-003a2da0d0c8/volumes" Nov 21 14:38:04 crc kubenswrapper[4897]: E1121 14:38:04.686723 4897 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = reading blob sha256:e7dd6fde598cae188aa8cee8fb9ba9ec8bdfb0cf60f8f8b85382265d6c51c5fd: Get \"https://quay.rdoproject.org/v2/podified-master-centos10/openstack-ceilometer-central/blobs/sha256:e7dd6fde598cae188aa8cee8fb9ba9ec8bdfb0cf60f8f8b85382265d6c51c5fd\": context canceled" image="quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested" Nov 21 14:38:04 crc kubenswrapper[4897]: E1121 14:38:04.686812 4897 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = reading blob sha256:e7dd6fde598cae188aa8cee8fb9ba9ec8bdfb0cf60f8f8b85382265d6c51c5fd: Get \"https://quay.rdoproject.org/v2/podified-master-centos10/openstack-ceilometer-central/blobs/sha256:e7dd6fde598cae188aa8cee8fb9ba9ec8bdfb0cf60f8f8b85382265d6c51c5fd\": context canceled" image="quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested" Nov 21 14:38:04 crc kubenswrapper[4897]: E1121 14:38:04.687012 4897 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n5bhcfh55fh65dh666h65fh66chc7hddh6fhddh7ch665hb6hcch584hfch698h5c5h54fh565h669h68ch5h96h57fh6dh5c9hfdh5fbh5fch59cq,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-k2qzq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(c628be92-b88f-4e02-a981-4b8fe6fd7c13): ErrImagePull: rpc error: code = Canceled desc = reading blob sha256:e7dd6fde598cae188aa8cee8fb9ba9ec8bdfb0cf60f8f8b85382265d6c51c5fd: Get \"https://quay.rdoproject.org/v2/podified-master-centos10/openstack-ceilometer-central/blobs/sha256:e7dd6fde598cae188aa8cee8fb9ba9ec8bdfb0cf60f8f8b85382265d6c51c5fd\": context canceled" logger="UnhandledError" Nov 21 14:38:04 crc kubenswrapper[4897]: I1121 14:38:04.808771 4897 scope.go:117] "RemoveContainer" containerID="5e97248a8dc790521af481c0714596ab753484b35fa256f425d03e7bac6e870a" Nov 21 14:38:05 crc kubenswrapper[4897]: I1121 14:38:05.089710 4897 scope.go:117] "RemoveContainer" containerID="c68f1b003d12b4a2964169a85a04e3f0272d86c891f52e6bebd52f3781520d1a" Nov 21 14:38:05 crc kubenswrapper[4897]: E1121 14:38:05.090387 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 14:38:05 crc kubenswrapper[4897]: I1121 14:38:05.410100 4897 scope.go:117] "RemoveContainer" containerID="4ded3f218b2e102f2165dac36295117473edb574550d19babb7b246c41c8316f" Nov 21 14:38:05 crc kubenswrapper[4897]: I1121 14:38:05.521411 4897 scope.go:117] "RemoveContainer" containerID="4ded3f218b2e102f2165dac36295117473edb574550d19babb7b246c41c8316f" Nov 21 14:38:05 crc kubenswrapper[4897]: I1121 14:38:05.690006 4897 scope.go:117] "RemoveContainer" containerID="99656890d61e67a0d062b6637a03db823a2e86706363f8d471ea687b544eb8ea" Nov 21 14:38:05 crc kubenswrapper[4897]: E1121 14:38:05.690001 4897 log.go:32] "RemoveContainer from runtime service failed" err="rpc error: code = Unknown desc = failed to delete container k8s_rabbitmq_rabbitmq-server-0_openstack_b513e4ee-7c30-4942-b997-4d52414e1ec0_0 in pod sandbox d57051f4a2d9e61d4d7988ce611e9c0574d0451a505583e0be5558754171da2b from index: no such id: '4ded3f218b2e102f2165dac36295117473edb574550d19babb7b246c41c8316f'" containerID="4ded3f218b2e102f2165dac36295117473edb574550d19babb7b246c41c8316f" Nov 21 14:38:05 crc kubenswrapper[4897]: I1121 14:38:05.690340 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ded3f218b2e102f2165dac36295117473edb574550d19babb7b246c41c8316f"} err="rpc error: code = Unknown desc = failed to delete container k8s_rabbitmq_rabbitmq-server-0_openstack_b513e4ee-7c30-4942-b997-4d52414e1ec0_0 in pod sandbox d57051f4a2d9e61d4d7988ce611e9c0574d0451a505583e0be5558754171da2b from index: no such id: '4ded3f218b2e102f2165dac36295117473edb574550d19babb7b246c41c8316f'" Nov 21 14:38:05 crc kubenswrapper[4897]: I1121 14:38:05.690367 4897 scope.go:117] "RemoveContainer" containerID="3140f08ac469f99b17c2aa56cad1a0ccf3db6d5b4c5110f9454241ae6294b460" Nov 21 14:38:05 crc kubenswrapper[4897]: E1121 14:38:05.692481 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"99656890d61e67a0d062b6637a03db823a2e86706363f8d471ea687b544eb8ea\": container with ID starting with 99656890d61e67a0d062b6637a03db823a2e86706363f8d471ea687b544eb8ea not found: ID does not exist" containerID="99656890d61e67a0d062b6637a03db823a2e86706363f8d471ea687b544eb8ea" Nov 21 14:38:05 crc kubenswrapper[4897]: E1121 14:38:05.692565 4897 kuberuntime_gc.go:150] "Failed to remove container" err="failed to get container status \"99656890d61e67a0d062b6637a03db823a2e86706363f8d471ea687b544eb8ea\": rpc error: code = NotFound desc = could not find container \"99656890d61e67a0d062b6637a03db823a2e86706363f8d471ea687b544eb8ea\": container with ID starting with 99656890d61e67a0d062b6637a03db823a2e86706363f8d471ea687b544eb8ea not found: ID does not exist" containerID="99656890d61e67a0d062b6637a03db823a2e86706363f8d471ea687b544eb8ea" Nov 21 14:38:05 crc kubenswrapper[4897]: I1121 14:38:05.692612 4897 scope.go:117] "RemoveContainer" containerID="5e97248a8dc790521af481c0714596ab753484b35fa256f425d03e7bac6e870a" Nov 21 14:38:05 crc kubenswrapper[4897]: E1121 14:38:05.693566 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5e97248a8dc790521af481c0714596ab753484b35fa256f425d03e7bac6e870a\": container with ID starting with 5e97248a8dc790521af481c0714596ab753484b35fa256f425d03e7bac6e870a not found: ID does not exist" containerID="5e97248a8dc790521af481c0714596ab753484b35fa256f425d03e7bac6e870a" Nov 21 14:38:05 crc kubenswrapper[4897]: E1121 14:38:05.693637 4897 kuberuntime_gc.go:150] "Failed to remove container" err="failed to get container status \"5e97248a8dc790521af481c0714596ab753484b35fa256f425d03e7bac6e870a\": rpc error: code = NotFound desc = could not find container \"5e97248a8dc790521af481c0714596ab753484b35fa256f425d03e7bac6e870a\": container with ID starting with 5e97248a8dc790521af481c0714596ab753484b35fa256f425d03e7bac6e870a not found: ID does not exist" containerID="5e97248a8dc790521af481c0714596ab753484b35fa256f425d03e7bac6e870a" Nov 21 14:38:05 crc kubenswrapper[4897]: I1121 14:38:05.693661 4897 scope.go:117] "RemoveContainer" containerID="3140f08ac469f99b17c2aa56cad1a0ccf3db6d5b4c5110f9454241ae6294b460" Nov 21 14:38:05 crc kubenswrapper[4897]: E1121 14:38:05.814666 4897 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested" Nov 21 14:38:05 crc kubenswrapper[4897]: E1121 14:38:05.814716 4897 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested" Nov 21 14:38:05 crc kubenswrapper[4897]: E1121 14:38:05.814831 4897 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:heat-db-sync,Image:quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested,Command:[/bin/bash],Args:[-c /usr/bin/heat-manage --config-dir /etc/heat/heat.conf.d db_sync],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/heat/heat.conf.d/00-default.conf,SubPath:00-default.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/heat/heat.conf.d/01-custom.conf,SubPath:01-custom.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-th9pr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42418,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*42418,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-db-sync-m55sf_openstack(c4c479c7-7728-41fc-bf98-ce2dcf800840): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 21 14:38:05 crc kubenswrapper[4897]: E1121 14:38:05.816075 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/heat-db-sync-m55sf" podUID="c4c479c7-7728-41fc-bf98-ce2dcf800840" Nov 21 14:38:05 crc kubenswrapper[4897]: E1121 14:38:05.907676 4897 log.go:32] "RemoveContainer from runtime service failed" err="rpc error: code = Unknown desc = failed to delete container k8s_setup-container_rabbitmq-server-0_openstack_b513e4ee-7c30-4942-b997-4d52414e1ec0_0 in pod sandbox d57051f4a2d9e61d4d7988ce611e9c0574d0451a505583e0be5558754171da2b: identifier is not a container" containerID="3140f08ac469f99b17c2aa56cad1a0ccf3db6d5b4c5110f9454241ae6294b460" Nov 21 14:38:05 crc kubenswrapper[4897]: E1121 14:38:05.907729 4897 kuberuntime_gc.go:150] "Failed to remove container" err="rpc error: code = Unknown desc = failed to delete container k8s_setup-container_rabbitmq-server-0_openstack_b513e4ee-7c30-4942-b997-4d52414e1ec0_0 in pod sandbox d57051f4a2d9e61d4d7988ce611e9c0574d0451a505583e0be5558754171da2b: identifier is not a container" containerID="3140f08ac469f99b17c2aa56cad1a0ccf3db6d5b4c5110f9454241ae6294b460" Nov 21 14:38:05 crc kubenswrapper[4897]: I1121 14:38:05.994819 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 21 14:38:06 crc kubenswrapper[4897]: I1121 14:38:06.009860 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 21 14:38:06 crc kubenswrapper[4897]: W1121 14:38:06.063087 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9fb2dc84_c75b_45e9_9873_081851fd8ea8.slice/crio-1681ff38e631731f4b8c16e60ddbb794d0b1b1219475c36ce6d15880f195a42a WatchSource:0}: Error finding container 1681ff38e631731f4b8c16e60ddbb794d0b1b1219475c36ce6d15880f195a42a: Status 404 returned error can't find the container with id 1681ff38e631731f4b8c16e60ddbb794d0b1b1219475c36ce6d15880f195a42a Nov 21 14:38:06 crc kubenswrapper[4897]: W1121 14:38:06.066052 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7a0fd0c8_15d2_45c0_9126_bf7e9e7926e8.slice/crio-f7b449e7869687aa6fc56ecc541c523af16e30163f4f7eaeedb7ea5a871bbeb3 WatchSource:0}: Error finding container f7b449e7869687aa6fc56ecc541c523af16e30163f4f7eaeedb7ea5a871bbeb3: Status 404 returned error can't find the container with id f7b449e7869687aa6fc56ecc541c523af16e30163f4f7eaeedb7ea5a871bbeb3 Nov 21 14:38:06 crc kubenswrapper[4897]: I1121 14:38:06.543987 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"9fb2dc84-c75b-45e9-9873-081851fd8ea8","Type":"ContainerStarted","Data":"1681ff38e631731f4b8c16e60ddbb794d0b1b1219475c36ce6d15880f195a42a"} Nov 21 14:38:06 crc kubenswrapper[4897]: I1121 14:38:06.547317 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"7a0fd0c8-15d2-45c0-9126-bf7e9e7926e8","Type":"ContainerStarted","Data":"f7b449e7869687aa6fc56ecc541c523af16e30163f4f7eaeedb7ea5a871bbeb3"} Nov 21 14:38:06 crc kubenswrapper[4897]: E1121 14:38:06.838441 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-m55sf" podUID="c4c479c7-7728-41fc-bf98-ce2dcf800840" Nov 21 14:38:10 crc kubenswrapper[4897]: I1121 14:38:10.603114 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"9fb2dc84-c75b-45e9-9873-081851fd8ea8","Type":"ContainerStarted","Data":"fed7b5464d8f589ebaee0f3363e246a6c9c10b313f04b7a18a572e4c4cbb0643"} Nov 21 14:38:10 crc kubenswrapper[4897]: I1121 14:38:10.605416 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"7a0fd0c8-15d2-45c0-9126-bf7e9e7926e8","Type":"ContainerStarted","Data":"e1f103b3423928d585b0c7cb85bbb2e1b44e99f2bb7a58a4ad2d8b9c13d89fff"} Nov 21 14:38:12 crc kubenswrapper[4897]: I1121 14:38:12.633654 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c628be92-b88f-4e02-a981-4b8fe6fd7c13","Type":"ContainerStarted","Data":"18ab8c7043c2c7119293c3ba8eae8c554737641d5cffdd36825d922495861d84"} Nov 21 14:38:18 crc kubenswrapper[4897]: I1121 14:38:18.704811 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c628be92-b88f-4e02-a981-4b8fe6fd7c13","Type":"ContainerStarted","Data":"8b9f2d5ce482cc848d33b478a6ba7cf699bfb5e680bdfb81fd4ae2f4891e35fc"} Nov 21 14:38:19 crc kubenswrapper[4897]: I1121 14:38:19.089163 4897 scope.go:117] "RemoveContainer" containerID="c68f1b003d12b4a2964169a85a04e3f0272d86c891f52e6bebd52f3781520d1a" Nov 21 14:38:19 crc kubenswrapper[4897]: E1121 14:38:19.089621 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 14:38:24 crc kubenswrapper[4897]: I1121 14:38:24.298260 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-2dwkg"] Nov 21 14:38:25 crc kubenswrapper[4897]: I1121 14:38:24.302493 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2dwkg" Nov 21 14:38:25 crc kubenswrapper[4897]: I1121 14:38:24.310626 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2dwkg"] Nov 21 14:38:25 crc kubenswrapper[4897]: I1121 14:38:24.380899 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a5df1f3-8748-4324-bf70-bf50dc091710-catalog-content\") pod \"redhat-marketplace-2dwkg\" (UID: \"6a5df1f3-8748-4324-bf70-bf50dc091710\") " pod="openshift-marketplace/redhat-marketplace-2dwkg" Nov 21 14:38:25 crc kubenswrapper[4897]: I1121 14:38:24.381081 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a5df1f3-8748-4324-bf70-bf50dc091710-utilities\") pod \"redhat-marketplace-2dwkg\" (UID: \"6a5df1f3-8748-4324-bf70-bf50dc091710\") " pod="openshift-marketplace/redhat-marketplace-2dwkg" Nov 21 14:38:25 crc kubenswrapper[4897]: I1121 14:38:24.381609 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ffz2d\" (UniqueName: \"kubernetes.io/projected/6a5df1f3-8748-4324-bf70-bf50dc091710-kube-api-access-ffz2d\") pod \"redhat-marketplace-2dwkg\" (UID: \"6a5df1f3-8748-4324-bf70-bf50dc091710\") " pod="openshift-marketplace/redhat-marketplace-2dwkg" Nov 21 14:38:25 crc kubenswrapper[4897]: I1121 14:38:24.483384 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a5df1f3-8748-4324-bf70-bf50dc091710-catalog-content\") pod \"redhat-marketplace-2dwkg\" (UID: \"6a5df1f3-8748-4324-bf70-bf50dc091710\") " pod="openshift-marketplace/redhat-marketplace-2dwkg" Nov 21 14:38:25 crc kubenswrapper[4897]: I1121 14:38:24.483823 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a5df1f3-8748-4324-bf70-bf50dc091710-utilities\") pod \"redhat-marketplace-2dwkg\" (UID: \"6a5df1f3-8748-4324-bf70-bf50dc091710\") " pod="openshift-marketplace/redhat-marketplace-2dwkg" Nov 21 14:38:25 crc kubenswrapper[4897]: I1121 14:38:24.483942 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a5df1f3-8748-4324-bf70-bf50dc091710-catalog-content\") pod \"redhat-marketplace-2dwkg\" (UID: \"6a5df1f3-8748-4324-bf70-bf50dc091710\") " pod="openshift-marketplace/redhat-marketplace-2dwkg" Nov 21 14:38:25 crc kubenswrapper[4897]: I1121 14:38:24.484174 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a5df1f3-8748-4324-bf70-bf50dc091710-utilities\") pod \"redhat-marketplace-2dwkg\" (UID: \"6a5df1f3-8748-4324-bf70-bf50dc091710\") " pod="openshift-marketplace/redhat-marketplace-2dwkg" Nov 21 14:38:25 crc kubenswrapper[4897]: I1121 14:38:24.484295 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ffz2d\" (UniqueName: \"kubernetes.io/projected/6a5df1f3-8748-4324-bf70-bf50dc091710-kube-api-access-ffz2d\") pod \"redhat-marketplace-2dwkg\" (UID: \"6a5df1f3-8748-4324-bf70-bf50dc091710\") " pod="openshift-marketplace/redhat-marketplace-2dwkg" Nov 21 14:38:25 crc kubenswrapper[4897]: I1121 14:38:24.505856 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ffz2d\" (UniqueName: \"kubernetes.io/projected/6a5df1f3-8748-4324-bf70-bf50dc091710-kube-api-access-ffz2d\") pod \"redhat-marketplace-2dwkg\" (UID: \"6a5df1f3-8748-4324-bf70-bf50dc091710\") " pod="openshift-marketplace/redhat-marketplace-2dwkg" Nov 21 14:38:25 crc kubenswrapper[4897]: I1121 14:38:24.638148 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2dwkg" Nov 21 14:38:25 crc kubenswrapper[4897]: I1121 14:38:25.916790 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-m55sf" event={"ID":"c4c479c7-7728-41fc-bf98-ce2dcf800840","Type":"ContainerStarted","Data":"c43eba824c1aabd68653c0fb2ebca80d3784c2763791bff3f6d120f3daaefcb6"} Nov 21 14:38:26 crc kubenswrapper[4897]: I1121 14:38:26.692238 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-d8nth"] Nov 21 14:38:26 crc kubenswrapper[4897]: I1121 14:38:26.694909 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-d8nth" Nov 21 14:38:26 crc kubenswrapper[4897]: I1121 14:38:26.706123 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-d8nth"] Nov 21 14:38:26 crc kubenswrapper[4897]: I1121 14:38:26.735386 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/985eca1a-139e-46ce-a2a9-3e68d76b84d0-catalog-content\") pod \"community-operators-d8nth\" (UID: \"985eca1a-139e-46ce-a2a9-3e68d76b84d0\") " pod="openshift-marketplace/community-operators-d8nth" Nov 21 14:38:26 crc kubenswrapper[4897]: I1121 14:38:26.735696 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/985eca1a-139e-46ce-a2a9-3e68d76b84d0-utilities\") pod \"community-operators-d8nth\" (UID: \"985eca1a-139e-46ce-a2a9-3e68d76b84d0\") " pod="openshift-marketplace/community-operators-d8nth" Nov 21 14:38:26 crc kubenswrapper[4897]: I1121 14:38:26.735748 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wqcw6\" (UniqueName: \"kubernetes.io/projected/985eca1a-139e-46ce-a2a9-3e68d76b84d0-kube-api-access-wqcw6\") pod \"community-operators-d8nth\" (UID: \"985eca1a-139e-46ce-a2a9-3e68d76b84d0\") " pod="openshift-marketplace/community-operators-d8nth" Nov 21 14:38:26 crc kubenswrapper[4897]: I1121 14:38:26.837355 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/985eca1a-139e-46ce-a2a9-3e68d76b84d0-utilities\") pod \"community-operators-d8nth\" (UID: \"985eca1a-139e-46ce-a2a9-3e68d76b84d0\") " pod="openshift-marketplace/community-operators-d8nth" Nov 21 14:38:26 crc kubenswrapper[4897]: I1121 14:38:26.837419 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wqcw6\" (UniqueName: \"kubernetes.io/projected/985eca1a-139e-46ce-a2a9-3e68d76b84d0-kube-api-access-wqcw6\") pod \"community-operators-d8nth\" (UID: \"985eca1a-139e-46ce-a2a9-3e68d76b84d0\") " pod="openshift-marketplace/community-operators-d8nth" Nov 21 14:38:26 crc kubenswrapper[4897]: I1121 14:38:26.837528 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/985eca1a-139e-46ce-a2a9-3e68d76b84d0-catalog-content\") pod \"community-operators-d8nth\" (UID: \"985eca1a-139e-46ce-a2a9-3e68d76b84d0\") " pod="openshift-marketplace/community-operators-d8nth" Nov 21 14:38:26 crc kubenswrapper[4897]: I1121 14:38:26.837984 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/985eca1a-139e-46ce-a2a9-3e68d76b84d0-utilities\") pod \"community-operators-d8nth\" (UID: \"985eca1a-139e-46ce-a2a9-3e68d76b84d0\") " pod="openshift-marketplace/community-operators-d8nth" Nov 21 14:38:26 crc kubenswrapper[4897]: I1121 14:38:26.838059 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/985eca1a-139e-46ce-a2a9-3e68d76b84d0-catalog-content\") pod \"community-operators-d8nth\" (UID: \"985eca1a-139e-46ce-a2a9-3e68d76b84d0\") " pod="openshift-marketplace/community-operators-d8nth" Nov 21 14:38:26 crc kubenswrapper[4897]: I1121 14:38:26.869174 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wqcw6\" (UniqueName: \"kubernetes.io/projected/985eca1a-139e-46ce-a2a9-3e68d76b84d0-kube-api-access-wqcw6\") pod \"community-operators-d8nth\" (UID: \"985eca1a-139e-46ce-a2a9-3e68d76b84d0\") " pod="openshift-marketplace/community-operators-d8nth" Nov 21 14:38:27 crc kubenswrapper[4897]: I1121 14:38:27.023653 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-d8nth" Nov 21 14:38:30 crc kubenswrapper[4897]: I1121 14:38:30.041828 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-db-sync-m55sf" podStartSLOduration=7.280378737 podStartE2EDuration="1m18.041809861s" podCreationTimestamp="2025-11-21 14:37:12 +0000 UTC" firstStartedPulling="2025-11-21 14:37:13.570229921 +0000 UTC m=+1710.854823396" lastFinishedPulling="2025-11-21 14:38:24.331661045 +0000 UTC m=+1781.616254520" observedRunningTime="2025-11-21 14:38:26.951844907 +0000 UTC m=+1784.236438392" watchObservedRunningTime="2025-11-21 14:38:30.041809861 +0000 UTC m=+1787.326403336" Nov 21 14:38:30 crc kubenswrapper[4897]: I1121 14:38:30.051797 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-d8nth"] Nov 21 14:38:30 crc kubenswrapper[4897]: I1121 14:38:30.173334 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2dwkg"] Nov 21 14:38:30 crc kubenswrapper[4897]: W1121 14:38:30.204888 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6a5df1f3_8748_4324_bf70_bf50dc091710.slice/crio-d9fa5c1a658cfbf528cddc9fd0c58c8f47e0af83e0729722328e41a07e9cd5aa WatchSource:0}: Error finding container d9fa5c1a658cfbf528cddc9fd0c58c8f47e0af83e0729722328e41a07e9cd5aa: Status 404 returned error can't find the container with id d9fa5c1a658cfbf528cddc9fd0c58c8f47e0af83e0729722328e41a07e9cd5aa Nov 21 14:38:30 crc kubenswrapper[4897]: I1121 14:38:30.999263 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2dwkg" event={"ID":"6a5df1f3-8748-4324-bf70-bf50dc091710","Type":"ContainerStarted","Data":"d9fa5c1a658cfbf528cddc9fd0c58c8f47e0af83e0729722328e41a07e9cd5aa"} Nov 21 14:38:31 crc kubenswrapper[4897]: I1121 14:38:31.000935 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d8nth" event={"ID":"985eca1a-139e-46ce-a2a9-3e68d76b84d0","Type":"ContainerStarted","Data":"759a1e9e975e70246247f659c1fff33f80e45802f416023cd91a8765a8fde2b6"} Nov 21 14:38:31 crc kubenswrapper[4897]: I1121 14:38:31.003137 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c628be92-b88f-4e02-a981-4b8fe6fd7c13","Type":"ContainerStarted","Data":"3c29e2800e41c1355f2252c2716f4f964d057f6035e22cee95ec3644eec31f4c"} Nov 21 14:38:31 crc kubenswrapper[4897]: I1121 14:38:31.089583 4897 scope.go:117] "RemoveContainer" containerID="c68f1b003d12b4a2964169a85a04e3f0272d86c891f52e6bebd52f3781520d1a" Nov 21 14:38:31 crc kubenswrapper[4897]: E1121 14:38:31.089857 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 14:38:31 crc kubenswrapper[4897]: E1121 14:38:31.898015 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = reading blob sha256:e7dd6fde598cae188aa8cee8fb9ba9ec8bdfb0cf60f8f8b85382265d6c51c5fd: Get \\\"https://quay.rdoproject.org/v2/podified-master-centos10/openstack-ceilometer-central/blobs/sha256:e7dd6fde598cae188aa8cee8fb9ba9ec8bdfb0cf60f8f8b85382265d6c51c5fd\\\": context canceled\"" pod="openstack/ceilometer-0" podUID="c628be92-b88f-4e02-a981-4b8fe6fd7c13" Nov 21 14:38:32 crc kubenswrapper[4897]: I1121 14:38:32.017780 4897 generic.go:334] "Generic (PLEG): container finished" podID="6a5df1f3-8748-4324-bf70-bf50dc091710" containerID="f765c507ff764d2639397755ebd799d092854e76e4abf34bed549d52af84ea43" exitCode=0 Nov 21 14:38:32 crc kubenswrapper[4897]: I1121 14:38:32.018074 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2dwkg" event={"ID":"6a5df1f3-8748-4324-bf70-bf50dc091710","Type":"ContainerDied","Data":"f765c507ff764d2639397755ebd799d092854e76e4abf34bed549d52af84ea43"} Nov 21 14:38:32 crc kubenswrapper[4897]: I1121 14:38:32.020466 4897 generic.go:334] "Generic (PLEG): container finished" podID="985eca1a-139e-46ce-a2a9-3e68d76b84d0" containerID="fe610035df51ba9b7ec1dd836f736c0f477e2aaee4d035bbc2b9dd97f160dd34" exitCode=0 Nov 21 14:38:32 crc kubenswrapper[4897]: I1121 14:38:32.020599 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d8nth" event={"ID":"985eca1a-139e-46ce-a2a9-3e68d76b84d0","Type":"ContainerDied","Data":"fe610035df51ba9b7ec1dd836f736c0f477e2aaee4d035bbc2b9dd97f160dd34"} Nov 21 14:38:32 crc kubenswrapper[4897]: I1121 14:38:32.020709 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 21 14:38:32 crc kubenswrapper[4897]: I1121 14:38:32.180164 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-594cb89c79-s76zb"] Nov 21 14:38:32 crc kubenswrapper[4897]: I1121 14:38:32.184707 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-594cb89c79-s76zb" Nov 21 14:38:32 crc kubenswrapper[4897]: I1121 14:38:32.189112 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Nov 21 14:38:32 crc kubenswrapper[4897]: I1121 14:38:32.235583 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-594cb89c79-s76zb"] Nov 21 14:38:32 crc kubenswrapper[4897]: I1121 14:38:32.284884 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/e7f013b4-7b5d-4568-afc5-6457ca1fb490-openstack-edpm-ipam\") pod \"dnsmasq-dns-594cb89c79-s76zb\" (UID: \"e7f013b4-7b5d-4568-afc5-6457ca1fb490\") " pod="openstack/dnsmasq-dns-594cb89c79-s76zb" Nov 21 14:38:32 crc kubenswrapper[4897]: I1121 14:38:32.284936 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e7f013b4-7b5d-4568-afc5-6457ca1fb490-ovsdbserver-sb\") pod \"dnsmasq-dns-594cb89c79-s76zb\" (UID: \"e7f013b4-7b5d-4568-afc5-6457ca1fb490\") " pod="openstack/dnsmasq-dns-594cb89c79-s76zb" Nov 21 14:38:32 crc kubenswrapper[4897]: I1121 14:38:32.285062 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e7f013b4-7b5d-4568-afc5-6457ca1fb490-ovsdbserver-nb\") pod \"dnsmasq-dns-594cb89c79-s76zb\" (UID: \"e7f013b4-7b5d-4568-afc5-6457ca1fb490\") " pod="openstack/dnsmasq-dns-594cb89c79-s76zb" Nov 21 14:38:32 crc kubenswrapper[4897]: I1121 14:38:32.285104 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g8txz\" (UniqueName: \"kubernetes.io/projected/e7f013b4-7b5d-4568-afc5-6457ca1fb490-kube-api-access-g8txz\") pod \"dnsmasq-dns-594cb89c79-s76zb\" (UID: \"e7f013b4-7b5d-4568-afc5-6457ca1fb490\") " pod="openstack/dnsmasq-dns-594cb89c79-s76zb" Nov 21 14:38:32 crc kubenswrapper[4897]: I1121 14:38:32.285126 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7f013b4-7b5d-4568-afc5-6457ca1fb490-config\") pod \"dnsmasq-dns-594cb89c79-s76zb\" (UID: \"e7f013b4-7b5d-4568-afc5-6457ca1fb490\") " pod="openstack/dnsmasq-dns-594cb89c79-s76zb" Nov 21 14:38:32 crc kubenswrapper[4897]: I1121 14:38:32.285169 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e7f013b4-7b5d-4568-afc5-6457ca1fb490-dns-swift-storage-0\") pod \"dnsmasq-dns-594cb89c79-s76zb\" (UID: \"e7f013b4-7b5d-4568-afc5-6457ca1fb490\") " pod="openstack/dnsmasq-dns-594cb89c79-s76zb" Nov 21 14:38:32 crc kubenswrapper[4897]: I1121 14:38:32.285216 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e7f013b4-7b5d-4568-afc5-6457ca1fb490-dns-svc\") pod \"dnsmasq-dns-594cb89c79-s76zb\" (UID: \"e7f013b4-7b5d-4568-afc5-6457ca1fb490\") " pod="openstack/dnsmasq-dns-594cb89c79-s76zb" Nov 21 14:38:32 crc kubenswrapper[4897]: I1121 14:38:32.386748 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e7f013b4-7b5d-4568-afc5-6457ca1fb490-ovsdbserver-nb\") pod \"dnsmasq-dns-594cb89c79-s76zb\" (UID: \"e7f013b4-7b5d-4568-afc5-6457ca1fb490\") " pod="openstack/dnsmasq-dns-594cb89c79-s76zb" Nov 21 14:38:32 crc kubenswrapper[4897]: I1121 14:38:32.386832 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g8txz\" (UniqueName: \"kubernetes.io/projected/e7f013b4-7b5d-4568-afc5-6457ca1fb490-kube-api-access-g8txz\") pod \"dnsmasq-dns-594cb89c79-s76zb\" (UID: \"e7f013b4-7b5d-4568-afc5-6457ca1fb490\") " pod="openstack/dnsmasq-dns-594cb89c79-s76zb" Nov 21 14:38:32 crc kubenswrapper[4897]: I1121 14:38:32.386858 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7f013b4-7b5d-4568-afc5-6457ca1fb490-config\") pod \"dnsmasq-dns-594cb89c79-s76zb\" (UID: \"e7f013b4-7b5d-4568-afc5-6457ca1fb490\") " pod="openstack/dnsmasq-dns-594cb89c79-s76zb" Nov 21 14:38:32 crc kubenswrapper[4897]: I1121 14:38:32.386902 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e7f013b4-7b5d-4568-afc5-6457ca1fb490-dns-swift-storage-0\") pod \"dnsmasq-dns-594cb89c79-s76zb\" (UID: \"e7f013b4-7b5d-4568-afc5-6457ca1fb490\") " pod="openstack/dnsmasq-dns-594cb89c79-s76zb" Nov 21 14:38:32 crc kubenswrapper[4897]: I1121 14:38:32.386957 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e7f013b4-7b5d-4568-afc5-6457ca1fb490-dns-svc\") pod \"dnsmasq-dns-594cb89c79-s76zb\" (UID: \"e7f013b4-7b5d-4568-afc5-6457ca1fb490\") " pod="openstack/dnsmasq-dns-594cb89c79-s76zb" Nov 21 14:38:32 crc kubenswrapper[4897]: I1121 14:38:32.386984 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/e7f013b4-7b5d-4568-afc5-6457ca1fb490-openstack-edpm-ipam\") pod \"dnsmasq-dns-594cb89c79-s76zb\" (UID: \"e7f013b4-7b5d-4568-afc5-6457ca1fb490\") " pod="openstack/dnsmasq-dns-594cb89c79-s76zb" Nov 21 14:38:32 crc kubenswrapper[4897]: I1121 14:38:32.387006 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e7f013b4-7b5d-4568-afc5-6457ca1fb490-ovsdbserver-sb\") pod \"dnsmasq-dns-594cb89c79-s76zb\" (UID: \"e7f013b4-7b5d-4568-afc5-6457ca1fb490\") " pod="openstack/dnsmasq-dns-594cb89c79-s76zb" Nov 21 14:38:32 crc kubenswrapper[4897]: I1121 14:38:32.387711 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e7f013b4-7b5d-4568-afc5-6457ca1fb490-ovsdbserver-nb\") pod \"dnsmasq-dns-594cb89c79-s76zb\" (UID: \"e7f013b4-7b5d-4568-afc5-6457ca1fb490\") " pod="openstack/dnsmasq-dns-594cb89c79-s76zb" Nov 21 14:38:32 crc kubenswrapper[4897]: I1121 14:38:32.387898 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7f013b4-7b5d-4568-afc5-6457ca1fb490-config\") pod \"dnsmasq-dns-594cb89c79-s76zb\" (UID: \"e7f013b4-7b5d-4568-afc5-6457ca1fb490\") " pod="openstack/dnsmasq-dns-594cb89c79-s76zb" Nov 21 14:38:32 crc kubenswrapper[4897]: I1121 14:38:32.387999 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/e7f013b4-7b5d-4568-afc5-6457ca1fb490-openstack-edpm-ipam\") pod \"dnsmasq-dns-594cb89c79-s76zb\" (UID: \"e7f013b4-7b5d-4568-afc5-6457ca1fb490\") " pod="openstack/dnsmasq-dns-594cb89c79-s76zb" Nov 21 14:38:32 crc kubenswrapper[4897]: I1121 14:38:32.388068 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e7f013b4-7b5d-4568-afc5-6457ca1fb490-dns-svc\") pod \"dnsmasq-dns-594cb89c79-s76zb\" (UID: \"e7f013b4-7b5d-4568-afc5-6457ca1fb490\") " pod="openstack/dnsmasq-dns-594cb89c79-s76zb" Nov 21 14:38:32 crc kubenswrapper[4897]: I1121 14:38:32.388118 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e7f013b4-7b5d-4568-afc5-6457ca1fb490-ovsdbserver-sb\") pod \"dnsmasq-dns-594cb89c79-s76zb\" (UID: \"e7f013b4-7b5d-4568-afc5-6457ca1fb490\") " pod="openstack/dnsmasq-dns-594cb89c79-s76zb" Nov 21 14:38:32 crc kubenswrapper[4897]: I1121 14:38:32.388227 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e7f013b4-7b5d-4568-afc5-6457ca1fb490-dns-swift-storage-0\") pod \"dnsmasq-dns-594cb89c79-s76zb\" (UID: \"e7f013b4-7b5d-4568-afc5-6457ca1fb490\") " pod="openstack/dnsmasq-dns-594cb89c79-s76zb" Nov 21 14:38:32 crc kubenswrapper[4897]: I1121 14:38:32.406943 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g8txz\" (UniqueName: \"kubernetes.io/projected/e7f013b4-7b5d-4568-afc5-6457ca1fb490-kube-api-access-g8txz\") pod \"dnsmasq-dns-594cb89c79-s76zb\" (UID: \"e7f013b4-7b5d-4568-afc5-6457ca1fb490\") " pod="openstack/dnsmasq-dns-594cb89c79-s76zb" Nov 21 14:38:32 crc kubenswrapper[4897]: I1121 14:38:32.512463 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-594cb89c79-s76zb" Nov 21 14:38:33 crc kubenswrapper[4897]: I1121 14:38:33.167988 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-594cb89c79-s76zb"] Nov 21 14:38:34 crc kubenswrapper[4897]: I1121 14:38:34.053053 4897 generic.go:334] "Generic (PLEG): container finished" podID="c4c479c7-7728-41fc-bf98-ce2dcf800840" containerID="c43eba824c1aabd68653c0fb2ebca80d3784c2763791bff3f6d120f3daaefcb6" exitCode=0 Nov 21 14:38:34 crc kubenswrapper[4897]: I1121 14:38:34.053423 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-m55sf" event={"ID":"c4c479c7-7728-41fc-bf98-ce2dcf800840","Type":"ContainerDied","Data":"c43eba824c1aabd68653c0fb2ebca80d3784c2763791bff3f6d120f3daaefcb6"} Nov 21 14:38:34 crc kubenswrapper[4897]: I1121 14:38:34.055896 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-594cb89c79-s76zb" event={"ID":"e7f013b4-7b5d-4568-afc5-6457ca1fb490","Type":"ContainerStarted","Data":"c0b5dc6d4353c0d327f7b2bac14edfe6e5360f74e9566b3874cb5b254d7ddb22"} Nov 21 14:38:35 crc kubenswrapper[4897]: I1121 14:38:35.066801 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-594cb89c79-s76zb" event={"ID":"e7f013b4-7b5d-4568-afc5-6457ca1fb490","Type":"ContainerStarted","Data":"832c6b1221c55e8116033ec5bcdab96f3fa8fea0dcd39368533587175f42ab61"} Nov 21 14:38:35 crc kubenswrapper[4897]: I1121 14:38:35.500912 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-m55sf" Nov 21 14:38:35 crc kubenswrapper[4897]: I1121 14:38:35.659993 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-th9pr\" (UniqueName: \"kubernetes.io/projected/c4c479c7-7728-41fc-bf98-ce2dcf800840-kube-api-access-th9pr\") pod \"c4c479c7-7728-41fc-bf98-ce2dcf800840\" (UID: \"c4c479c7-7728-41fc-bf98-ce2dcf800840\") " Nov 21 14:38:35 crc kubenswrapper[4897]: I1121 14:38:35.660584 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4c479c7-7728-41fc-bf98-ce2dcf800840-combined-ca-bundle\") pod \"c4c479c7-7728-41fc-bf98-ce2dcf800840\" (UID: \"c4c479c7-7728-41fc-bf98-ce2dcf800840\") " Nov 21 14:38:35 crc kubenswrapper[4897]: I1121 14:38:35.660658 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c4c479c7-7728-41fc-bf98-ce2dcf800840-config-data\") pod \"c4c479c7-7728-41fc-bf98-ce2dcf800840\" (UID: \"c4c479c7-7728-41fc-bf98-ce2dcf800840\") " Nov 21 14:38:35 crc kubenswrapper[4897]: I1121 14:38:35.681680 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c4c479c7-7728-41fc-bf98-ce2dcf800840-kube-api-access-th9pr" (OuterVolumeSpecName: "kube-api-access-th9pr") pod "c4c479c7-7728-41fc-bf98-ce2dcf800840" (UID: "c4c479c7-7728-41fc-bf98-ce2dcf800840"). InnerVolumeSpecName "kube-api-access-th9pr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:38:35 crc kubenswrapper[4897]: I1121 14:38:35.764223 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-th9pr\" (UniqueName: \"kubernetes.io/projected/c4c479c7-7728-41fc-bf98-ce2dcf800840-kube-api-access-th9pr\") on node \"crc\" DevicePath \"\"" Nov 21 14:38:35 crc kubenswrapper[4897]: I1121 14:38:35.769742 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4c479c7-7728-41fc-bf98-ce2dcf800840-config-data" (OuterVolumeSpecName: "config-data") pod "c4c479c7-7728-41fc-bf98-ce2dcf800840" (UID: "c4c479c7-7728-41fc-bf98-ce2dcf800840"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:38:35 crc kubenswrapper[4897]: I1121 14:38:35.867176 4897 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c4c479c7-7728-41fc-bf98-ce2dcf800840-config-data\") on node \"crc\" DevicePath \"\"" Nov 21 14:38:35 crc kubenswrapper[4897]: I1121 14:38:35.950893 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4c479c7-7728-41fc-bf98-ce2dcf800840-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c4c479c7-7728-41fc-bf98-ce2dcf800840" (UID: "c4c479c7-7728-41fc-bf98-ce2dcf800840"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:38:35 crc kubenswrapper[4897]: I1121 14:38:35.969656 4897 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4c479c7-7728-41fc-bf98-ce2dcf800840-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 21 14:38:36 crc kubenswrapper[4897]: I1121 14:38:36.077880 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-m55sf" event={"ID":"c4c479c7-7728-41fc-bf98-ce2dcf800840","Type":"ContainerDied","Data":"68dfec735f2c3af1ffe088b4d082a02a473347f95c141f1c45d36b2a00fb59ce"} Nov 21 14:38:36 crc kubenswrapper[4897]: I1121 14:38:36.077922 4897 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="68dfec735f2c3af1ffe088b4d082a02a473347f95c141f1c45d36b2a00fb59ce" Nov 21 14:38:36 crc kubenswrapper[4897]: I1121 14:38:36.077933 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-m55sf" Nov 21 14:38:36 crc kubenswrapper[4897]: I1121 14:38:36.080853 4897 generic.go:334] "Generic (PLEG): container finished" podID="e7f013b4-7b5d-4568-afc5-6457ca1fb490" containerID="832c6b1221c55e8116033ec5bcdab96f3fa8fea0dcd39368533587175f42ab61" exitCode=0 Nov 21 14:38:36 crc kubenswrapper[4897]: I1121 14:38:36.080880 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-594cb89c79-s76zb" event={"ID":"e7f013b4-7b5d-4568-afc5-6457ca1fb490","Type":"ContainerDied","Data":"832c6b1221c55e8116033ec5bcdab96f3fa8fea0dcd39368533587175f42ab61"} Nov 21 14:38:36 crc kubenswrapper[4897]: I1121 14:38:36.974867 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-7b87584448-bxhzq"] Nov 21 14:38:36 crc kubenswrapper[4897]: E1121 14:38:36.975846 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4c479c7-7728-41fc-bf98-ce2dcf800840" containerName="heat-db-sync" Nov 21 14:38:36 crc kubenswrapper[4897]: I1121 14:38:36.975896 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4c479c7-7728-41fc-bf98-ce2dcf800840" containerName="heat-db-sync" Nov 21 14:38:36 crc kubenswrapper[4897]: I1121 14:38:36.976247 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4c479c7-7728-41fc-bf98-ce2dcf800840" containerName="heat-db-sync" Nov 21 14:38:36 crc kubenswrapper[4897]: I1121 14:38:36.977360 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-7b87584448-bxhzq" Nov 21 14:38:36 crc kubenswrapper[4897]: I1121 14:38:36.998022 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-7b87584448-bxhzq"] Nov 21 14:38:37 crc kubenswrapper[4897]: I1121 14:38:37.078010 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-5b7df7f864-2x8r2"] Nov 21 14:38:37 crc kubenswrapper[4897]: I1121 14:38:37.079695 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-5b7df7f864-2x8r2" Nov 21 14:38:37 crc kubenswrapper[4897]: I1121 14:38:37.090865 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-7b74b8db5b-7lvl5"] Nov 21 14:38:37 crc kubenswrapper[4897]: I1121 14:38:37.093884 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-7b74b8db5b-7lvl5" Nov 21 14:38:37 crc kubenswrapper[4897]: I1121 14:38:37.099858 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5b65730-8328-4c6f-b661-b27f46aa1ef7-combined-ca-bundle\") pod \"heat-engine-7b87584448-bxhzq\" (UID: \"e5b65730-8328-4c6f-b661-b27f46aa1ef7\") " pod="openstack/heat-engine-7b87584448-bxhzq" Nov 21 14:38:37 crc kubenswrapper[4897]: I1121 14:38:37.099917 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sccrn\" (UniqueName: \"kubernetes.io/projected/e5b65730-8328-4c6f-b661-b27f46aa1ef7-kube-api-access-sccrn\") pod \"heat-engine-7b87584448-bxhzq\" (UID: \"e5b65730-8328-4c6f-b661-b27f46aa1ef7\") " pod="openstack/heat-engine-7b87584448-bxhzq" Nov 21 14:38:37 crc kubenswrapper[4897]: I1121 14:38:37.100084 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5b65730-8328-4c6f-b661-b27f46aa1ef7-config-data\") pod \"heat-engine-7b87584448-bxhzq\" (UID: \"e5b65730-8328-4c6f-b661-b27f46aa1ef7\") " pod="openstack/heat-engine-7b87584448-bxhzq" Nov 21 14:38:37 crc kubenswrapper[4897]: I1121 14:38:37.100106 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e5b65730-8328-4c6f-b661-b27f46aa1ef7-config-data-custom\") pod \"heat-engine-7b87584448-bxhzq\" (UID: \"e5b65730-8328-4c6f-b661-b27f46aa1ef7\") " pod="openstack/heat-engine-7b87584448-bxhzq" Nov 21 14:38:37 crc kubenswrapper[4897]: I1121 14:38:37.106740 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-5b7df7f864-2x8r2"] Nov 21 14:38:37 crc kubenswrapper[4897]: I1121 14:38:37.128721 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-7b74b8db5b-7lvl5"] Nov 21 14:38:37 crc kubenswrapper[4897]: I1121 14:38:37.202273 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5b65730-8328-4c6f-b661-b27f46aa1ef7-combined-ca-bundle\") pod \"heat-engine-7b87584448-bxhzq\" (UID: \"e5b65730-8328-4c6f-b661-b27f46aa1ef7\") " pod="openstack/heat-engine-7b87584448-bxhzq" Nov 21 14:38:37 crc kubenswrapper[4897]: I1121 14:38:37.202328 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-htbqj\" (UniqueName: \"kubernetes.io/projected/d3271c9a-4dec-4da7-bce2-29ea6ae30200-kube-api-access-htbqj\") pod \"heat-api-7b74b8db5b-7lvl5\" (UID: \"d3271c9a-4dec-4da7-bce2-29ea6ae30200\") " pod="openstack/heat-api-7b74b8db5b-7lvl5" Nov 21 14:38:37 crc kubenswrapper[4897]: I1121 14:38:37.202350 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d3271c9a-4dec-4da7-bce2-29ea6ae30200-config-data-custom\") pod \"heat-api-7b74b8db5b-7lvl5\" (UID: \"d3271c9a-4dec-4da7-bce2-29ea6ae30200\") " pod="openstack/heat-api-7b74b8db5b-7lvl5" Nov 21 14:38:37 crc kubenswrapper[4897]: I1121 14:38:37.202374 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d86e4524-d096-48c1-b407-f77b4bd9bc5d-config-data\") pod \"heat-cfnapi-5b7df7f864-2x8r2\" (UID: \"d86e4524-d096-48c1-b407-f77b4bd9bc5d\") " pod="openstack/heat-cfnapi-5b7df7f864-2x8r2" Nov 21 14:38:37 crc kubenswrapper[4897]: I1121 14:38:37.202394 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sccrn\" (UniqueName: \"kubernetes.io/projected/e5b65730-8328-4c6f-b661-b27f46aa1ef7-kube-api-access-sccrn\") pod \"heat-engine-7b87584448-bxhzq\" (UID: \"e5b65730-8328-4c6f-b661-b27f46aa1ef7\") " pod="openstack/heat-engine-7b87584448-bxhzq" Nov 21 14:38:37 crc kubenswrapper[4897]: I1121 14:38:37.202422 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-srg2r\" (UniqueName: \"kubernetes.io/projected/d86e4524-d096-48c1-b407-f77b4bd9bc5d-kube-api-access-srg2r\") pod \"heat-cfnapi-5b7df7f864-2x8r2\" (UID: \"d86e4524-d096-48c1-b407-f77b4bd9bc5d\") " pod="openstack/heat-cfnapi-5b7df7f864-2x8r2" Nov 21 14:38:37 crc kubenswrapper[4897]: I1121 14:38:37.202466 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d86e4524-d096-48c1-b407-f77b4bd9bc5d-internal-tls-certs\") pod \"heat-cfnapi-5b7df7f864-2x8r2\" (UID: \"d86e4524-d096-48c1-b407-f77b4bd9bc5d\") " pod="openstack/heat-cfnapi-5b7df7f864-2x8r2" Nov 21 14:38:37 crc kubenswrapper[4897]: I1121 14:38:37.202527 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d3271c9a-4dec-4da7-bce2-29ea6ae30200-public-tls-certs\") pod \"heat-api-7b74b8db5b-7lvl5\" (UID: \"d3271c9a-4dec-4da7-bce2-29ea6ae30200\") " pod="openstack/heat-api-7b74b8db5b-7lvl5" Nov 21 14:38:37 crc kubenswrapper[4897]: I1121 14:38:37.202626 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d86e4524-d096-48c1-b407-f77b4bd9bc5d-public-tls-certs\") pod \"heat-cfnapi-5b7df7f864-2x8r2\" (UID: \"d86e4524-d096-48c1-b407-f77b4bd9bc5d\") " pod="openstack/heat-cfnapi-5b7df7f864-2x8r2" Nov 21 14:38:37 crc kubenswrapper[4897]: I1121 14:38:37.202722 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d3271c9a-4dec-4da7-bce2-29ea6ae30200-internal-tls-certs\") pod \"heat-api-7b74b8db5b-7lvl5\" (UID: \"d3271c9a-4dec-4da7-bce2-29ea6ae30200\") " pod="openstack/heat-api-7b74b8db5b-7lvl5" Nov 21 14:38:37 crc kubenswrapper[4897]: I1121 14:38:37.202804 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d86e4524-d096-48c1-b407-f77b4bd9bc5d-combined-ca-bundle\") pod \"heat-cfnapi-5b7df7f864-2x8r2\" (UID: \"d86e4524-d096-48c1-b407-f77b4bd9bc5d\") " pod="openstack/heat-cfnapi-5b7df7f864-2x8r2" Nov 21 14:38:37 crc kubenswrapper[4897]: I1121 14:38:37.202888 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3271c9a-4dec-4da7-bce2-29ea6ae30200-combined-ca-bundle\") pod \"heat-api-7b74b8db5b-7lvl5\" (UID: \"d3271c9a-4dec-4da7-bce2-29ea6ae30200\") " pod="openstack/heat-api-7b74b8db5b-7lvl5" Nov 21 14:38:37 crc kubenswrapper[4897]: I1121 14:38:37.202924 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5b65730-8328-4c6f-b661-b27f46aa1ef7-config-data\") pod \"heat-engine-7b87584448-bxhzq\" (UID: \"e5b65730-8328-4c6f-b661-b27f46aa1ef7\") " pod="openstack/heat-engine-7b87584448-bxhzq" Nov 21 14:38:37 crc kubenswrapper[4897]: I1121 14:38:37.202965 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e5b65730-8328-4c6f-b661-b27f46aa1ef7-config-data-custom\") pod \"heat-engine-7b87584448-bxhzq\" (UID: \"e5b65730-8328-4c6f-b661-b27f46aa1ef7\") " pod="openstack/heat-engine-7b87584448-bxhzq" Nov 21 14:38:37 crc kubenswrapper[4897]: I1121 14:38:37.202987 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3271c9a-4dec-4da7-bce2-29ea6ae30200-config-data\") pod \"heat-api-7b74b8db5b-7lvl5\" (UID: \"d3271c9a-4dec-4da7-bce2-29ea6ae30200\") " pod="openstack/heat-api-7b74b8db5b-7lvl5" Nov 21 14:38:37 crc kubenswrapper[4897]: I1121 14:38:37.203067 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d86e4524-d096-48c1-b407-f77b4bd9bc5d-config-data-custom\") pod \"heat-cfnapi-5b7df7f864-2x8r2\" (UID: \"d86e4524-d096-48c1-b407-f77b4bd9bc5d\") " pod="openstack/heat-cfnapi-5b7df7f864-2x8r2" Nov 21 14:38:37 crc kubenswrapper[4897]: I1121 14:38:37.209409 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e5b65730-8328-4c6f-b661-b27f46aa1ef7-config-data-custom\") pod \"heat-engine-7b87584448-bxhzq\" (UID: \"e5b65730-8328-4c6f-b661-b27f46aa1ef7\") " pod="openstack/heat-engine-7b87584448-bxhzq" Nov 21 14:38:37 crc kubenswrapper[4897]: I1121 14:38:37.209473 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5b65730-8328-4c6f-b661-b27f46aa1ef7-combined-ca-bundle\") pod \"heat-engine-7b87584448-bxhzq\" (UID: \"e5b65730-8328-4c6f-b661-b27f46aa1ef7\") " pod="openstack/heat-engine-7b87584448-bxhzq" Nov 21 14:38:37 crc kubenswrapper[4897]: I1121 14:38:37.210427 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5b65730-8328-4c6f-b661-b27f46aa1ef7-config-data\") pod \"heat-engine-7b87584448-bxhzq\" (UID: \"e5b65730-8328-4c6f-b661-b27f46aa1ef7\") " pod="openstack/heat-engine-7b87584448-bxhzq" Nov 21 14:38:37 crc kubenswrapper[4897]: I1121 14:38:37.225411 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sccrn\" (UniqueName: \"kubernetes.io/projected/e5b65730-8328-4c6f-b661-b27f46aa1ef7-kube-api-access-sccrn\") pod \"heat-engine-7b87584448-bxhzq\" (UID: \"e5b65730-8328-4c6f-b661-b27f46aa1ef7\") " pod="openstack/heat-engine-7b87584448-bxhzq" Nov 21 14:38:37 crc kubenswrapper[4897]: I1121 14:38:37.305465 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d3271c9a-4dec-4da7-bce2-29ea6ae30200-public-tls-certs\") pod \"heat-api-7b74b8db5b-7lvl5\" (UID: \"d3271c9a-4dec-4da7-bce2-29ea6ae30200\") " pod="openstack/heat-api-7b74b8db5b-7lvl5" Nov 21 14:38:37 crc kubenswrapper[4897]: I1121 14:38:37.305555 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d86e4524-d096-48c1-b407-f77b4bd9bc5d-public-tls-certs\") pod \"heat-cfnapi-5b7df7f864-2x8r2\" (UID: \"d86e4524-d096-48c1-b407-f77b4bd9bc5d\") " pod="openstack/heat-cfnapi-5b7df7f864-2x8r2" Nov 21 14:38:37 crc kubenswrapper[4897]: I1121 14:38:37.305585 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d3271c9a-4dec-4da7-bce2-29ea6ae30200-internal-tls-certs\") pod \"heat-api-7b74b8db5b-7lvl5\" (UID: \"d3271c9a-4dec-4da7-bce2-29ea6ae30200\") " pod="openstack/heat-api-7b74b8db5b-7lvl5" Nov 21 14:38:37 crc kubenswrapper[4897]: I1121 14:38:37.305617 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d86e4524-d096-48c1-b407-f77b4bd9bc5d-combined-ca-bundle\") pod \"heat-cfnapi-5b7df7f864-2x8r2\" (UID: \"d86e4524-d096-48c1-b407-f77b4bd9bc5d\") " pod="openstack/heat-cfnapi-5b7df7f864-2x8r2" Nov 21 14:38:37 crc kubenswrapper[4897]: I1121 14:38:37.305647 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3271c9a-4dec-4da7-bce2-29ea6ae30200-combined-ca-bundle\") pod \"heat-api-7b74b8db5b-7lvl5\" (UID: \"d3271c9a-4dec-4da7-bce2-29ea6ae30200\") " pod="openstack/heat-api-7b74b8db5b-7lvl5" Nov 21 14:38:37 crc kubenswrapper[4897]: I1121 14:38:37.305669 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3271c9a-4dec-4da7-bce2-29ea6ae30200-config-data\") pod \"heat-api-7b74b8db5b-7lvl5\" (UID: \"d3271c9a-4dec-4da7-bce2-29ea6ae30200\") " pod="openstack/heat-api-7b74b8db5b-7lvl5" Nov 21 14:38:37 crc kubenswrapper[4897]: I1121 14:38:37.305709 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d86e4524-d096-48c1-b407-f77b4bd9bc5d-config-data-custom\") pod \"heat-cfnapi-5b7df7f864-2x8r2\" (UID: \"d86e4524-d096-48c1-b407-f77b4bd9bc5d\") " pod="openstack/heat-cfnapi-5b7df7f864-2x8r2" Nov 21 14:38:37 crc kubenswrapper[4897]: I1121 14:38:37.305760 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-htbqj\" (UniqueName: \"kubernetes.io/projected/d3271c9a-4dec-4da7-bce2-29ea6ae30200-kube-api-access-htbqj\") pod \"heat-api-7b74b8db5b-7lvl5\" (UID: \"d3271c9a-4dec-4da7-bce2-29ea6ae30200\") " pod="openstack/heat-api-7b74b8db5b-7lvl5" Nov 21 14:38:37 crc kubenswrapper[4897]: I1121 14:38:37.305779 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d3271c9a-4dec-4da7-bce2-29ea6ae30200-config-data-custom\") pod \"heat-api-7b74b8db5b-7lvl5\" (UID: \"d3271c9a-4dec-4da7-bce2-29ea6ae30200\") " pod="openstack/heat-api-7b74b8db5b-7lvl5" Nov 21 14:38:37 crc kubenswrapper[4897]: I1121 14:38:37.305798 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d86e4524-d096-48c1-b407-f77b4bd9bc5d-config-data\") pod \"heat-cfnapi-5b7df7f864-2x8r2\" (UID: \"d86e4524-d096-48c1-b407-f77b4bd9bc5d\") " pod="openstack/heat-cfnapi-5b7df7f864-2x8r2" Nov 21 14:38:37 crc kubenswrapper[4897]: I1121 14:38:37.305824 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-srg2r\" (UniqueName: \"kubernetes.io/projected/d86e4524-d096-48c1-b407-f77b4bd9bc5d-kube-api-access-srg2r\") pod \"heat-cfnapi-5b7df7f864-2x8r2\" (UID: \"d86e4524-d096-48c1-b407-f77b4bd9bc5d\") " pod="openstack/heat-cfnapi-5b7df7f864-2x8r2" Nov 21 14:38:37 crc kubenswrapper[4897]: I1121 14:38:37.305864 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d86e4524-d096-48c1-b407-f77b4bd9bc5d-internal-tls-certs\") pod \"heat-cfnapi-5b7df7f864-2x8r2\" (UID: \"d86e4524-d096-48c1-b407-f77b4bd9bc5d\") " pod="openstack/heat-cfnapi-5b7df7f864-2x8r2" Nov 21 14:38:37 crc kubenswrapper[4897]: I1121 14:38:37.307006 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-7b87584448-bxhzq" Nov 21 14:38:37 crc kubenswrapper[4897]: I1121 14:38:37.310828 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d86e4524-d096-48c1-b407-f77b4bd9bc5d-config-data-custom\") pod \"heat-cfnapi-5b7df7f864-2x8r2\" (UID: \"d86e4524-d096-48c1-b407-f77b4bd9bc5d\") " pod="openstack/heat-cfnapi-5b7df7f864-2x8r2" Nov 21 14:38:37 crc kubenswrapper[4897]: I1121 14:38:37.316308 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3271c9a-4dec-4da7-bce2-29ea6ae30200-config-data\") pod \"heat-api-7b74b8db5b-7lvl5\" (UID: \"d3271c9a-4dec-4da7-bce2-29ea6ae30200\") " pod="openstack/heat-api-7b74b8db5b-7lvl5" Nov 21 14:38:37 crc kubenswrapper[4897]: I1121 14:38:37.318054 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d3271c9a-4dec-4da7-bce2-29ea6ae30200-internal-tls-certs\") pod \"heat-api-7b74b8db5b-7lvl5\" (UID: \"d3271c9a-4dec-4da7-bce2-29ea6ae30200\") " pod="openstack/heat-api-7b74b8db5b-7lvl5" Nov 21 14:38:37 crc kubenswrapper[4897]: I1121 14:38:37.318395 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d3271c9a-4dec-4da7-bce2-29ea6ae30200-config-data-custom\") pod \"heat-api-7b74b8db5b-7lvl5\" (UID: \"d3271c9a-4dec-4da7-bce2-29ea6ae30200\") " pod="openstack/heat-api-7b74b8db5b-7lvl5" Nov 21 14:38:37 crc kubenswrapper[4897]: I1121 14:38:37.318755 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3271c9a-4dec-4da7-bce2-29ea6ae30200-combined-ca-bundle\") pod \"heat-api-7b74b8db5b-7lvl5\" (UID: \"d3271c9a-4dec-4da7-bce2-29ea6ae30200\") " pod="openstack/heat-api-7b74b8db5b-7lvl5" Nov 21 14:38:37 crc kubenswrapper[4897]: I1121 14:38:37.324089 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d86e4524-d096-48c1-b407-f77b4bd9bc5d-public-tls-certs\") pod \"heat-cfnapi-5b7df7f864-2x8r2\" (UID: \"d86e4524-d096-48c1-b407-f77b4bd9bc5d\") " pod="openstack/heat-cfnapi-5b7df7f864-2x8r2" Nov 21 14:38:37 crc kubenswrapper[4897]: I1121 14:38:37.326824 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d86e4524-d096-48c1-b407-f77b4bd9bc5d-internal-tls-certs\") pod \"heat-cfnapi-5b7df7f864-2x8r2\" (UID: \"d86e4524-d096-48c1-b407-f77b4bd9bc5d\") " pod="openstack/heat-cfnapi-5b7df7f864-2x8r2" Nov 21 14:38:37 crc kubenswrapper[4897]: I1121 14:38:37.326914 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d3271c9a-4dec-4da7-bce2-29ea6ae30200-public-tls-certs\") pod \"heat-api-7b74b8db5b-7lvl5\" (UID: \"d3271c9a-4dec-4da7-bce2-29ea6ae30200\") " pod="openstack/heat-api-7b74b8db5b-7lvl5" Nov 21 14:38:37 crc kubenswrapper[4897]: I1121 14:38:37.327045 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d86e4524-d096-48c1-b407-f77b4bd9bc5d-combined-ca-bundle\") pod \"heat-cfnapi-5b7df7f864-2x8r2\" (UID: \"d86e4524-d096-48c1-b407-f77b4bd9bc5d\") " pod="openstack/heat-cfnapi-5b7df7f864-2x8r2" Nov 21 14:38:37 crc kubenswrapper[4897]: I1121 14:38:37.327142 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d86e4524-d096-48c1-b407-f77b4bd9bc5d-config-data\") pod \"heat-cfnapi-5b7df7f864-2x8r2\" (UID: \"d86e4524-d096-48c1-b407-f77b4bd9bc5d\") " pod="openstack/heat-cfnapi-5b7df7f864-2x8r2" Nov 21 14:38:37 crc kubenswrapper[4897]: I1121 14:38:37.329592 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-htbqj\" (UniqueName: \"kubernetes.io/projected/d3271c9a-4dec-4da7-bce2-29ea6ae30200-kube-api-access-htbqj\") pod \"heat-api-7b74b8db5b-7lvl5\" (UID: \"d3271c9a-4dec-4da7-bce2-29ea6ae30200\") " pod="openstack/heat-api-7b74b8db5b-7lvl5" Nov 21 14:38:37 crc kubenswrapper[4897]: I1121 14:38:37.332653 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-srg2r\" (UniqueName: \"kubernetes.io/projected/d86e4524-d096-48c1-b407-f77b4bd9bc5d-kube-api-access-srg2r\") pod \"heat-cfnapi-5b7df7f864-2x8r2\" (UID: \"d86e4524-d096-48c1-b407-f77b4bd9bc5d\") " pod="openstack/heat-cfnapi-5b7df7f864-2x8r2" Nov 21 14:38:37 crc kubenswrapper[4897]: I1121 14:38:37.402923 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-5b7df7f864-2x8r2" Nov 21 14:38:37 crc kubenswrapper[4897]: I1121 14:38:37.417892 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-7b74b8db5b-7lvl5" Nov 21 14:38:37 crc kubenswrapper[4897]: I1121 14:38:37.845228 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-7b87584448-bxhzq"] Nov 21 14:38:37 crc kubenswrapper[4897]: I1121 14:38:37.954316 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-5b7df7f864-2x8r2"] Nov 21 14:38:38 crc kubenswrapper[4897]: W1121 14:38:38.063704 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd3271c9a_4dec_4da7_bce2_29ea6ae30200.slice/crio-3afe401c929a29290a58335a4d9de8e82078e1340f3fe0b3600f5356547c9c62 WatchSource:0}: Error finding container 3afe401c929a29290a58335a4d9de8e82078e1340f3fe0b3600f5356547c9c62: Status 404 returned error can't find the container with id 3afe401c929a29290a58335a4d9de8e82078e1340f3fe0b3600f5356547c9c62 Nov 21 14:38:38 crc kubenswrapper[4897]: I1121 14:38:38.069450 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-7b74b8db5b-7lvl5"] Nov 21 14:38:38 crc kubenswrapper[4897]: I1121 14:38:38.125065 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-7b87584448-bxhzq" event={"ID":"e5b65730-8328-4c6f-b661-b27f46aa1ef7","Type":"ContainerStarted","Data":"e835f424639094461daa0c39e87a50e96d3a6dfba51d7af733fb1b571e6a56c0"} Nov 21 14:38:38 crc kubenswrapper[4897]: I1121 14:38:38.128958 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-5b7df7f864-2x8r2" event={"ID":"d86e4524-d096-48c1-b407-f77b4bd9bc5d","Type":"ContainerStarted","Data":"0e4a2c010d90264b1b065fa4a4e7efc016fec9adf9e328aa238a23d0097838c7"} Nov 21 14:38:38 crc kubenswrapper[4897]: I1121 14:38:38.131688 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-594cb89c79-s76zb" event={"ID":"e7f013b4-7b5d-4568-afc5-6457ca1fb490","Type":"ContainerStarted","Data":"3cc6da84cb754f06e3a1d3ce1b30f5051575cc4f67de1db2881ae6cc17cec4c5"} Nov 21 14:38:38 crc kubenswrapper[4897]: I1121 14:38:38.133217 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-7b74b8db5b-7lvl5" event={"ID":"d3271c9a-4dec-4da7-bce2-29ea6ae30200","Type":"ContainerStarted","Data":"3afe401c929a29290a58335a4d9de8e82078e1340f3fe0b3600f5356547c9c62"} Nov 21 14:38:39 crc kubenswrapper[4897]: I1121 14:38:39.154473 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-7b87584448-bxhzq" event={"ID":"e5b65730-8328-4c6f-b661-b27f46aa1ef7","Type":"ContainerStarted","Data":"97cb3a53522dc3f02c4efc605a56bc3cd3a803d152b3ab5742922d2c92bcef82"} Nov 21 14:38:39 crc kubenswrapper[4897]: I1121 14:38:39.154948 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-594cb89c79-s76zb" Nov 21 14:38:39 crc kubenswrapper[4897]: I1121 14:38:39.190811 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-594cb89c79-s76zb" podStartSLOduration=7.190793367 podStartE2EDuration="7.190793367s" podCreationTimestamp="2025-11-21 14:38:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:38:39.17568954 +0000 UTC m=+1796.460283035" watchObservedRunningTime="2025-11-21 14:38:39.190793367 +0000 UTC m=+1796.475386842" Nov 21 14:38:40 crc kubenswrapper[4897]: I1121 14:38:40.165329 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-7b87584448-bxhzq" Nov 21 14:38:40 crc kubenswrapper[4897]: I1121 14:38:40.195619 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-7b87584448-bxhzq" podStartSLOduration=4.19560005 podStartE2EDuration="4.19560005s" podCreationTimestamp="2025-11-21 14:38:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:38:40.184367508 +0000 UTC m=+1797.468960983" watchObservedRunningTime="2025-11-21 14:38:40.19560005 +0000 UTC m=+1797.480193525" Nov 21 14:38:42 crc kubenswrapper[4897]: I1121 14:38:42.513741 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-594cb89c79-s76zb" Nov 21 14:38:42 crc kubenswrapper[4897]: I1121 14:38:42.588730 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d99f6bc7f-k7pkd"] Nov 21 14:38:42 crc kubenswrapper[4897]: I1121 14:38:42.588955 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6d99f6bc7f-k7pkd" podUID="139262e5-8014-495f-b877-8a355481b63e" containerName="dnsmasq-dns" containerID="cri-o://68ad1c7ead606d090872be2914843b6f3b65ad129c1977e39d9cb84fe5b3e99b" gracePeriod=10 Nov 21 14:38:42 crc kubenswrapper[4897]: I1121 14:38:42.805833 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5596c69fcc-kxx8n"] Nov 21 14:38:42 crc kubenswrapper[4897]: I1121 14:38:42.807940 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5596c69fcc-kxx8n" Nov 21 14:38:42 crc kubenswrapper[4897]: I1121 14:38:42.825209 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5596c69fcc-kxx8n"] Nov 21 14:38:42 crc kubenswrapper[4897]: I1121 14:38:42.941090 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/38ba447a-68ff-4d5f-b826-4bd60240d570-openstack-edpm-ipam\") pod \"dnsmasq-dns-5596c69fcc-kxx8n\" (UID: \"38ba447a-68ff-4d5f-b826-4bd60240d570\") " pod="openstack/dnsmasq-dns-5596c69fcc-kxx8n" Nov 21 14:38:42 crc kubenswrapper[4897]: I1121 14:38:42.941142 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/38ba447a-68ff-4d5f-b826-4bd60240d570-ovsdbserver-sb\") pod \"dnsmasq-dns-5596c69fcc-kxx8n\" (UID: \"38ba447a-68ff-4d5f-b826-4bd60240d570\") " pod="openstack/dnsmasq-dns-5596c69fcc-kxx8n" Nov 21 14:38:42 crc kubenswrapper[4897]: I1121 14:38:42.941173 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/38ba447a-68ff-4d5f-b826-4bd60240d570-config\") pod \"dnsmasq-dns-5596c69fcc-kxx8n\" (UID: \"38ba447a-68ff-4d5f-b826-4bd60240d570\") " pod="openstack/dnsmasq-dns-5596c69fcc-kxx8n" Nov 21 14:38:42 crc kubenswrapper[4897]: I1121 14:38:42.941306 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rsbfl\" (UniqueName: \"kubernetes.io/projected/38ba447a-68ff-4d5f-b826-4bd60240d570-kube-api-access-rsbfl\") pod \"dnsmasq-dns-5596c69fcc-kxx8n\" (UID: \"38ba447a-68ff-4d5f-b826-4bd60240d570\") " pod="openstack/dnsmasq-dns-5596c69fcc-kxx8n" Nov 21 14:38:42 crc kubenswrapper[4897]: I1121 14:38:42.941436 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/38ba447a-68ff-4d5f-b826-4bd60240d570-dns-swift-storage-0\") pod \"dnsmasq-dns-5596c69fcc-kxx8n\" (UID: \"38ba447a-68ff-4d5f-b826-4bd60240d570\") " pod="openstack/dnsmasq-dns-5596c69fcc-kxx8n" Nov 21 14:38:42 crc kubenswrapper[4897]: I1121 14:38:42.941679 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/38ba447a-68ff-4d5f-b826-4bd60240d570-ovsdbserver-nb\") pod \"dnsmasq-dns-5596c69fcc-kxx8n\" (UID: \"38ba447a-68ff-4d5f-b826-4bd60240d570\") " pod="openstack/dnsmasq-dns-5596c69fcc-kxx8n" Nov 21 14:38:42 crc kubenswrapper[4897]: I1121 14:38:42.942007 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/38ba447a-68ff-4d5f-b826-4bd60240d570-dns-svc\") pod \"dnsmasq-dns-5596c69fcc-kxx8n\" (UID: \"38ba447a-68ff-4d5f-b826-4bd60240d570\") " pod="openstack/dnsmasq-dns-5596c69fcc-kxx8n" Nov 21 14:38:43 crc kubenswrapper[4897]: I1121 14:38:43.045061 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/38ba447a-68ff-4d5f-b826-4bd60240d570-openstack-edpm-ipam\") pod \"dnsmasq-dns-5596c69fcc-kxx8n\" (UID: \"38ba447a-68ff-4d5f-b826-4bd60240d570\") " pod="openstack/dnsmasq-dns-5596c69fcc-kxx8n" Nov 21 14:38:43 crc kubenswrapper[4897]: I1121 14:38:43.045097 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/38ba447a-68ff-4d5f-b826-4bd60240d570-ovsdbserver-sb\") pod \"dnsmasq-dns-5596c69fcc-kxx8n\" (UID: \"38ba447a-68ff-4d5f-b826-4bd60240d570\") " pod="openstack/dnsmasq-dns-5596c69fcc-kxx8n" Nov 21 14:38:43 crc kubenswrapper[4897]: I1121 14:38:43.045125 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/38ba447a-68ff-4d5f-b826-4bd60240d570-config\") pod \"dnsmasq-dns-5596c69fcc-kxx8n\" (UID: \"38ba447a-68ff-4d5f-b826-4bd60240d570\") " pod="openstack/dnsmasq-dns-5596c69fcc-kxx8n" Nov 21 14:38:43 crc kubenswrapper[4897]: I1121 14:38:43.045165 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rsbfl\" (UniqueName: \"kubernetes.io/projected/38ba447a-68ff-4d5f-b826-4bd60240d570-kube-api-access-rsbfl\") pod \"dnsmasq-dns-5596c69fcc-kxx8n\" (UID: \"38ba447a-68ff-4d5f-b826-4bd60240d570\") " pod="openstack/dnsmasq-dns-5596c69fcc-kxx8n" Nov 21 14:38:43 crc kubenswrapper[4897]: I1121 14:38:43.045219 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/38ba447a-68ff-4d5f-b826-4bd60240d570-dns-swift-storage-0\") pod \"dnsmasq-dns-5596c69fcc-kxx8n\" (UID: \"38ba447a-68ff-4d5f-b826-4bd60240d570\") " pod="openstack/dnsmasq-dns-5596c69fcc-kxx8n" Nov 21 14:38:43 crc kubenswrapper[4897]: I1121 14:38:43.045291 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/38ba447a-68ff-4d5f-b826-4bd60240d570-ovsdbserver-nb\") pod \"dnsmasq-dns-5596c69fcc-kxx8n\" (UID: \"38ba447a-68ff-4d5f-b826-4bd60240d570\") " pod="openstack/dnsmasq-dns-5596c69fcc-kxx8n" Nov 21 14:38:43 crc kubenswrapper[4897]: I1121 14:38:43.045365 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/38ba447a-68ff-4d5f-b826-4bd60240d570-dns-svc\") pod \"dnsmasq-dns-5596c69fcc-kxx8n\" (UID: \"38ba447a-68ff-4d5f-b826-4bd60240d570\") " pod="openstack/dnsmasq-dns-5596c69fcc-kxx8n" Nov 21 14:38:43 crc kubenswrapper[4897]: I1121 14:38:43.045995 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/38ba447a-68ff-4d5f-b826-4bd60240d570-openstack-edpm-ipam\") pod \"dnsmasq-dns-5596c69fcc-kxx8n\" (UID: \"38ba447a-68ff-4d5f-b826-4bd60240d570\") " pod="openstack/dnsmasq-dns-5596c69fcc-kxx8n" Nov 21 14:38:43 crc kubenswrapper[4897]: I1121 14:38:43.050402 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/38ba447a-68ff-4d5f-b826-4bd60240d570-ovsdbserver-sb\") pod \"dnsmasq-dns-5596c69fcc-kxx8n\" (UID: \"38ba447a-68ff-4d5f-b826-4bd60240d570\") " pod="openstack/dnsmasq-dns-5596c69fcc-kxx8n" Nov 21 14:38:43 crc kubenswrapper[4897]: I1121 14:38:43.050906 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/38ba447a-68ff-4d5f-b826-4bd60240d570-dns-svc\") pod \"dnsmasq-dns-5596c69fcc-kxx8n\" (UID: \"38ba447a-68ff-4d5f-b826-4bd60240d570\") " pod="openstack/dnsmasq-dns-5596c69fcc-kxx8n" Nov 21 14:38:43 crc kubenswrapper[4897]: I1121 14:38:43.051285 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/38ba447a-68ff-4d5f-b826-4bd60240d570-dns-swift-storage-0\") pod \"dnsmasq-dns-5596c69fcc-kxx8n\" (UID: \"38ba447a-68ff-4d5f-b826-4bd60240d570\") " pod="openstack/dnsmasq-dns-5596c69fcc-kxx8n" Nov 21 14:38:43 crc kubenswrapper[4897]: I1121 14:38:43.051395 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/38ba447a-68ff-4d5f-b826-4bd60240d570-ovsdbserver-nb\") pod \"dnsmasq-dns-5596c69fcc-kxx8n\" (UID: \"38ba447a-68ff-4d5f-b826-4bd60240d570\") " pod="openstack/dnsmasq-dns-5596c69fcc-kxx8n" Nov 21 14:38:43 crc kubenswrapper[4897]: I1121 14:38:43.052000 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/38ba447a-68ff-4d5f-b826-4bd60240d570-config\") pod \"dnsmasq-dns-5596c69fcc-kxx8n\" (UID: \"38ba447a-68ff-4d5f-b826-4bd60240d570\") " pod="openstack/dnsmasq-dns-5596c69fcc-kxx8n" Nov 21 14:38:43 crc kubenswrapper[4897]: I1121 14:38:43.063014 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rsbfl\" (UniqueName: \"kubernetes.io/projected/38ba447a-68ff-4d5f-b826-4bd60240d570-kube-api-access-rsbfl\") pod \"dnsmasq-dns-5596c69fcc-kxx8n\" (UID: \"38ba447a-68ff-4d5f-b826-4bd60240d570\") " pod="openstack/dnsmasq-dns-5596c69fcc-kxx8n" Nov 21 14:38:43 crc kubenswrapper[4897]: I1121 14:38:43.089986 4897 scope.go:117] "RemoveContainer" containerID="c68f1b003d12b4a2964169a85a04e3f0272d86c891f52e6bebd52f3781520d1a" Nov 21 14:38:43 crc kubenswrapper[4897]: E1121 14:38:43.092261 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 14:38:43 crc kubenswrapper[4897]: I1121 14:38:43.138933 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5596c69fcc-kxx8n" Nov 21 14:38:43 crc kubenswrapper[4897]: I1121 14:38:43.196492 4897 generic.go:334] "Generic (PLEG): container finished" podID="9fb2dc84-c75b-45e9-9873-081851fd8ea8" containerID="fed7b5464d8f589ebaee0f3363e246a6c9c10b313f04b7a18a572e4c4cbb0643" exitCode=0 Nov 21 14:38:43 crc kubenswrapper[4897]: I1121 14:38:43.196564 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"9fb2dc84-c75b-45e9-9873-081851fd8ea8","Type":"ContainerDied","Data":"fed7b5464d8f589ebaee0f3363e246a6c9c10b313f04b7a18a572e4c4cbb0643"} Nov 21 14:38:43 crc kubenswrapper[4897]: I1121 14:38:43.199753 4897 generic.go:334] "Generic (PLEG): container finished" podID="7a0fd0c8-15d2-45c0-9126-bf7e9e7926e8" containerID="e1f103b3423928d585b0c7cb85bbb2e1b44e99f2bb7a58a4ad2d8b9c13d89fff" exitCode=0 Nov 21 14:38:43 crc kubenswrapper[4897]: I1121 14:38:43.199787 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"7a0fd0c8-15d2-45c0-9126-bf7e9e7926e8","Type":"ContainerDied","Data":"e1f103b3423928d585b0c7cb85bbb2e1b44e99f2bb7a58a4ad2d8b9c13d89fff"} Nov 21 14:38:44 crc kubenswrapper[4897]: I1121 14:38:44.221013 4897 generic.go:334] "Generic (PLEG): container finished" podID="139262e5-8014-495f-b877-8a355481b63e" containerID="68ad1c7ead606d090872be2914843b6f3b65ad129c1977e39d9cb84fe5b3e99b" exitCode=0 Nov 21 14:38:44 crc kubenswrapper[4897]: I1121 14:38:44.221469 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d99f6bc7f-k7pkd" event={"ID":"139262e5-8014-495f-b877-8a355481b63e","Type":"ContainerDied","Data":"68ad1c7ead606d090872be2914843b6f3b65ad129c1977e39d9cb84fe5b3e99b"} Nov 21 14:38:47 crc kubenswrapper[4897]: I1121 14:38:47.295983 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6d99f6bc7f-k7pkd" podUID="139262e5-8014-495f-b877-8a355481b63e" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.252:5353: connect: connection refused" Nov 21 14:38:53 crc kubenswrapper[4897]: I1121 14:38:53.143173 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d99f6bc7f-k7pkd" Nov 21 14:38:53 crc kubenswrapper[4897]: E1121 14:38:53.157682 4897 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested" Nov 21 14:38:53 crc kubenswrapper[4897]: E1121 14:38:53.157736 4897 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested" Nov 21 14:38:53 crc kubenswrapper[4897]: E1121 14:38:53.157862 4897 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n5bhcfh55fh65dh666h65fh66chc7hddh6fhddh7ch665hb6hcch584hfch698h5c5h54fh565h669h68ch5h96h57fh6dh5c9hfdh5fbh5fch59cq,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-k2qzq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(c628be92-b88f-4e02-a981-4b8fe6fd7c13): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 21 14:38:53 crc kubenswrapper[4897]: E1121 14:38:53.159799 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ceilometer-0" podUID="c628be92-b88f-4e02-a981-4b8fe6fd7c13" Nov 21 14:38:53 crc kubenswrapper[4897]: I1121 14:38:53.230604 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/139262e5-8014-495f-b877-8a355481b63e-ovsdbserver-sb\") pod \"139262e5-8014-495f-b877-8a355481b63e\" (UID: \"139262e5-8014-495f-b877-8a355481b63e\") " Nov 21 14:38:53 crc kubenswrapper[4897]: I1121 14:38:53.230836 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/139262e5-8014-495f-b877-8a355481b63e-config\") pod \"139262e5-8014-495f-b877-8a355481b63e\" (UID: \"139262e5-8014-495f-b877-8a355481b63e\") " Nov 21 14:38:53 crc kubenswrapper[4897]: I1121 14:38:53.230954 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/139262e5-8014-495f-b877-8a355481b63e-ovsdbserver-nb\") pod \"139262e5-8014-495f-b877-8a355481b63e\" (UID: \"139262e5-8014-495f-b877-8a355481b63e\") " Nov 21 14:38:53 crc kubenswrapper[4897]: I1121 14:38:53.231012 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/139262e5-8014-495f-b877-8a355481b63e-dns-swift-storage-0\") pod \"139262e5-8014-495f-b877-8a355481b63e\" (UID: \"139262e5-8014-495f-b877-8a355481b63e\") " Nov 21 14:38:53 crc kubenswrapper[4897]: I1121 14:38:53.231132 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hj92f\" (UniqueName: \"kubernetes.io/projected/139262e5-8014-495f-b877-8a355481b63e-kube-api-access-hj92f\") pod \"139262e5-8014-495f-b877-8a355481b63e\" (UID: \"139262e5-8014-495f-b877-8a355481b63e\") " Nov 21 14:38:53 crc kubenswrapper[4897]: I1121 14:38:53.232130 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/139262e5-8014-495f-b877-8a355481b63e-dns-svc\") pod \"139262e5-8014-495f-b877-8a355481b63e\" (UID: \"139262e5-8014-495f-b877-8a355481b63e\") " Nov 21 14:38:53 crc kubenswrapper[4897]: I1121 14:38:53.238732 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/139262e5-8014-495f-b877-8a355481b63e-kube-api-access-hj92f" (OuterVolumeSpecName: "kube-api-access-hj92f") pod "139262e5-8014-495f-b877-8a355481b63e" (UID: "139262e5-8014-495f-b877-8a355481b63e"). InnerVolumeSpecName "kube-api-access-hj92f". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:38:53 crc kubenswrapper[4897]: I1121 14:38:53.337495 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hj92f\" (UniqueName: \"kubernetes.io/projected/139262e5-8014-495f-b877-8a355481b63e-kube-api-access-hj92f\") on node \"crc\" DevicePath \"\"" Nov 21 14:38:53 crc kubenswrapper[4897]: I1121 14:38:53.347075 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d99f6bc7f-k7pkd" Nov 21 14:38:53 crc kubenswrapper[4897]: I1121 14:38:53.347258 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d99f6bc7f-k7pkd" event={"ID":"139262e5-8014-495f-b877-8a355481b63e","Type":"ContainerDied","Data":"85e7d1770a28110921693cfd5e03b12873fbe5da7589b8d80b2aabb15c6a8239"} Nov 21 14:38:53 crc kubenswrapper[4897]: I1121 14:38:53.347310 4897 scope.go:117] "RemoveContainer" containerID="68ad1c7ead606d090872be2914843b6f3b65ad129c1977e39d9cb84fe5b3e99b" Nov 21 14:38:53 crc kubenswrapper[4897]: E1121 14:38:53.353750 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="c628be92-b88f-4e02-a981-4b8fe6fd7c13" Nov 21 14:38:53 crc kubenswrapper[4897]: I1121 14:38:53.396223 4897 scope.go:117] "RemoveContainer" containerID="3f71d8a78981c2b1871f233e3fb130b2a5ddbb19d9a79b4a171c893b7f7a6432" Nov 21 14:38:53 crc kubenswrapper[4897]: I1121 14:38:53.451831 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5596c69fcc-kxx8n"] Nov 21 14:38:53 crc kubenswrapper[4897]: W1121 14:38:53.467753 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod38ba447a_68ff_4d5f_b826_4bd60240d570.slice/crio-a2b502a2d4b77031d84339e9f325834254061b69df8156a18bab7de4a1bb4051 WatchSource:0}: Error finding container a2b502a2d4b77031d84339e9f325834254061b69df8156a18bab7de4a1bb4051: Status 404 returned error can't find the container with id a2b502a2d4b77031d84339e9f325834254061b69df8156a18bab7de4a1bb4051 Nov 21 14:38:53 crc kubenswrapper[4897]: I1121 14:38:53.537187 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/139262e5-8014-495f-b877-8a355481b63e-config" (OuterVolumeSpecName: "config") pod "139262e5-8014-495f-b877-8a355481b63e" (UID: "139262e5-8014-495f-b877-8a355481b63e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:38:53 crc kubenswrapper[4897]: I1121 14:38:53.545735 4897 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/139262e5-8014-495f-b877-8a355481b63e-config\") on node \"crc\" DevicePath \"\"" Nov 21 14:38:53 crc kubenswrapper[4897]: I1121 14:38:53.562148 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/139262e5-8014-495f-b877-8a355481b63e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "139262e5-8014-495f-b877-8a355481b63e" (UID: "139262e5-8014-495f-b877-8a355481b63e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:38:53 crc kubenswrapper[4897]: I1121 14:38:53.563806 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/139262e5-8014-495f-b877-8a355481b63e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "139262e5-8014-495f-b877-8a355481b63e" (UID: "139262e5-8014-495f-b877-8a355481b63e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:38:53 crc kubenswrapper[4897]: I1121 14:38:53.570723 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/139262e5-8014-495f-b877-8a355481b63e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "139262e5-8014-495f-b877-8a355481b63e" (UID: "139262e5-8014-495f-b877-8a355481b63e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:38:53 crc kubenswrapper[4897]: I1121 14:38:53.590601 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/139262e5-8014-495f-b877-8a355481b63e-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "139262e5-8014-495f-b877-8a355481b63e" (UID: "139262e5-8014-495f-b877-8a355481b63e"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:38:53 crc kubenswrapper[4897]: I1121 14:38:53.651800 4897 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/139262e5-8014-495f-b877-8a355481b63e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 21 14:38:53 crc kubenswrapper[4897]: I1121 14:38:53.651840 4897 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/139262e5-8014-495f-b877-8a355481b63e-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 21 14:38:53 crc kubenswrapper[4897]: I1121 14:38:53.651855 4897 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/139262e5-8014-495f-b877-8a355481b63e-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 21 14:38:53 crc kubenswrapper[4897]: I1121 14:38:53.651866 4897 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/139262e5-8014-495f-b877-8a355481b63e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 21 14:38:53 crc kubenswrapper[4897]: I1121 14:38:53.703408 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d99f6bc7f-k7pkd"] Nov 21 14:38:53 crc kubenswrapper[4897]: I1121 14:38:53.716787 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6d99f6bc7f-k7pkd"] Nov 21 14:38:53 crc kubenswrapper[4897]: I1121 14:38:53.764218 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 21 14:38:54 crc kubenswrapper[4897]: I1121 14:38:54.115857 4897 scope.go:117] "RemoveContainer" containerID="c68f1b003d12b4a2964169a85a04e3f0272d86c891f52e6bebd52f3781520d1a" Nov 21 14:38:54 crc kubenswrapper[4897]: I1121 14:38:54.115718 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="139262e5-8014-495f-b877-8a355481b63e" path="/var/lib/kubelet/pods/139262e5-8014-495f-b877-8a355481b63e/volumes" Nov 21 14:38:54 crc kubenswrapper[4897]: E1121 14:38:54.116898 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 14:38:54 crc kubenswrapper[4897]: I1121 14:38:54.361438 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"9fb2dc84-c75b-45e9-9873-081851fd8ea8","Type":"ContainerStarted","Data":"3af6fc00ff61b2c420d11abd1f9d18f9d49f723a03d15375b79727b47279b3e4"} Nov 21 14:38:54 crc kubenswrapper[4897]: I1121 14:38:54.372318 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5596c69fcc-kxx8n" event={"ID":"38ba447a-68ff-4d5f-b826-4bd60240d570","Type":"ContainerStarted","Data":"a2b502a2d4b77031d84339e9f325834254061b69df8156a18bab7de4a1bb4051"} Nov 21 14:38:54 crc kubenswrapper[4897]: E1121 14:38:54.373555 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="c628be92-b88f-4e02-a981-4b8fe6fd7c13" Nov 21 14:38:55 crc kubenswrapper[4897]: I1121 14:38:55.382708 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"7a0fd0c8-15d2-45c0-9126-bf7e9e7926e8","Type":"ContainerStarted","Data":"7d5d502ffc03cdc0add2fea6891afffe60b38664334abfa2faab7c390389c320"} Nov 21 14:38:55 crc kubenswrapper[4897]: I1121 14:38:55.383208 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Nov 21 14:38:55 crc kubenswrapper[4897]: I1121 14:38:55.386187 4897 generic.go:334] "Generic (PLEG): container finished" podID="38ba447a-68ff-4d5f-b826-4bd60240d570" containerID="6d247c03ca0e11bf99e3ebcc5f70ff76cbf737dba1401e259c93e96e5b83838b" exitCode=0 Nov 21 14:38:55 crc kubenswrapper[4897]: I1121 14:38:55.386244 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5596c69fcc-kxx8n" event={"ID":"38ba447a-68ff-4d5f-b826-4bd60240d570","Type":"ContainerDied","Data":"6d247c03ca0e11bf99e3ebcc5f70ff76cbf737dba1401e259c93e96e5b83838b"} Nov 21 14:38:55 crc kubenswrapper[4897]: I1121 14:38:55.389003 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2dwkg" event={"ID":"6a5df1f3-8748-4324-bf70-bf50dc091710","Type":"ContainerStarted","Data":"4029a1138125f161ccbc27d75d495ce088edf070cac2356d9bffe44c829d4868"} Nov 21 14:38:55 crc kubenswrapper[4897]: I1121 14:38:55.395766 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d8nth" event={"ID":"985eca1a-139e-46ce-a2a9-3e68d76b84d0","Type":"ContainerStarted","Data":"e91c668b832ac90da6a2db93f34a74105ac8f124d7251d1e057949d2b4fc96c6"} Nov 21 14:38:55 crc kubenswrapper[4897]: I1121 14:38:55.396613 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Nov 21 14:38:55 crc kubenswrapper[4897]: I1121 14:38:55.450588 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=53.450550822 podStartE2EDuration="53.450550822s" podCreationTimestamp="2025-11-21 14:38:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:38:55.408122809 +0000 UTC m=+1812.692716294" watchObservedRunningTime="2025-11-21 14:38:55.450550822 +0000 UTC m=+1812.735144297" Nov 21 14:38:55 crc kubenswrapper[4897]: E1121 14:38:55.459269 4897 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod38ba447a_68ff_4d5f_b826_4bd60240d570.slice/crio-conmon-6d247c03ca0e11bf99e3ebcc5f70ff76cbf737dba1401e259c93e96e5b83838b.scope\": RecentStats: unable to find data in memory cache]" Nov 21 14:38:55 crc kubenswrapper[4897]: I1121 14:38:55.501273 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=53.501252947 podStartE2EDuration="53.501252947s" podCreationTimestamp="2025-11-21 14:38:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:38:55.441999942 +0000 UTC m=+1812.726593427" watchObservedRunningTime="2025-11-21 14:38:55.501252947 +0000 UTC m=+1812.785846422" Nov 21 14:38:57 crc kubenswrapper[4897]: I1121 14:38:57.296424 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6d99f6bc7f-k7pkd" podUID="139262e5-8014-495f-b877-8a355481b63e" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.252:5353: i/o timeout" Nov 21 14:38:57 crc kubenswrapper[4897]: I1121 14:38:57.352989 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-7b87584448-bxhzq" Nov 21 14:38:57 crc kubenswrapper[4897]: I1121 14:38:57.412244 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-fbfff68bc-mgc44"] Nov 21 14:38:57 crc kubenswrapper[4897]: I1121 14:38:57.412722 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-engine-fbfff68bc-mgc44" podUID="a31a2ff8-2330-4e11-90cb-83d8b22244b4" containerName="heat-engine" containerID="cri-o://a5a70bbd5dd0125b9fb42409f1a2b98e5d87380ea8ab39814d38e7fe71ee27b8" gracePeriod=60 Nov 21 14:39:01 crc kubenswrapper[4897]: I1121 14:39:01.480937 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5596c69fcc-kxx8n" event={"ID":"38ba447a-68ff-4d5f-b826-4bd60240d570","Type":"ContainerStarted","Data":"abddfca5c9642bac1329858b279ef09c26a95c006128740d0655b9cd3c4a537d"} Nov 21 14:39:01 crc kubenswrapper[4897]: I1121 14:39:01.481555 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5596c69fcc-kxx8n" Nov 21 14:39:01 crc kubenswrapper[4897]: I1121 14:39:01.507113 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5596c69fcc-kxx8n" podStartSLOduration=19.507085367 podStartE2EDuration="19.507085367s" podCreationTimestamp="2025-11-21 14:38:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:39:01.498383262 +0000 UTC m=+1818.782976757" watchObservedRunningTime="2025-11-21 14:39:01.507085367 +0000 UTC m=+1818.791678862" Nov 21 14:39:03 crc kubenswrapper[4897]: E1121 14:39:03.960879 4897 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a5a70bbd5dd0125b9fb42409f1a2b98e5d87380ea8ab39814d38e7fe71ee27b8" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Nov 21 14:39:03 crc kubenswrapper[4897]: E1121 14:39:03.963129 4897 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a5a70bbd5dd0125b9fb42409f1a2b98e5d87380ea8ab39814d38e7fe71ee27b8" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Nov 21 14:39:03 crc kubenswrapper[4897]: E1121 14:39:03.964846 4897 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a5a70bbd5dd0125b9fb42409f1a2b98e5d87380ea8ab39814d38e7fe71ee27b8" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Nov 21 14:39:03 crc kubenswrapper[4897]: E1121 14:39:03.964910 4897 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/heat-engine-fbfff68bc-mgc44" podUID="a31a2ff8-2330-4e11-90cb-83d8b22244b4" containerName="heat-engine" Nov 21 14:39:05 crc kubenswrapper[4897]: E1121 14:39:05.498743 4897 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-master-centos10/openstack-heat-api:current-tested" Nov 21 14:39:05 crc kubenswrapper[4897]: E1121 14:39:05.498901 4897 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-master-centos10/openstack-heat-api:current-tested" Nov 21 14:39:05 crc kubenswrapper[4897]: E1121 14:39:05.499082 4897 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:heat-api,Image:quay.rdoproject.org/podified-master-centos10/openstack-heat-api:current-tested,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_httpd_setup && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n9bh89h698h669h58chdchf5h5bdhc6h54fh576h56fh76h5ffh8h649hfh695hf9h655h67bh549h68chc5h557h9dh5fh9dh669hf7hdbh57cq,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:heat-api-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-custom,ReadOnly:true,MountPath:/etc/heat/heat.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/default,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:internal-tls-certs,ReadOnly:true,MountPath:/var/lib/config-data/tls/certs/internal.crt,SubPath:tls.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:internal-tls-certs,ReadOnly:true,MountPath:/var/lib/config-data/tls/private/internal.key,SubPath:tls.key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:public-tls-certs,ReadOnly:true,MountPath:/var/lib/config-data/tls/certs/public.crt,SubPath:tls.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:public-tls-certs,ReadOnly:true,MountPath:/var/lib/config-data/tls/private/public.key,SubPath:tls.key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-htbqj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthcheck,Port:{0 8004 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:10,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthcheck,Port:{0 8004 },Host:,Scheme:HTTPS,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:10,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42418,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42418,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-api-7b74b8db5b-7lvl5_openstack(d3271c9a-4dec-4da7-bce2-29ea6ae30200): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 21 14:39:05 crc kubenswrapper[4897]: E1121 14:39:05.500280 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/heat-api-7b74b8db5b-7lvl5" podUID="d3271c9a-4dec-4da7-bce2-29ea6ae30200" Nov 21 14:39:06 crc kubenswrapper[4897]: I1121 14:39:06.512877 4897 scope.go:117] "RemoveContainer" containerID="d620b5b8baa9dcb6069695a000305aaac4895ad653cdba53e611946149acab93" Nov 21 14:39:07 crc kubenswrapper[4897]: I1121 14:39:07.090307 4897 scope.go:117] "RemoveContainer" containerID="c68f1b003d12b4a2964169a85a04e3f0272d86c891f52e6bebd52f3781520d1a" Nov 21 14:39:07 crc kubenswrapper[4897]: E1121 14:39:07.090603 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 14:39:08 crc kubenswrapper[4897]: I1121 14:39:08.140750 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5596c69fcc-kxx8n" Nov 21 14:39:08 crc kubenswrapper[4897]: I1121 14:39:08.211976 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-594cb89c79-s76zb"] Nov 21 14:39:08 crc kubenswrapper[4897]: I1121 14:39:08.212223 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-594cb89c79-s76zb" podUID="e7f013b4-7b5d-4568-afc5-6457ca1fb490" containerName="dnsmasq-dns" containerID="cri-o://3cc6da84cb754f06e3a1d3ce1b30f5051575cc4f67de1db2881ae6cc17cec4c5" gracePeriod=10 Nov 21 14:39:08 crc kubenswrapper[4897]: E1121 14:39:08.401050 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-api:current-tested\\\"\"" pod="openstack/heat-api-7b74b8db5b-7lvl5" podUID="d3271c9a-4dec-4da7-bce2-29ea6ae30200" Nov 21 14:39:08 crc kubenswrapper[4897]: E1121 14:39:08.401180 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="c628be92-b88f-4e02-a981-4b8fe6fd7c13" Nov 21 14:39:09 crc kubenswrapper[4897]: I1121 14:39:09.149862 4897 scope.go:117] "RemoveContainer" containerID="e3f7f26e10be327c1a46809d507483a15eb0aca60ab79afe47a1fa3b5e10885f" Nov 21 14:39:09 crc kubenswrapper[4897]: I1121 14:39:09.450189 4897 scope.go:117] "RemoveContainer" containerID="ccda44133f50fc08043f235aaf8d793190f2ad7da677e86d25005b5384e7f032" Nov 21 14:39:09 crc kubenswrapper[4897]: I1121 14:39:09.579432 4897 generic.go:334] "Generic (PLEG): container finished" podID="e7f013b4-7b5d-4568-afc5-6457ca1fb490" containerID="3cc6da84cb754f06e3a1d3ce1b30f5051575cc4f67de1db2881ae6cc17cec4c5" exitCode=0 Nov 21 14:39:09 crc kubenswrapper[4897]: I1121 14:39:09.579518 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-594cb89c79-s76zb" event={"ID":"e7f013b4-7b5d-4568-afc5-6457ca1fb490","Type":"ContainerDied","Data":"3cc6da84cb754f06e3a1d3ce1b30f5051575cc4f67de1db2881ae6cc17cec4c5"} Nov 21 14:39:09 crc kubenswrapper[4897]: I1121 14:39:09.581757 4897 generic.go:334] "Generic (PLEG): container finished" podID="6a5df1f3-8748-4324-bf70-bf50dc091710" containerID="4029a1138125f161ccbc27d75d495ce088edf070cac2356d9bffe44c829d4868" exitCode=0 Nov 21 14:39:09 crc kubenswrapper[4897]: I1121 14:39:09.581832 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2dwkg" event={"ID":"6a5df1f3-8748-4324-bf70-bf50dc091710","Type":"ContainerDied","Data":"4029a1138125f161ccbc27d75d495ce088edf070cac2356d9bffe44c829d4868"} Nov 21 14:39:09 crc kubenswrapper[4897]: I1121 14:39:09.687947 4897 scope.go:117] "RemoveContainer" containerID="7e1b93554f3c8ad375016f453d6de46e6965cbc1cf41b512bb5c0432ea3f9e3d" Nov 21 14:39:09 crc kubenswrapper[4897]: I1121 14:39:09.810390 4897 scope.go:117] "RemoveContainer" containerID="dc5724ef2786b81a902bc7373794e6066c91ce62272f7e949534d19dee9f3fe0" Nov 21 14:39:10 crc kubenswrapper[4897]: I1121 14:39:10.274629 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-594cb89c79-s76zb" Nov 21 14:39:10 crc kubenswrapper[4897]: I1121 14:39:10.447214 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e7f013b4-7b5d-4568-afc5-6457ca1fb490-dns-swift-storage-0\") pod \"e7f013b4-7b5d-4568-afc5-6457ca1fb490\" (UID: \"e7f013b4-7b5d-4568-afc5-6457ca1fb490\") " Nov 21 14:39:10 crc kubenswrapper[4897]: I1121 14:39:10.447343 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e7f013b4-7b5d-4568-afc5-6457ca1fb490-ovsdbserver-sb\") pod \"e7f013b4-7b5d-4568-afc5-6457ca1fb490\" (UID: \"e7f013b4-7b5d-4568-afc5-6457ca1fb490\") " Nov 21 14:39:10 crc kubenswrapper[4897]: I1121 14:39:10.447389 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/e7f013b4-7b5d-4568-afc5-6457ca1fb490-openstack-edpm-ipam\") pod \"e7f013b4-7b5d-4568-afc5-6457ca1fb490\" (UID: \"e7f013b4-7b5d-4568-afc5-6457ca1fb490\") " Nov 21 14:39:10 crc kubenswrapper[4897]: I1121 14:39:10.447484 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7f013b4-7b5d-4568-afc5-6457ca1fb490-config\") pod \"e7f013b4-7b5d-4568-afc5-6457ca1fb490\" (UID: \"e7f013b4-7b5d-4568-afc5-6457ca1fb490\") " Nov 21 14:39:10 crc kubenswrapper[4897]: I1121 14:39:10.447753 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e7f013b4-7b5d-4568-afc5-6457ca1fb490-ovsdbserver-nb\") pod \"e7f013b4-7b5d-4568-afc5-6457ca1fb490\" (UID: \"e7f013b4-7b5d-4568-afc5-6457ca1fb490\") " Nov 21 14:39:10 crc kubenswrapper[4897]: I1121 14:39:10.447814 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e7f013b4-7b5d-4568-afc5-6457ca1fb490-dns-svc\") pod \"e7f013b4-7b5d-4568-afc5-6457ca1fb490\" (UID: \"e7f013b4-7b5d-4568-afc5-6457ca1fb490\") " Nov 21 14:39:10 crc kubenswrapper[4897]: I1121 14:39:10.447943 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g8txz\" (UniqueName: \"kubernetes.io/projected/e7f013b4-7b5d-4568-afc5-6457ca1fb490-kube-api-access-g8txz\") pod \"e7f013b4-7b5d-4568-afc5-6457ca1fb490\" (UID: \"e7f013b4-7b5d-4568-afc5-6457ca1fb490\") " Nov 21 14:39:10 crc kubenswrapper[4897]: I1121 14:39:10.454747 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7f013b4-7b5d-4568-afc5-6457ca1fb490-kube-api-access-g8txz" (OuterVolumeSpecName: "kube-api-access-g8txz") pod "e7f013b4-7b5d-4568-afc5-6457ca1fb490" (UID: "e7f013b4-7b5d-4568-afc5-6457ca1fb490"). InnerVolumeSpecName "kube-api-access-g8txz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:39:10 crc kubenswrapper[4897]: I1121 14:39:10.529681 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7f013b4-7b5d-4568-afc5-6457ca1fb490-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e7f013b4-7b5d-4568-afc5-6457ca1fb490" (UID: "e7f013b4-7b5d-4568-afc5-6457ca1fb490"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:39:10 crc kubenswrapper[4897]: I1121 14:39:10.541196 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7f013b4-7b5d-4568-afc5-6457ca1fb490-config" (OuterVolumeSpecName: "config") pod "e7f013b4-7b5d-4568-afc5-6457ca1fb490" (UID: "e7f013b4-7b5d-4568-afc5-6457ca1fb490"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:39:10 crc kubenswrapper[4897]: I1121 14:39:10.542523 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7f013b4-7b5d-4568-afc5-6457ca1fb490-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "e7f013b4-7b5d-4568-afc5-6457ca1fb490" (UID: "e7f013b4-7b5d-4568-afc5-6457ca1fb490"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:39:10 crc kubenswrapper[4897]: I1121 14:39:10.550851 4897 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e7f013b4-7b5d-4568-afc5-6457ca1fb490-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 21 14:39:10 crc kubenswrapper[4897]: I1121 14:39:10.551187 4897 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7f013b4-7b5d-4568-afc5-6457ca1fb490-config\") on node \"crc\" DevicePath \"\"" Nov 21 14:39:10 crc kubenswrapper[4897]: I1121 14:39:10.551814 4897 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e7f013b4-7b5d-4568-afc5-6457ca1fb490-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 21 14:39:10 crc kubenswrapper[4897]: I1121 14:39:10.551905 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g8txz\" (UniqueName: \"kubernetes.io/projected/e7f013b4-7b5d-4568-afc5-6457ca1fb490-kube-api-access-g8txz\") on node \"crc\" DevicePath \"\"" Nov 21 14:39:10 crc kubenswrapper[4897]: I1121 14:39:10.552196 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7f013b4-7b5d-4568-afc5-6457ca1fb490-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "e7f013b4-7b5d-4568-afc5-6457ca1fb490" (UID: "e7f013b4-7b5d-4568-afc5-6457ca1fb490"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:39:10 crc kubenswrapper[4897]: I1121 14:39:10.569000 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7f013b4-7b5d-4568-afc5-6457ca1fb490-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "e7f013b4-7b5d-4568-afc5-6457ca1fb490" (UID: "e7f013b4-7b5d-4568-afc5-6457ca1fb490"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:39:10 crc kubenswrapper[4897]: I1121 14:39:10.577112 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7f013b4-7b5d-4568-afc5-6457ca1fb490-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "e7f013b4-7b5d-4568-afc5-6457ca1fb490" (UID: "e7f013b4-7b5d-4568-afc5-6457ca1fb490"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:39:10 crc kubenswrapper[4897]: I1121 14:39:10.597760 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-594cb89c79-s76zb" event={"ID":"e7f013b4-7b5d-4568-afc5-6457ca1fb490","Type":"ContainerDied","Data":"c0b5dc6d4353c0d327f7b2bac14edfe6e5360f74e9566b3874cb5b254d7ddb22"} Nov 21 14:39:10 crc kubenswrapper[4897]: I1121 14:39:10.597803 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-594cb89c79-s76zb" Nov 21 14:39:10 crc kubenswrapper[4897]: I1121 14:39:10.597818 4897 scope.go:117] "RemoveContainer" containerID="3cc6da84cb754f06e3a1d3ce1b30f5051575cc4f67de1db2881ae6cc17cec4c5" Nov 21 14:39:10 crc kubenswrapper[4897]: I1121 14:39:10.637557 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-594cb89c79-s76zb"] Nov 21 14:39:10 crc kubenswrapper[4897]: I1121 14:39:10.637698 4897 scope.go:117] "RemoveContainer" containerID="832c6b1221c55e8116033ec5bcdab96f3fa8fea0dcd39368533587175f42ab61" Nov 21 14:39:10 crc kubenswrapper[4897]: I1121 14:39:10.649842 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-594cb89c79-s76zb"] Nov 21 14:39:10 crc kubenswrapper[4897]: I1121 14:39:10.653706 4897 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e7f013b4-7b5d-4568-afc5-6457ca1fb490-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 21 14:39:10 crc kubenswrapper[4897]: I1121 14:39:10.653735 4897 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/e7f013b4-7b5d-4568-afc5-6457ca1fb490-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Nov 21 14:39:10 crc kubenswrapper[4897]: I1121 14:39:10.653745 4897 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e7f013b4-7b5d-4568-afc5-6457ca1fb490-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 21 14:39:11 crc kubenswrapper[4897]: I1121 14:39:11.615517 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-5b7df7f864-2x8r2" event={"ID":"d86e4524-d096-48c1-b407-f77b4bd9bc5d","Type":"ContainerStarted","Data":"fae2fb5a087fb360feca4a4edcfe28131fb7683149351099e1ca713b48da24dc"} Nov 21 14:39:12 crc kubenswrapper[4897]: I1121 14:39:12.108650 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7f013b4-7b5d-4568-afc5-6457ca1fb490" path="/var/lib/kubelet/pods/e7f013b4-7b5d-4568-afc5-6457ca1fb490/volumes" Nov 21 14:39:13 crc kubenswrapper[4897]: I1121 14:39:13.459224 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="9fb2dc84-c75b-45e9-9873-081851fd8ea8" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.1.7:5671: connect: connection refused" Nov 21 14:39:13 crc kubenswrapper[4897]: I1121 14:39:13.664539 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-5b7df7f864-2x8r2" Nov 21 14:39:13 crc kubenswrapper[4897]: I1121 14:39:13.699312 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-5b7df7f864-2x8r2" podStartSLOduration=6.517912258 podStartE2EDuration="37.699286847s" podCreationTimestamp="2025-11-21 14:38:36 +0000 UTC" firstStartedPulling="2025-11-21 14:38:37.971225279 +0000 UTC m=+1795.255818754" lastFinishedPulling="2025-11-21 14:39:09.152599868 +0000 UTC m=+1826.437193343" observedRunningTime="2025-11-21 14:39:13.689437912 +0000 UTC m=+1830.974031407" watchObservedRunningTime="2025-11-21 14:39:13.699286847 +0000 UTC m=+1830.983880322" Nov 21 14:39:13 crc kubenswrapper[4897]: I1121 14:39:13.734639 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="7a0fd0c8-15d2-45c0-9126-bf7e9e7926e8" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.1.6:5671: connect: connection refused" Nov 21 14:39:13 crc kubenswrapper[4897]: E1121 14:39:13.957585 4897 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a5a70bbd5dd0125b9fb42409f1a2b98e5d87380ea8ab39814d38e7fe71ee27b8" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Nov 21 14:39:13 crc kubenswrapper[4897]: E1121 14:39:13.960458 4897 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a5a70bbd5dd0125b9fb42409f1a2b98e5d87380ea8ab39814d38e7fe71ee27b8" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Nov 21 14:39:13 crc kubenswrapper[4897]: E1121 14:39:13.961872 4897 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a5a70bbd5dd0125b9fb42409f1a2b98e5d87380ea8ab39814d38e7fe71ee27b8" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Nov 21 14:39:13 crc kubenswrapper[4897]: E1121 14:39:13.961905 4897 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/heat-engine-fbfff68bc-mgc44" podUID="a31a2ff8-2330-4e11-90cb-83d8b22244b4" containerName="heat-engine" Nov 21 14:39:20 crc kubenswrapper[4897]: I1121 14:39:20.736008 4897 generic.go:334] "Generic (PLEG): container finished" podID="a31a2ff8-2330-4e11-90cb-83d8b22244b4" containerID="a5a70bbd5dd0125b9fb42409f1a2b98e5d87380ea8ab39814d38e7fe71ee27b8" exitCode=0 Nov 21 14:39:20 crc kubenswrapper[4897]: I1121 14:39:20.736082 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-fbfff68bc-mgc44" event={"ID":"a31a2ff8-2330-4e11-90cb-83d8b22244b4","Type":"ContainerDied","Data":"a5a70bbd5dd0125b9fb42409f1a2b98e5d87380ea8ab39814d38e7fe71ee27b8"} Nov 21 14:39:21 crc kubenswrapper[4897]: I1121 14:39:21.089812 4897 scope.go:117] "RemoveContainer" containerID="c68f1b003d12b4a2964169a85a04e3f0272d86c891f52e6bebd52f3781520d1a" Nov 21 14:39:21 crc kubenswrapper[4897]: E1121 14:39:21.090091 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 14:39:21 crc kubenswrapper[4897]: I1121 14:39:21.757535 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2dwkg" event={"ID":"6a5df1f3-8748-4324-bf70-bf50dc091710","Type":"ContainerStarted","Data":"4b7c7032f53554bf9ea435c42eefadcb42080f89085d6513fcc89bf7f4bc2140"} Nov 21 14:39:22 crc kubenswrapper[4897]: I1121 14:39:22.050256 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-fbfff68bc-mgc44" Nov 21 14:39:22 crc kubenswrapper[4897]: I1121 14:39:22.211191 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a31a2ff8-2330-4e11-90cb-83d8b22244b4-config-data\") pod \"a31a2ff8-2330-4e11-90cb-83d8b22244b4\" (UID: \"a31a2ff8-2330-4e11-90cb-83d8b22244b4\") " Nov 21 14:39:22 crc kubenswrapper[4897]: I1121 14:39:22.211331 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a31a2ff8-2330-4e11-90cb-83d8b22244b4-config-data-custom\") pod \"a31a2ff8-2330-4e11-90cb-83d8b22244b4\" (UID: \"a31a2ff8-2330-4e11-90cb-83d8b22244b4\") " Nov 21 14:39:22 crc kubenswrapper[4897]: I1121 14:39:22.211563 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a31a2ff8-2330-4e11-90cb-83d8b22244b4-combined-ca-bundle\") pod \"a31a2ff8-2330-4e11-90cb-83d8b22244b4\" (UID: \"a31a2ff8-2330-4e11-90cb-83d8b22244b4\") " Nov 21 14:39:22 crc kubenswrapper[4897]: I1121 14:39:22.211612 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s7l7z\" (UniqueName: \"kubernetes.io/projected/a31a2ff8-2330-4e11-90cb-83d8b22244b4-kube-api-access-s7l7z\") pod \"a31a2ff8-2330-4e11-90cb-83d8b22244b4\" (UID: \"a31a2ff8-2330-4e11-90cb-83d8b22244b4\") " Nov 21 14:39:22 crc kubenswrapper[4897]: I1121 14:39:22.220384 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31a2ff8-2330-4e11-90cb-83d8b22244b4-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "a31a2ff8-2330-4e11-90cb-83d8b22244b4" (UID: "a31a2ff8-2330-4e11-90cb-83d8b22244b4"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:39:22 crc kubenswrapper[4897]: I1121 14:39:22.233226 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31a2ff8-2330-4e11-90cb-83d8b22244b4-kube-api-access-s7l7z" (OuterVolumeSpecName: "kube-api-access-s7l7z") pod "a31a2ff8-2330-4e11-90cb-83d8b22244b4" (UID: "a31a2ff8-2330-4e11-90cb-83d8b22244b4"). InnerVolumeSpecName "kube-api-access-s7l7z". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:39:22 crc kubenswrapper[4897]: I1121 14:39:22.259603 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31a2ff8-2330-4e11-90cb-83d8b22244b4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a31a2ff8-2330-4e11-90cb-83d8b22244b4" (UID: "a31a2ff8-2330-4e11-90cb-83d8b22244b4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:39:22 crc kubenswrapper[4897]: I1121 14:39:22.297172 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31a2ff8-2330-4e11-90cb-83d8b22244b4-config-data" (OuterVolumeSpecName: "config-data") pod "a31a2ff8-2330-4e11-90cb-83d8b22244b4" (UID: "a31a2ff8-2330-4e11-90cb-83d8b22244b4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:39:22 crc kubenswrapper[4897]: I1121 14:39:22.320193 4897 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a31a2ff8-2330-4e11-90cb-83d8b22244b4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 21 14:39:22 crc kubenswrapper[4897]: I1121 14:39:22.320234 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s7l7z\" (UniqueName: \"kubernetes.io/projected/a31a2ff8-2330-4e11-90cb-83d8b22244b4-kube-api-access-s7l7z\") on node \"crc\" DevicePath \"\"" Nov 21 14:39:22 crc kubenswrapper[4897]: I1121 14:39:22.320247 4897 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a31a2ff8-2330-4e11-90cb-83d8b22244b4-config-data\") on node \"crc\" DevicePath \"\"" Nov 21 14:39:22 crc kubenswrapper[4897]: I1121 14:39:22.320255 4897 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a31a2ff8-2330-4e11-90cb-83d8b22244b4-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 21 14:39:22 crc kubenswrapper[4897]: I1121 14:39:22.773854 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-fbfff68bc-mgc44" event={"ID":"a31a2ff8-2330-4e11-90cb-83d8b22244b4","Type":"ContainerDied","Data":"705df789c7edcc79398449bb8e4671d25534bb44738dd71beffd1d28562bf88a"} Nov 21 14:39:22 crc kubenswrapper[4897]: I1121 14:39:22.774303 4897 scope.go:117] "RemoveContainer" containerID="a5a70bbd5dd0125b9fb42409f1a2b98e5d87380ea8ab39814d38e7fe71ee27b8" Nov 21 14:39:22 crc kubenswrapper[4897]: I1121 14:39:22.773876 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-fbfff68bc-mgc44" Nov 21 14:39:22 crc kubenswrapper[4897]: I1121 14:39:22.813665 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-2dwkg" podStartSLOduration=10.691161484 podStartE2EDuration="58.81364395s" podCreationTimestamp="2025-11-21 14:38:24 +0000 UTC" firstStartedPulling="2025-11-21 14:38:32.019844786 +0000 UTC m=+1789.304438261" lastFinishedPulling="2025-11-21 14:39:20.142327262 +0000 UTC m=+1837.426920727" observedRunningTime="2025-11-21 14:39:22.807370912 +0000 UTC m=+1840.091964387" watchObservedRunningTime="2025-11-21 14:39:22.81364395 +0000 UTC m=+1840.098237435" Nov 21 14:39:22 crc kubenswrapper[4897]: I1121 14:39:22.835540 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-fbfff68bc-mgc44"] Nov 21 14:39:22 crc kubenswrapper[4897]: I1121 14:39:22.845227 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-engine-fbfff68bc-mgc44"] Nov 21 14:39:23 crc kubenswrapper[4897]: I1121 14:39:23.457606 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="9fb2dc84-c75b-45e9-9873-081851fd8ea8" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.1.7:5671: connect: connection refused" Nov 21 14:39:23 crc kubenswrapper[4897]: I1121 14:39:23.731097 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="7a0fd0c8-15d2-45c0-9126-bf7e9e7926e8" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.1.6:5671: connect: connection refused" Nov 21 14:39:24 crc kubenswrapper[4897]: I1121 14:39:24.144567 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31a2ff8-2330-4e11-90cb-83d8b22244b4" path="/var/lib/kubelet/pods/a31a2ff8-2330-4e11-90cb-83d8b22244b4/volumes" Nov 21 14:39:24 crc kubenswrapper[4897]: I1121 14:39:24.639477 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-2dwkg" Nov 21 14:39:24 crc kubenswrapper[4897]: I1121 14:39:24.639564 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-2dwkg" Nov 21 14:39:25 crc kubenswrapper[4897]: I1121 14:39:25.713664 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-2dwkg" podUID="6a5df1f3-8748-4324-bf70-bf50dc091710" containerName="registry-server" probeResult="failure" output=< Nov 21 14:39:25 crc kubenswrapper[4897]: timeout: failed to connect service ":50051" within 1s Nov 21 14:39:25 crc kubenswrapper[4897]: > Nov 21 14:39:27 crc kubenswrapper[4897]: I1121 14:39:27.175374 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ttw7z"] Nov 21 14:39:27 crc kubenswrapper[4897]: E1121 14:39:27.176333 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="139262e5-8014-495f-b877-8a355481b63e" containerName="init" Nov 21 14:39:27 crc kubenswrapper[4897]: I1121 14:39:27.176354 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="139262e5-8014-495f-b877-8a355481b63e" containerName="init" Nov 21 14:39:27 crc kubenswrapper[4897]: E1121 14:39:27.176382 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7f013b4-7b5d-4568-afc5-6457ca1fb490" containerName="dnsmasq-dns" Nov 21 14:39:27 crc kubenswrapper[4897]: I1121 14:39:27.176391 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7f013b4-7b5d-4568-afc5-6457ca1fb490" containerName="dnsmasq-dns" Nov 21 14:39:27 crc kubenswrapper[4897]: E1121 14:39:27.176419 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="139262e5-8014-495f-b877-8a355481b63e" containerName="dnsmasq-dns" Nov 21 14:39:27 crc kubenswrapper[4897]: I1121 14:39:27.176428 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="139262e5-8014-495f-b877-8a355481b63e" containerName="dnsmasq-dns" Nov 21 14:39:27 crc kubenswrapper[4897]: E1121 14:39:27.176463 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a31a2ff8-2330-4e11-90cb-83d8b22244b4" containerName="heat-engine" Nov 21 14:39:27 crc kubenswrapper[4897]: I1121 14:39:27.176471 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="a31a2ff8-2330-4e11-90cb-83d8b22244b4" containerName="heat-engine" Nov 21 14:39:27 crc kubenswrapper[4897]: E1121 14:39:27.176487 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7f013b4-7b5d-4568-afc5-6457ca1fb490" containerName="init" Nov 21 14:39:27 crc kubenswrapper[4897]: I1121 14:39:27.176495 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7f013b4-7b5d-4568-afc5-6457ca1fb490" containerName="init" Nov 21 14:39:27 crc kubenswrapper[4897]: I1121 14:39:27.176824 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="139262e5-8014-495f-b877-8a355481b63e" containerName="dnsmasq-dns" Nov 21 14:39:27 crc kubenswrapper[4897]: I1121 14:39:27.176847 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="e7f013b4-7b5d-4568-afc5-6457ca1fb490" containerName="dnsmasq-dns" Nov 21 14:39:27 crc kubenswrapper[4897]: I1121 14:39:27.176875 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="a31a2ff8-2330-4e11-90cb-83d8b22244b4" containerName="heat-engine" Nov 21 14:39:27 crc kubenswrapper[4897]: I1121 14:39:27.177840 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ttw7z" Nov 21 14:39:27 crc kubenswrapper[4897]: I1121 14:39:27.179652 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-dj88m" Nov 21 14:39:27 crc kubenswrapper[4897]: I1121 14:39:27.180374 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 21 14:39:27 crc kubenswrapper[4897]: I1121 14:39:27.180467 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 21 14:39:27 crc kubenswrapper[4897]: I1121 14:39:27.181043 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 21 14:39:27 crc kubenswrapper[4897]: I1121 14:39:27.264412 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jwlpq\" (UniqueName: \"kubernetes.io/projected/eea9b320-c0ba-4978-961a-a0e69803ffbf-kube-api-access-jwlpq\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ttw7z\" (UID: \"eea9b320-c0ba-4978-961a-a0e69803ffbf\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ttw7z" Nov 21 14:39:27 crc kubenswrapper[4897]: I1121 14:39:27.264853 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/eea9b320-c0ba-4978-961a-a0e69803ffbf-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ttw7z\" (UID: \"eea9b320-c0ba-4978-961a-a0e69803ffbf\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ttw7z" Nov 21 14:39:27 crc kubenswrapper[4897]: I1121 14:39:27.264880 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eea9b320-c0ba-4978-961a-a0e69803ffbf-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ttw7z\" (UID: \"eea9b320-c0ba-4978-961a-a0e69803ffbf\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ttw7z" Nov 21 14:39:27 crc kubenswrapper[4897]: I1121 14:39:27.264898 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/eea9b320-c0ba-4978-961a-a0e69803ffbf-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ttw7z\" (UID: \"eea9b320-c0ba-4978-961a-a0e69803ffbf\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ttw7z" Nov 21 14:39:27 crc kubenswrapper[4897]: I1121 14:39:27.281389 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ttw7z"] Nov 21 14:39:27 crc kubenswrapper[4897]: I1121 14:39:27.368020 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/eea9b320-c0ba-4978-961a-a0e69803ffbf-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ttw7z\" (UID: \"eea9b320-c0ba-4978-961a-a0e69803ffbf\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ttw7z" Nov 21 14:39:27 crc kubenswrapper[4897]: I1121 14:39:27.368090 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eea9b320-c0ba-4978-961a-a0e69803ffbf-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ttw7z\" (UID: \"eea9b320-c0ba-4978-961a-a0e69803ffbf\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ttw7z" Nov 21 14:39:27 crc kubenswrapper[4897]: I1121 14:39:27.368122 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/eea9b320-c0ba-4978-961a-a0e69803ffbf-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ttw7z\" (UID: \"eea9b320-c0ba-4978-961a-a0e69803ffbf\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ttw7z" Nov 21 14:39:27 crc kubenswrapper[4897]: I1121 14:39:27.369257 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jwlpq\" (UniqueName: \"kubernetes.io/projected/eea9b320-c0ba-4978-961a-a0e69803ffbf-kube-api-access-jwlpq\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ttw7z\" (UID: \"eea9b320-c0ba-4978-961a-a0e69803ffbf\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ttw7z" Nov 21 14:39:27 crc kubenswrapper[4897]: I1121 14:39:27.417779 4897 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/heat-cfnapi-5b7df7f864-2x8r2" podUID="d86e4524-d096-48c1-b407-f77b4bd9bc5d" containerName="heat-cfnapi" probeResult="failure" output="Get \"https://10.217.1.12:8000/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 21 14:39:27 crc kubenswrapper[4897]: I1121 14:39:27.418151 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-cfnapi-5b7df7f864-2x8r2" podUID="d86e4524-d096-48c1-b407-f77b4bd9bc5d" containerName="heat-cfnapi" probeResult="failure" output="Get \"https://10.217.1.12:8000/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 21 14:39:27 crc kubenswrapper[4897]: I1121 14:39:27.541223 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jwlpq\" (UniqueName: \"kubernetes.io/projected/eea9b320-c0ba-4978-961a-a0e69803ffbf-kube-api-access-jwlpq\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ttw7z\" (UID: \"eea9b320-c0ba-4978-961a-a0e69803ffbf\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ttw7z" Nov 21 14:39:27 crc kubenswrapper[4897]: I1121 14:39:27.541898 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/eea9b320-c0ba-4978-961a-a0e69803ffbf-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ttw7z\" (UID: \"eea9b320-c0ba-4978-961a-a0e69803ffbf\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ttw7z" Nov 21 14:39:27 crc kubenswrapper[4897]: I1121 14:39:27.542168 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/eea9b320-c0ba-4978-961a-a0e69803ffbf-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ttw7z\" (UID: \"eea9b320-c0ba-4978-961a-a0e69803ffbf\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ttw7z" Nov 21 14:39:27 crc kubenswrapper[4897]: I1121 14:39:27.542301 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eea9b320-c0ba-4978-961a-a0e69803ffbf-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ttw7z\" (UID: \"eea9b320-c0ba-4978-961a-a0e69803ffbf\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ttw7z" Nov 21 14:39:27 crc kubenswrapper[4897]: I1121 14:39:27.800910 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ttw7z" Nov 21 14:39:29 crc kubenswrapper[4897]: I1121 14:39:29.906894 4897 generic.go:334] "Generic (PLEG): container finished" podID="985eca1a-139e-46ce-a2a9-3e68d76b84d0" containerID="e91c668b832ac90da6a2db93f34a74105ac8f124d7251d1e057949d2b4fc96c6" exitCode=0 Nov 21 14:39:29 crc kubenswrapper[4897]: I1121 14:39:29.906942 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d8nth" event={"ID":"985eca1a-139e-46ce-a2a9-3e68d76b84d0","Type":"ContainerDied","Data":"e91c668b832ac90da6a2db93f34a74105ac8f124d7251d1e057949d2b4fc96c6"} Nov 21 14:39:33 crc kubenswrapper[4897]: I1121 14:39:33.457769 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="9fb2dc84-c75b-45e9-9873-081851fd8ea8" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.1.7:5671: connect: connection refused" Nov 21 14:39:33 crc kubenswrapper[4897]: I1121 14:39:33.730492 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="7a0fd0c8-15d2-45c0-9126-bf7e9e7926e8" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.1.6:5671: connect: connection refused" Nov 21 14:39:34 crc kubenswrapper[4897]: I1121 14:39:34.690902 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-2dwkg" Nov 21 14:39:34 crc kubenswrapper[4897]: I1121 14:39:34.740744 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-2dwkg" Nov 21 14:39:34 crc kubenswrapper[4897]: I1121 14:39:34.930075 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2dwkg"] Nov 21 14:39:35 crc kubenswrapper[4897]: I1121 14:39:35.089967 4897 scope.go:117] "RemoveContainer" containerID="c68f1b003d12b4a2964169a85a04e3f0272d86c891f52e6bebd52f3781520d1a" Nov 21 14:39:35 crc kubenswrapper[4897]: E1121 14:39:35.090317 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 14:39:35 crc kubenswrapper[4897]: I1121 14:39:35.977711 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-2dwkg" podUID="6a5df1f3-8748-4324-bf70-bf50dc091710" containerName="registry-server" containerID="cri-o://4b7c7032f53554bf9ea435c42eefadcb42080f89085d6513fcc89bf7f4bc2140" gracePeriod=2 Nov 21 14:39:37 crc kubenswrapper[4897]: I1121 14:39:37.422790 4897 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/heat-cfnapi-5b7df7f864-2x8r2" podUID="d86e4524-d096-48c1-b407-f77b4bd9bc5d" containerName="heat-cfnapi" probeResult="failure" output="Get \"https://10.217.1.12:8000/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 21 14:39:37 crc kubenswrapper[4897]: I1121 14:39:37.426920 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-cfnapi-5b7df7f864-2x8r2" podUID="d86e4524-d096-48c1-b407-f77b4bd9bc5d" containerName="heat-cfnapi" probeResult="failure" output="Get \"https://10.217.1.12:8000/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 21 14:39:38 crc kubenswrapper[4897]: I1121 14:39:38.001201 4897 generic.go:334] "Generic (PLEG): container finished" podID="6a5df1f3-8748-4324-bf70-bf50dc091710" containerID="4b7c7032f53554bf9ea435c42eefadcb42080f89085d6513fcc89bf7f4bc2140" exitCode=0 Nov 21 14:39:38 crc kubenswrapper[4897]: I1121 14:39:38.001244 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2dwkg" event={"ID":"6a5df1f3-8748-4324-bf70-bf50dc091710","Type":"ContainerDied","Data":"4b7c7032f53554bf9ea435c42eefadcb42080f89085d6513fcc89bf7f4bc2140"} Nov 21 14:39:38 crc kubenswrapper[4897]: I1121 14:39:38.003835 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c628be92-b88f-4e02-a981-4b8fe6fd7c13","Type":"ContainerStarted","Data":"1b91ca93b064103150f552df19e1b49753741160e2419d26236cb60ac048cfe4"} Nov 21 14:39:38 crc kubenswrapper[4897]: I1121 14:39:38.005468 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-7b74b8db5b-7lvl5" event={"ID":"d3271c9a-4dec-4da7-bce2-29ea6ae30200","Type":"ContainerStarted","Data":"c588624636dc8b1d325dddd96a5ad6cd85241cec35209951cf357529f3e4e0a1"} Nov 21 14:39:38 crc kubenswrapper[4897]: I1121 14:39:38.976955 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2dwkg" Nov 21 14:39:39 crc kubenswrapper[4897]: I1121 14:39:39.021810 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2dwkg" event={"ID":"6a5df1f3-8748-4324-bf70-bf50dc091710","Type":"ContainerDied","Data":"d9fa5c1a658cfbf528cddc9fd0c58c8f47e0af83e0729722328e41a07e9cd5aa"} Nov 21 14:39:39 crc kubenswrapper[4897]: I1121 14:39:39.021860 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2dwkg" Nov 21 14:39:39 crc kubenswrapper[4897]: I1121 14:39:39.021908 4897 scope.go:117] "RemoveContainer" containerID="4b7c7032f53554bf9ea435c42eefadcb42080f89085d6513fcc89bf7f4bc2140" Nov 21 14:39:39 crc kubenswrapper[4897]: I1121 14:39:39.022109 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-7b74b8db5b-7lvl5" Nov 21 14:39:39 crc kubenswrapper[4897]: I1121 14:39:39.054961 4897 scope.go:117] "RemoveContainer" containerID="4029a1138125f161ccbc27d75d495ce088edf070cac2356d9bffe44c829d4868" Nov 21 14:39:39 crc kubenswrapper[4897]: I1121 14:39:39.072872 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-7b74b8db5b-7lvl5" podStartSLOduration=5.887574544 podStartE2EDuration="1m2.072853131s" podCreationTimestamp="2025-11-21 14:38:37 +0000 UTC" firstStartedPulling="2025-11-21 14:38:39.285032055 +0000 UTC m=+1796.569625530" lastFinishedPulling="2025-11-21 14:39:35.470310642 +0000 UTC m=+1852.754904117" observedRunningTime="2025-11-21 14:39:39.044548518 +0000 UTC m=+1856.329141993" watchObservedRunningTime="2025-11-21 14:39:39.072853131 +0000 UTC m=+1856.357446606" Nov 21 14:39:39 crc kubenswrapper[4897]: I1121 14:39:39.085653 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=5.476545102 podStartE2EDuration="2m16.085626865s" podCreationTimestamp="2025-11-21 14:37:23 +0000 UTC" firstStartedPulling="2025-11-21 14:37:24.860046757 +0000 UTC m=+1722.144640232" lastFinishedPulling="2025-11-21 14:39:35.46912851 +0000 UTC m=+1852.753721995" observedRunningTime="2025-11-21 14:39:39.062942704 +0000 UTC m=+1856.347536179" watchObservedRunningTime="2025-11-21 14:39:39.085626865 +0000 UTC m=+1856.370220340" Nov 21 14:39:39 crc kubenswrapper[4897]: I1121 14:39:39.102209 4897 scope.go:117] "RemoveContainer" containerID="f765c507ff764d2639397755ebd799d092854e76e4abf34bed549d52af84ea43" Nov 21 14:39:39 crc kubenswrapper[4897]: I1121 14:39:39.151769 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a5df1f3-8748-4324-bf70-bf50dc091710-catalog-content\") pod \"6a5df1f3-8748-4324-bf70-bf50dc091710\" (UID: \"6a5df1f3-8748-4324-bf70-bf50dc091710\") " Nov 21 14:39:39 crc kubenswrapper[4897]: I1121 14:39:39.152304 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ffz2d\" (UniqueName: \"kubernetes.io/projected/6a5df1f3-8748-4324-bf70-bf50dc091710-kube-api-access-ffz2d\") pod \"6a5df1f3-8748-4324-bf70-bf50dc091710\" (UID: \"6a5df1f3-8748-4324-bf70-bf50dc091710\") " Nov 21 14:39:39 crc kubenswrapper[4897]: I1121 14:39:39.152569 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a5df1f3-8748-4324-bf70-bf50dc091710-utilities\") pod \"6a5df1f3-8748-4324-bf70-bf50dc091710\" (UID: \"6a5df1f3-8748-4324-bf70-bf50dc091710\") " Nov 21 14:39:39 crc kubenswrapper[4897]: I1121 14:39:39.157432 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6a5df1f3-8748-4324-bf70-bf50dc091710-utilities" (OuterVolumeSpecName: "utilities") pod "6a5df1f3-8748-4324-bf70-bf50dc091710" (UID: "6a5df1f3-8748-4324-bf70-bf50dc091710"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:39:39 crc kubenswrapper[4897]: I1121 14:39:39.166746 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a5df1f3-8748-4324-bf70-bf50dc091710-kube-api-access-ffz2d" (OuterVolumeSpecName: "kube-api-access-ffz2d") pod "6a5df1f3-8748-4324-bf70-bf50dc091710" (UID: "6a5df1f3-8748-4324-bf70-bf50dc091710"). InnerVolumeSpecName "kube-api-access-ffz2d". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:39:39 crc kubenswrapper[4897]: I1121 14:39:39.255829 4897 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a5df1f3-8748-4324-bf70-bf50dc091710-utilities\") on node \"crc\" DevicePath \"\"" Nov 21 14:39:39 crc kubenswrapper[4897]: I1121 14:39:39.256055 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ffz2d\" (UniqueName: \"kubernetes.io/projected/6a5df1f3-8748-4324-bf70-bf50dc091710-kube-api-access-ffz2d\") on node \"crc\" DevicePath \"\"" Nov 21 14:39:39 crc kubenswrapper[4897]: I1121 14:39:39.260024 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6a5df1f3-8748-4324-bf70-bf50dc091710-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6a5df1f3-8748-4324-bf70-bf50dc091710" (UID: "6a5df1f3-8748-4324-bf70-bf50dc091710"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:39:39 crc kubenswrapper[4897]: I1121 14:39:39.357886 4897 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a5df1f3-8748-4324-bf70-bf50dc091710-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 21 14:39:39 crc kubenswrapper[4897]: I1121 14:39:39.364223 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2dwkg"] Nov 21 14:39:39 crc kubenswrapper[4897]: I1121 14:39:39.376747 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-2dwkg"] Nov 21 14:39:40 crc kubenswrapper[4897]: I1121 14:39:40.108043 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a5df1f3-8748-4324-bf70-bf50dc091710" path="/var/lib/kubelet/pods/6a5df1f3-8748-4324-bf70-bf50dc091710/volumes" Nov 21 14:39:40 crc kubenswrapper[4897]: I1121 14:39:40.692977 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-cfnapi-5b7df7f864-2x8r2" Nov 21 14:39:40 crc kubenswrapper[4897]: I1121 14:39:40.762206 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-8459c6d79f-p9s4l"] Nov 21 14:39:40 crc kubenswrapper[4897]: I1121 14:39:40.762728 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-cfnapi-8459c6d79f-p9s4l" podUID="0bf363b7-2fe8-4676-aa3c-4ae74c2c7d05" containerName="heat-cfnapi" containerID="cri-o://8c9411e8019fc5114e4bb8a9ea116f4cf8b427fbab481635f9a469fe2705ff35" gracePeriod=60 Nov 21 14:39:43 crc kubenswrapper[4897]: I1121 14:39:43.457887 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="9fb2dc84-c75b-45e9-9873-081851fd8ea8" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.1.7:5671: connect: connection refused" Nov 21 14:39:43 crc kubenswrapper[4897]: I1121 14:39:43.730896 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="7a0fd0c8-15d2-45c0-9126-bf7e9e7926e8" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.1.6:5671: connect: connection refused" Nov 21 14:39:44 crc kubenswrapper[4897]: I1121 14:39:44.480648 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ttw7z"] Nov 21 14:39:44 crc kubenswrapper[4897]: I1121 14:39:44.587087 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-api-7b74b8db5b-7lvl5" Nov 21 14:39:44 crc kubenswrapper[4897]: I1121 14:39:44.656656 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-5d8b84cf9f-4lz8g"] Nov 21 14:39:44 crc kubenswrapper[4897]: I1121 14:39:44.657027 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-api-5d8b84cf9f-4lz8g" podUID="d7cfe6c1-f519-438c-b6f1-fea0fcc70cc3" containerName="heat-api" containerID="cri-o://762160a7ac40ed22f4f6b9f64c7579c5e61b7da769018e0e19d8fedd5c0382ff" gracePeriod=60 Nov 21 14:39:45 crc kubenswrapper[4897]: I1121 14:39:45.101789 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ttw7z" event={"ID":"eea9b320-c0ba-4978-961a-a0e69803ffbf","Type":"ContainerStarted","Data":"eb1f9642817079761f02315242af83860b2839c8bf0c50a266cc44a905ee84c4"} Nov 21 14:39:45 crc kubenswrapper[4897]: I1121 14:39:45.456126 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-cfnapi-8459c6d79f-p9s4l" podUID="0bf363b7-2fe8-4676-aa3c-4ae74c2c7d05" containerName="heat-cfnapi" probeResult="failure" output="Get \"https://10.217.0.217:8000/healthcheck\": read tcp 10.217.0.2:43102->10.217.0.217:8000: read: connection reset by peer" Nov 21 14:39:46 crc kubenswrapper[4897]: I1121 14:39:46.090041 4897 scope.go:117] "RemoveContainer" containerID="c68f1b003d12b4a2964169a85a04e3f0272d86c891f52e6bebd52f3781520d1a" Nov 21 14:39:46 crc kubenswrapper[4897]: E1121 14:39:46.090628 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 14:39:46 crc kubenswrapper[4897]: I1121 14:39:46.112744 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d8nth" event={"ID":"985eca1a-139e-46ce-a2a9-3e68d76b84d0","Type":"ContainerStarted","Data":"d687cc3a224dedde2763d0f27045e20eb89735f70f16352147e95af90a6720c2"} Nov 21 14:39:46 crc kubenswrapper[4897]: I1121 14:39:46.116167 4897 generic.go:334] "Generic (PLEG): container finished" podID="0bf363b7-2fe8-4676-aa3c-4ae74c2c7d05" containerID="8c9411e8019fc5114e4bb8a9ea116f4cf8b427fbab481635f9a469fe2705ff35" exitCode=0 Nov 21 14:39:46 crc kubenswrapper[4897]: I1121 14:39:46.116219 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-8459c6d79f-p9s4l" event={"ID":"0bf363b7-2fe8-4676-aa3c-4ae74c2c7d05","Type":"ContainerDied","Data":"8c9411e8019fc5114e4bb8a9ea116f4cf8b427fbab481635f9a469fe2705ff35"} Nov 21 14:39:46 crc kubenswrapper[4897]: I1121 14:39:46.132522 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-d8nth" podStartSLOduration=9.250302705 podStartE2EDuration="1m20.132484903s" podCreationTimestamp="2025-11-21 14:38:26 +0000 UTC" firstStartedPulling="2025-11-21 14:38:33.044039672 +0000 UTC m=+1790.328633147" lastFinishedPulling="2025-11-21 14:39:43.92622187 +0000 UTC m=+1861.210815345" observedRunningTime="2025-11-21 14:39:46.129911644 +0000 UTC m=+1863.414505139" watchObservedRunningTime="2025-11-21 14:39:46.132484903 +0000 UTC m=+1863.417078368" Nov 21 14:39:47 crc kubenswrapper[4897]: I1121 14:39:47.023714 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-d8nth" Nov 21 14:39:47 crc kubenswrapper[4897]: I1121 14:39:47.024134 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-d8nth" Nov 21 14:39:47 crc kubenswrapper[4897]: I1121 14:39:47.144935 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-8459c6d79f-p9s4l" event={"ID":"0bf363b7-2fe8-4676-aa3c-4ae74c2c7d05","Type":"ContainerDied","Data":"607ee2bb0af6ef3c38af0d5d8727a89ce71ded12cff0147f16381028331f393f"} Nov 21 14:39:47 crc kubenswrapper[4897]: I1121 14:39:47.145000 4897 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="607ee2bb0af6ef3c38af0d5d8727a89ce71ded12cff0147f16381028331f393f" Nov 21 14:39:47 crc kubenswrapper[4897]: I1121 14:39:47.236345 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-8459c6d79f-p9s4l" Nov 21 14:39:47 crc kubenswrapper[4897]: I1121 14:39:47.347817 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0bf363b7-2fe8-4676-aa3c-4ae74c2c7d05-public-tls-certs\") pod \"0bf363b7-2fe8-4676-aa3c-4ae74c2c7d05\" (UID: \"0bf363b7-2fe8-4676-aa3c-4ae74c2c7d05\") " Nov 21 14:39:47 crc kubenswrapper[4897]: I1121 14:39:47.347943 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0bf363b7-2fe8-4676-aa3c-4ae74c2c7d05-internal-tls-certs\") pod \"0bf363b7-2fe8-4676-aa3c-4ae74c2c7d05\" (UID: \"0bf363b7-2fe8-4676-aa3c-4ae74c2c7d05\") " Nov 21 14:39:47 crc kubenswrapper[4897]: I1121 14:39:47.348244 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0bf363b7-2fe8-4676-aa3c-4ae74c2c7d05-config-data-custom\") pod \"0bf363b7-2fe8-4676-aa3c-4ae74c2c7d05\" (UID: \"0bf363b7-2fe8-4676-aa3c-4ae74c2c7d05\") " Nov 21 14:39:47 crc kubenswrapper[4897]: I1121 14:39:47.348276 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4mgtw\" (UniqueName: \"kubernetes.io/projected/0bf363b7-2fe8-4676-aa3c-4ae74c2c7d05-kube-api-access-4mgtw\") pod \"0bf363b7-2fe8-4676-aa3c-4ae74c2c7d05\" (UID: \"0bf363b7-2fe8-4676-aa3c-4ae74c2c7d05\") " Nov 21 14:39:47 crc kubenswrapper[4897]: I1121 14:39:47.348301 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0bf363b7-2fe8-4676-aa3c-4ae74c2c7d05-combined-ca-bundle\") pod \"0bf363b7-2fe8-4676-aa3c-4ae74c2c7d05\" (UID: \"0bf363b7-2fe8-4676-aa3c-4ae74c2c7d05\") " Nov 21 14:39:47 crc kubenswrapper[4897]: I1121 14:39:47.348331 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0bf363b7-2fe8-4676-aa3c-4ae74c2c7d05-config-data\") pod \"0bf363b7-2fe8-4676-aa3c-4ae74c2c7d05\" (UID: \"0bf363b7-2fe8-4676-aa3c-4ae74c2c7d05\") " Nov 21 14:39:47 crc kubenswrapper[4897]: I1121 14:39:47.355777 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0bf363b7-2fe8-4676-aa3c-4ae74c2c7d05-kube-api-access-4mgtw" (OuterVolumeSpecName: "kube-api-access-4mgtw") pod "0bf363b7-2fe8-4676-aa3c-4ae74c2c7d05" (UID: "0bf363b7-2fe8-4676-aa3c-4ae74c2c7d05"). InnerVolumeSpecName "kube-api-access-4mgtw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:39:47 crc kubenswrapper[4897]: I1121 14:39:47.356237 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0bf363b7-2fe8-4676-aa3c-4ae74c2c7d05-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "0bf363b7-2fe8-4676-aa3c-4ae74c2c7d05" (UID: "0bf363b7-2fe8-4676-aa3c-4ae74c2c7d05"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:39:47 crc kubenswrapper[4897]: I1121 14:39:47.389308 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0bf363b7-2fe8-4676-aa3c-4ae74c2c7d05-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0bf363b7-2fe8-4676-aa3c-4ae74c2c7d05" (UID: "0bf363b7-2fe8-4676-aa3c-4ae74c2c7d05"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:39:47 crc kubenswrapper[4897]: I1121 14:39:47.428910 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0bf363b7-2fe8-4676-aa3c-4ae74c2c7d05-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "0bf363b7-2fe8-4676-aa3c-4ae74c2c7d05" (UID: "0bf363b7-2fe8-4676-aa3c-4ae74c2c7d05"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:39:47 crc kubenswrapper[4897]: I1121 14:39:47.452935 4897 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0bf363b7-2fe8-4676-aa3c-4ae74c2c7d05-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 21 14:39:47 crc kubenswrapper[4897]: I1121 14:39:47.452974 4897 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0bf363b7-2fe8-4676-aa3c-4ae74c2c7d05-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 21 14:39:47 crc kubenswrapper[4897]: I1121 14:39:47.452989 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4mgtw\" (UniqueName: \"kubernetes.io/projected/0bf363b7-2fe8-4676-aa3c-4ae74c2c7d05-kube-api-access-4mgtw\") on node \"crc\" DevicePath \"\"" Nov 21 14:39:47 crc kubenswrapper[4897]: I1121 14:39:47.453001 4897 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0bf363b7-2fe8-4676-aa3c-4ae74c2c7d05-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 21 14:39:47 crc kubenswrapper[4897]: I1121 14:39:47.492964 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0bf363b7-2fe8-4676-aa3c-4ae74c2c7d05-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "0bf363b7-2fe8-4676-aa3c-4ae74c2c7d05" (UID: "0bf363b7-2fe8-4676-aa3c-4ae74c2c7d05"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:39:47 crc kubenswrapper[4897]: I1121 14:39:47.493081 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0bf363b7-2fe8-4676-aa3c-4ae74c2c7d05-config-data" (OuterVolumeSpecName: "config-data") pod "0bf363b7-2fe8-4676-aa3c-4ae74c2c7d05" (UID: "0bf363b7-2fe8-4676-aa3c-4ae74c2c7d05"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:39:47 crc kubenswrapper[4897]: I1121 14:39:47.554955 4897 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0bf363b7-2fe8-4676-aa3c-4ae74c2c7d05-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 21 14:39:47 crc kubenswrapper[4897]: I1121 14:39:47.555337 4897 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0bf363b7-2fe8-4676-aa3c-4ae74c2c7d05-config-data\") on node \"crc\" DevicePath \"\"" Nov 21 14:39:47 crc kubenswrapper[4897]: I1121 14:39:47.927413 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-api-5d8b84cf9f-4lz8g" podUID="d7cfe6c1-f519-438c-b6f1-fea0fcc70cc3" containerName="heat-api" probeResult="failure" output="Get \"https://10.217.0.218:8004/healthcheck\": read tcp 10.217.0.2:38716->10.217.0.218:8004: read: connection reset by peer" Nov 21 14:39:48 crc kubenswrapper[4897]: I1121 14:39:48.085044 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-d8nth" podUID="985eca1a-139e-46ce-a2a9-3e68d76b84d0" containerName="registry-server" probeResult="failure" output=< Nov 21 14:39:48 crc kubenswrapper[4897]: timeout: failed to connect service ":50051" within 1s Nov 21 14:39:48 crc kubenswrapper[4897]: > Nov 21 14:39:48 crc kubenswrapper[4897]: I1121 14:39:48.160672 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-8459c6d79f-p9s4l" Nov 21 14:39:48 crc kubenswrapper[4897]: I1121 14:39:48.197231 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-8459c6d79f-p9s4l"] Nov 21 14:39:48 crc kubenswrapper[4897]: I1121 14:39:48.214882 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-cfnapi-8459c6d79f-p9s4l"] Nov 21 14:39:49 crc kubenswrapper[4897]: I1121 14:39:49.177643 4897 generic.go:334] "Generic (PLEG): container finished" podID="d7cfe6c1-f519-438c-b6f1-fea0fcc70cc3" containerID="762160a7ac40ed22f4f6b9f64c7579c5e61b7da769018e0e19d8fedd5c0382ff" exitCode=0 Nov 21 14:39:49 crc kubenswrapper[4897]: I1121 14:39:49.177716 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-5d8b84cf9f-4lz8g" event={"ID":"d7cfe6c1-f519-438c-b6f1-fea0fcc70cc3","Type":"ContainerDied","Data":"762160a7ac40ed22f4f6b9f64c7579c5e61b7da769018e0e19d8fedd5c0382ff"} Nov 21 14:39:49 crc kubenswrapper[4897]: I1121 14:39:49.659741 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/redhat-marketplace-jxqbm" podUID="0fcfc1e5-8927-4b4b-bd19-234497f86f23" containerName="registry-server" probeResult="failure" output=< Nov 21 14:39:49 crc kubenswrapper[4897]: timeout: failed to connect service ":50051" within 1s Nov 21 14:39:49 crc kubenswrapper[4897]: > Nov 21 14:39:49 crc kubenswrapper[4897]: I1121 14:39:49.674436 4897 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/redhat-marketplace-jxqbm" podUID="0fcfc1e5-8927-4b4b-bd19-234497f86f23" containerName="registry-server" probeResult="failure" output=< Nov 21 14:39:49 crc kubenswrapper[4897]: timeout: failed to connect service ":50051" within 1s Nov 21 14:39:49 crc kubenswrapper[4897]: > Nov 21 14:39:50 crc kubenswrapper[4897]: I1121 14:39:50.114146 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0bf363b7-2fe8-4676-aa3c-4ae74c2c7d05" path="/var/lib/kubelet/pods/0bf363b7-2fe8-4676-aa3c-4ae74c2c7d05/volumes" Nov 21 14:39:50 crc kubenswrapper[4897]: I1121 14:39:50.265583 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-5d8b84cf9f-4lz8g" Nov 21 14:39:50 crc kubenswrapper[4897]: I1121 14:39:50.342195 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7cfe6c1-f519-438c-b6f1-fea0fcc70cc3-combined-ca-bundle\") pod \"d7cfe6c1-f519-438c-b6f1-fea0fcc70cc3\" (UID: \"d7cfe6c1-f519-438c-b6f1-fea0fcc70cc3\") " Nov 21 14:39:50 crc kubenswrapper[4897]: I1121 14:39:50.374497 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7cfe6c1-f519-438c-b6f1-fea0fcc70cc3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d7cfe6c1-f519-438c-b6f1-fea0fcc70cc3" (UID: "d7cfe6c1-f519-438c-b6f1-fea0fcc70cc3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:39:50 crc kubenswrapper[4897]: I1121 14:39:50.402843 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-5d8b84cf9f-4lz8g" event={"ID":"d7cfe6c1-f519-438c-b6f1-fea0fcc70cc3","Type":"ContainerDied","Data":"995365c351d53d709c10f9af1ca948be80a0140b54f3a3db2db29fa35ccf582e"} Nov 21 14:39:50 crc kubenswrapper[4897]: I1121 14:39:50.402923 4897 scope.go:117] "RemoveContainer" containerID="762160a7ac40ed22f4f6b9f64c7579c5e61b7da769018e0e19d8fedd5c0382ff" Nov 21 14:39:50 crc kubenswrapper[4897]: I1121 14:39:50.445326 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7cfe6c1-f519-438c-b6f1-fea0fcc70cc3-public-tls-certs\") pod \"d7cfe6c1-f519-438c-b6f1-fea0fcc70cc3\" (UID: \"d7cfe6c1-f519-438c-b6f1-fea0fcc70cc3\") " Nov 21 14:39:50 crc kubenswrapper[4897]: I1121 14:39:50.445822 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7cfe6c1-f519-438c-b6f1-fea0fcc70cc3-config-data\") pod \"d7cfe6c1-f519-438c-b6f1-fea0fcc70cc3\" (UID: \"d7cfe6c1-f519-438c-b6f1-fea0fcc70cc3\") " Nov 21 14:39:50 crc kubenswrapper[4897]: I1121 14:39:50.447080 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5g4k9\" (UniqueName: \"kubernetes.io/projected/d7cfe6c1-f519-438c-b6f1-fea0fcc70cc3-kube-api-access-5g4k9\") pod \"d7cfe6c1-f519-438c-b6f1-fea0fcc70cc3\" (UID: \"d7cfe6c1-f519-438c-b6f1-fea0fcc70cc3\") " Nov 21 14:39:50 crc kubenswrapper[4897]: I1121 14:39:50.447468 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d7cfe6c1-f519-438c-b6f1-fea0fcc70cc3-config-data-custom\") pod \"d7cfe6c1-f519-438c-b6f1-fea0fcc70cc3\" (UID: \"d7cfe6c1-f519-438c-b6f1-fea0fcc70cc3\") " Nov 21 14:39:50 crc kubenswrapper[4897]: I1121 14:39:50.447811 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7cfe6c1-f519-438c-b6f1-fea0fcc70cc3-internal-tls-certs\") pod \"d7cfe6c1-f519-438c-b6f1-fea0fcc70cc3\" (UID: \"d7cfe6c1-f519-438c-b6f1-fea0fcc70cc3\") " Nov 21 14:39:50 crc kubenswrapper[4897]: I1121 14:39:50.450614 4897 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7cfe6c1-f519-438c-b6f1-fea0fcc70cc3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 21 14:39:50 crc kubenswrapper[4897]: I1121 14:39:50.468211 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7cfe6c1-f519-438c-b6f1-fea0fcc70cc3-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "d7cfe6c1-f519-438c-b6f1-fea0fcc70cc3" (UID: "d7cfe6c1-f519-438c-b6f1-fea0fcc70cc3"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:39:50 crc kubenswrapper[4897]: I1121 14:39:50.529651 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7cfe6c1-f519-438c-b6f1-fea0fcc70cc3-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "d7cfe6c1-f519-438c-b6f1-fea0fcc70cc3" (UID: "d7cfe6c1-f519-438c-b6f1-fea0fcc70cc3"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:39:50 crc kubenswrapper[4897]: I1121 14:39:50.531912 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7cfe6c1-f519-438c-b6f1-fea0fcc70cc3-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "d7cfe6c1-f519-438c-b6f1-fea0fcc70cc3" (UID: "d7cfe6c1-f519-438c-b6f1-fea0fcc70cc3"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:39:50 crc kubenswrapper[4897]: I1121 14:39:50.538791 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7cfe6c1-f519-438c-b6f1-fea0fcc70cc3-config-data" (OuterVolumeSpecName: "config-data") pod "d7cfe6c1-f519-438c-b6f1-fea0fcc70cc3" (UID: "d7cfe6c1-f519-438c-b6f1-fea0fcc70cc3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:39:50 crc kubenswrapper[4897]: I1121 14:39:50.551992 4897 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7cfe6c1-f519-438c-b6f1-fea0fcc70cc3-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 21 14:39:50 crc kubenswrapper[4897]: I1121 14:39:50.552942 4897 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7cfe6c1-f519-438c-b6f1-fea0fcc70cc3-config-data\") on node \"crc\" DevicePath \"\"" Nov 21 14:39:50 crc kubenswrapper[4897]: I1121 14:39:50.552966 4897 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d7cfe6c1-f519-438c-b6f1-fea0fcc70cc3-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 21 14:39:50 crc kubenswrapper[4897]: I1121 14:39:50.552978 4897 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7cfe6c1-f519-438c-b6f1-fea0fcc70cc3-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 21 14:39:50 crc kubenswrapper[4897]: I1121 14:39:50.641415 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7cfe6c1-f519-438c-b6f1-fea0fcc70cc3-kube-api-access-5g4k9" (OuterVolumeSpecName: "kube-api-access-5g4k9") pod "d7cfe6c1-f519-438c-b6f1-fea0fcc70cc3" (UID: "d7cfe6c1-f519-438c-b6f1-fea0fcc70cc3"). InnerVolumeSpecName "kube-api-access-5g4k9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:39:50 crc kubenswrapper[4897]: I1121 14:39:50.655585 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5g4k9\" (UniqueName: \"kubernetes.io/projected/d7cfe6c1-f519-438c-b6f1-fea0fcc70cc3-kube-api-access-5g4k9\") on node \"crc\" DevicePath \"\"" Nov 21 14:39:51 crc kubenswrapper[4897]: I1121 14:39:51.207156 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-5d8b84cf9f-4lz8g" Nov 21 14:39:51 crc kubenswrapper[4897]: I1121 14:39:51.259027 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-5d8b84cf9f-4lz8g"] Nov 21 14:39:51 crc kubenswrapper[4897]: I1121 14:39:51.276248 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-api-5d8b84cf9f-4lz8g"] Nov 21 14:39:52 crc kubenswrapper[4897]: I1121 14:39:52.104711 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d7cfe6c1-f519-438c-b6f1-fea0fcc70cc3" path="/var/lib/kubelet/pods/d7cfe6c1-f519-438c-b6f1-fea0fcc70cc3/volumes" Nov 21 14:39:52 crc kubenswrapper[4897]: I1121 14:39:52.974342 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-sync-ctn4r"] Nov 21 14:39:52 crc kubenswrapper[4897]: I1121 14:39:52.991588 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-sync-ctn4r"] Nov 21 14:39:53 crc kubenswrapper[4897]: I1121 14:39:53.055875 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-sync-slsml"] Nov 21 14:39:53 crc kubenswrapper[4897]: E1121 14:39:53.056345 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a5df1f3-8748-4324-bf70-bf50dc091710" containerName="registry-server" Nov 21 14:39:53 crc kubenswrapper[4897]: I1121 14:39:53.056362 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a5df1f3-8748-4324-bf70-bf50dc091710" containerName="registry-server" Nov 21 14:39:53 crc kubenswrapper[4897]: E1121 14:39:53.056376 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7cfe6c1-f519-438c-b6f1-fea0fcc70cc3" containerName="heat-api" Nov 21 14:39:53 crc kubenswrapper[4897]: I1121 14:39:53.056382 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7cfe6c1-f519-438c-b6f1-fea0fcc70cc3" containerName="heat-api" Nov 21 14:39:53 crc kubenswrapper[4897]: E1121 14:39:53.056393 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a5df1f3-8748-4324-bf70-bf50dc091710" containerName="extract-content" Nov 21 14:39:53 crc kubenswrapper[4897]: I1121 14:39:53.056399 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a5df1f3-8748-4324-bf70-bf50dc091710" containerName="extract-content" Nov 21 14:39:53 crc kubenswrapper[4897]: E1121 14:39:53.056413 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a5df1f3-8748-4324-bf70-bf50dc091710" containerName="extract-utilities" Nov 21 14:39:53 crc kubenswrapper[4897]: I1121 14:39:53.056419 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a5df1f3-8748-4324-bf70-bf50dc091710" containerName="extract-utilities" Nov 21 14:39:53 crc kubenswrapper[4897]: E1121 14:39:53.056454 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0bf363b7-2fe8-4676-aa3c-4ae74c2c7d05" containerName="heat-cfnapi" Nov 21 14:39:53 crc kubenswrapper[4897]: I1121 14:39:53.056460 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="0bf363b7-2fe8-4676-aa3c-4ae74c2c7d05" containerName="heat-cfnapi" Nov 21 14:39:53 crc kubenswrapper[4897]: I1121 14:39:53.056721 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7cfe6c1-f519-438c-b6f1-fea0fcc70cc3" containerName="heat-api" Nov 21 14:39:53 crc kubenswrapper[4897]: I1121 14:39:53.056754 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a5df1f3-8748-4324-bf70-bf50dc091710" containerName="registry-server" Nov 21 14:39:53 crc kubenswrapper[4897]: I1121 14:39:53.056781 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="0bf363b7-2fe8-4676-aa3c-4ae74c2c7d05" containerName="heat-cfnapi" Nov 21 14:39:53 crc kubenswrapper[4897]: I1121 14:39:53.057620 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-slsml" Nov 21 14:39:53 crc kubenswrapper[4897]: I1121 14:39:53.066355 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 21 14:39:53 crc kubenswrapper[4897]: I1121 14:39:53.072251 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-slsml"] Nov 21 14:39:53 crc kubenswrapper[4897]: I1121 14:39:53.210471 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9422887-78a4-47c5-99ed-89f4ed9a97e9-combined-ca-bundle\") pod \"aodh-db-sync-slsml\" (UID: \"f9422887-78a4-47c5-99ed-89f4ed9a97e9\") " pod="openstack/aodh-db-sync-slsml" Nov 21 14:39:53 crc kubenswrapper[4897]: I1121 14:39:53.211234 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4q5kw\" (UniqueName: \"kubernetes.io/projected/f9422887-78a4-47c5-99ed-89f4ed9a97e9-kube-api-access-4q5kw\") pod \"aodh-db-sync-slsml\" (UID: \"f9422887-78a4-47c5-99ed-89f4ed9a97e9\") " pod="openstack/aodh-db-sync-slsml" Nov 21 14:39:53 crc kubenswrapper[4897]: I1121 14:39:53.211355 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9422887-78a4-47c5-99ed-89f4ed9a97e9-config-data\") pod \"aodh-db-sync-slsml\" (UID: \"f9422887-78a4-47c5-99ed-89f4ed9a97e9\") " pod="openstack/aodh-db-sync-slsml" Nov 21 14:39:53 crc kubenswrapper[4897]: I1121 14:39:53.211457 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f9422887-78a4-47c5-99ed-89f4ed9a97e9-scripts\") pod \"aodh-db-sync-slsml\" (UID: \"f9422887-78a4-47c5-99ed-89f4ed9a97e9\") " pod="openstack/aodh-db-sync-slsml" Nov 21 14:39:53 crc kubenswrapper[4897]: I1121 14:39:53.315883 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f9422887-78a4-47c5-99ed-89f4ed9a97e9-scripts\") pod \"aodh-db-sync-slsml\" (UID: \"f9422887-78a4-47c5-99ed-89f4ed9a97e9\") " pod="openstack/aodh-db-sync-slsml" Nov 21 14:39:53 crc kubenswrapper[4897]: I1121 14:39:53.316059 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9422887-78a4-47c5-99ed-89f4ed9a97e9-combined-ca-bundle\") pod \"aodh-db-sync-slsml\" (UID: \"f9422887-78a4-47c5-99ed-89f4ed9a97e9\") " pod="openstack/aodh-db-sync-slsml" Nov 21 14:39:53 crc kubenswrapper[4897]: I1121 14:39:53.316180 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4q5kw\" (UniqueName: \"kubernetes.io/projected/f9422887-78a4-47c5-99ed-89f4ed9a97e9-kube-api-access-4q5kw\") pod \"aodh-db-sync-slsml\" (UID: \"f9422887-78a4-47c5-99ed-89f4ed9a97e9\") " pod="openstack/aodh-db-sync-slsml" Nov 21 14:39:53 crc kubenswrapper[4897]: I1121 14:39:53.316219 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9422887-78a4-47c5-99ed-89f4ed9a97e9-config-data\") pod \"aodh-db-sync-slsml\" (UID: \"f9422887-78a4-47c5-99ed-89f4ed9a97e9\") " pod="openstack/aodh-db-sync-slsml" Nov 21 14:39:53 crc kubenswrapper[4897]: I1121 14:39:53.321894 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9422887-78a4-47c5-99ed-89f4ed9a97e9-combined-ca-bundle\") pod \"aodh-db-sync-slsml\" (UID: \"f9422887-78a4-47c5-99ed-89f4ed9a97e9\") " pod="openstack/aodh-db-sync-slsml" Nov 21 14:39:53 crc kubenswrapper[4897]: I1121 14:39:53.326258 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f9422887-78a4-47c5-99ed-89f4ed9a97e9-scripts\") pod \"aodh-db-sync-slsml\" (UID: \"f9422887-78a4-47c5-99ed-89f4ed9a97e9\") " pod="openstack/aodh-db-sync-slsml" Nov 21 14:39:53 crc kubenswrapper[4897]: I1121 14:39:53.332106 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9422887-78a4-47c5-99ed-89f4ed9a97e9-config-data\") pod \"aodh-db-sync-slsml\" (UID: \"f9422887-78a4-47c5-99ed-89f4ed9a97e9\") " pod="openstack/aodh-db-sync-slsml" Nov 21 14:39:53 crc kubenswrapper[4897]: I1121 14:39:53.336171 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4q5kw\" (UniqueName: \"kubernetes.io/projected/f9422887-78a4-47c5-99ed-89f4ed9a97e9-kube-api-access-4q5kw\") pod \"aodh-db-sync-slsml\" (UID: \"f9422887-78a4-47c5-99ed-89f4ed9a97e9\") " pod="openstack/aodh-db-sync-slsml" Nov 21 14:39:53 crc kubenswrapper[4897]: I1121 14:39:53.388590 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-slsml" Nov 21 14:39:53 crc kubenswrapper[4897]: I1121 14:39:53.457184 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="9fb2dc84-c75b-45e9-9873-081851fd8ea8" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.1.7:5671: connect: connection refused" Nov 21 14:39:53 crc kubenswrapper[4897]: I1121 14:39:53.730401 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="7a0fd0c8-15d2-45c0-9126-bf7e9e7926e8" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.1.6:5671: connect: connection refused" Nov 21 14:39:54 crc kubenswrapper[4897]: I1121 14:39:54.101710 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c729e1a-4290-4cbf-b38b-d235f229f166" path="/var/lib/kubelet/pods/4c729e1a-4290-4cbf-b38b-d235f229f166/volumes" Nov 21 14:39:58 crc kubenswrapper[4897]: I1121 14:39:58.081553 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-d8nth" podUID="985eca1a-139e-46ce-a2a9-3e68d76b84d0" containerName="registry-server" probeResult="failure" output=< Nov 21 14:39:58 crc kubenswrapper[4897]: timeout: failed to connect service ":50051" within 1s Nov 21 14:39:58 crc kubenswrapper[4897]: > Nov 21 14:39:59 crc kubenswrapper[4897]: I1121 14:39:59.088897 4897 scope.go:117] "RemoveContainer" containerID="c68f1b003d12b4a2964169a85a04e3f0272d86c891f52e6bebd52f3781520d1a" Nov 21 14:39:59 crc kubenswrapper[4897]: E1121 14:39:59.089610 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 14:40:03 crc kubenswrapper[4897]: I1121 14:40:03.457670 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="9fb2dc84-c75b-45e9-9873-081851fd8ea8" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.1.7:5671: connect: connection refused" Nov 21 14:40:03 crc kubenswrapper[4897]: I1121 14:40:03.731579 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="7a0fd0c8-15d2-45c0-9126-bf7e9e7926e8" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.1.6:5671: connect: connection refused" Nov 21 14:40:05 crc kubenswrapper[4897]: I1121 14:40:05.089468 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="9fb2dc84-c75b-45e9-9873-081851fd8ea8" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.1.7:5671: connect: connection refused" Nov 21 14:40:08 crc kubenswrapper[4897]: I1121 14:40:08.071862 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-d8nth" podUID="985eca1a-139e-46ce-a2a9-3e68d76b84d0" containerName="registry-server" probeResult="failure" output=< Nov 21 14:40:08 crc kubenswrapper[4897]: timeout: failed to connect service ":50051" within 1s Nov 21 14:40:08 crc kubenswrapper[4897]: > Nov 21 14:40:08 crc kubenswrapper[4897]: I1121 14:40:08.098590 4897 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 21 14:40:08 crc kubenswrapper[4897]: I1121 14:40:08.102754 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-slsml"] Nov 21 14:40:08 crc kubenswrapper[4897]: I1121 14:40:08.474155 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-slsml" event={"ID":"f9422887-78a4-47c5-99ed-89f4ed9a97e9","Type":"ContainerStarted","Data":"b4572781fb394aa9fd7084435da75123f3dd969172f9f6a556bedc6dff9e871f"} Nov 21 14:40:09 crc kubenswrapper[4897]: E1121 14:40:09.368963 4897 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/openstack-ansibleee-runner:latest" Nov 21 14:40:09 crc kubenswrapper[4897]: E1121 14:40:09.369396 4897 kuberuntime_manager.go:1274] "Unhandled Error" err=< Nov 21 14:40:09 crc kubenswrapper[4897]: container &Container{Name:repo-setup-edpm-deployment-openstack-edpm-ipam,Image:quay.io/openstack-k8s-operators/openstack-ansibleee-runner:latest,Command:[],Args:[ansible-runner run /runner -p playbook.yaml -i repo-setup-edpm-deployment-openstack-edpm-ipam],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:ANSIBLE_VERBOSITY,Value:2,ValueFrom:nil,},EnvVar{Name:RUNNER_PLAYBOOK,Value: Nov 21 14:40:09 crc kubenswrapper[4897]: - hosts: all Nov 21 14:40:09 crc kubenswrapper[4897]: strategy: linear Nov 21 14:40:09 crc kubenswrapper[4897]: tasks: Nov 21 14:40:09 crc kubenswrapper[4897]: - name: Enable podified-repos Nov 21 14:40:09 crc kubenswrapper[4897]: become: true Nov 21 14:40:09 crc kubenswrapper[4897]: ansible.builtin.shell: | Nov 21 14:40:09 crc kubenswrapper[4897]: set -euxo pipefail Nov 21 14:40:09 crc kubenswrapper[4897]: pushd /var/tmp Nov 21 14:40:09 crc kubenswrapper[4897]: curl -sL https://github.com/openstack-k8s-operators/repo-setup/archive/refs/heads/main.tar.gz | tar -xz Nov 21 14:40:09 crc kubenswrapper[4897]: pushd repo-setup-main Nov 21 14:40:09 crc kubenswrapper[4897]: python3 -m venv ./venv Nov 21 14:40:09 crc kubenswrapper[4897]: PBR_VERSION=0.0.0 ./venv/bin/pip install ./ Nov 21 14:40:09 crc kubenswrapper[4897]: ./venv/bin/repo-setup current-podified -b antelope Nov 21 14:40:09 crc kubenswrapper[4897]: popd Nov 21 14:40:09 crc kubenswrapper[4897]: rm -rf repo-setup-main Nov 21 14:40:09 crc kubenswrapper[4897]: Nov 21 14:40:09 crc kubenswrapper[4897]: Nov 21 14:40:09 crc kubenswrapper[4897]: ,ValueFrom:nil,},EnvVar{Name:RUNNER_EXTRA_VARS,Value: Nov 21 14:40:09 crc kubenswrapper[4897]: edpm_override_hosts: openstack-edpm-ipam Nov 21 14:40:09 crc kubenswrapper[4897]: edpm_service_type: repo-setup Nov 21 14:40:09 crc kubenswrapper[4897]: Nov 21 14:40:09 crc kubenswrapper[4897]: Nov 21 14:40:09 crc kubenswrapper[4897]: ,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:repo-setup-combined-ca-bundle,ReadOnly:false,MountPath:/var/lib/openstack/cacerts/repo-setup,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/runner/env/ssh_key,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:inventory,ReadOnly:false,MountPath:/runner/inventory/hosts,SubPath:inventory,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jwlpq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:openstack-aee-default-env,},Optional:*true,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod repo-setup-edpm-deployment-openstack-edpm-ipam-ttw7z_openstack(eea9b320-c0ba-4978-961a-a0e69803ffbf): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled Nov 21 14:40:09 crc kubenswrapper[4897]: > logger="UnhandledError" Nov 21 14:40:09 crc kubenswrapper[4897]: E1121 14:40:09.370540 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"repo-setup-edpm-deployment-openstack-edpm-ipam\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ttw7z" podUID="eea9b320-c0ba-4978-961a-a0e69803ffbf" Nov 21 14:40:09 crc kubenswrapper[4897]: E1121 14:40:09.488204 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"repo-setup-edpm-deployment-openstack-edpm-ipam\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-ansibleee-runner:latest\\\"\"" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ttw7z" podUID="eea9b320-c0ba-4978-961a-a0e69803ffbf" Nov 21 14:40:10 crc kubenswrapper[4897]: I1121 14:40:10.456104 4897 scope.go:117] "RemoveContainer" containerID="8c9411e8019fc5114e4bb8a9ea116f4cf8b427fbab481635f9a469fe2705ff35" Nov 21 14:40:11 crc kubenswrapper[4897]: I1121 14:40:11.089769 4897 scope.go:117] "RemoveContainer" containerID="c68f1b003d12b4a2964169a85a04e3f0272d86c891f52e6bebd52f3781520d1a" Nov 21 14:40:12 crc kubenswrapper[4897]: I1121 14:40:12.537766 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" event={"ID":"e7670227-d280-4847-b882-754429f56b0f","Type":"ContainerStarted","Data":"a4d7f2b343d4dace3a7a2ec8aeabc1074996aaa7dc843ecf94e43c56479ac344"} Nov 21 14:40:13 crc kubenswrapper[4897]: I1121 14:40:13.458175 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="9fb2dc84-c75b-45e9-9873-081851fd8ea8" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.1.7:5671: connect: connection refused" Nov 21 14:40:13 crc kubenswrapper[4897]: I1121 14:40:13.732129 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="7a0fd0c8-15d2-45c0-9126-bf7e9e7926e8" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.1.6:5671: connect: connection refused" Nov 21 14:40:18 crc kubenswrapper[4897]: I1121 14:40:18.074111 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-d8nth" podUID="985eca1a-139e-46ce-a2a9-3e68d76b84d0" containerName="registry-server" probeResult="failure" output=< Nov 21 14:40:18 crc kubenswrapper[4897]: timeout: failed to connect service ":50051" within 1s Nov 21 14:40:18 crc kubenswrapper[4897]: > Nov 21 14:40:22 crc kubenswrapper[4897]: I1121 14:40:22.048021 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-sbnf2"] Nov 21 14:40:22 crc kubenswrapper[4897]: I1121 14:40:22.064579 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-2snxf"] Nov 21 14:40:22 crc kubenswrapper[4897]: I1121 14:40:22.075002 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-sbnf2"] Nov 21 14:40:22 crc kubenswrapper[4897]: I1121 14:40:22.084952 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-2snxf"] Nov 21 14:40:22 crc kubenswrapper[4897]: I1121 14:40:22.102904 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3441939-1c13-4b55-a129-a39f4bed36cd" path="/var/lib/kubelet/pods/d3441939-1c13-4b55-a129-a39f4bed36cd/volumes" Nov 21 14:40:22 crc kubenswrapper[4897]: I1121 14:40:22.849709 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d7a5e302-fb61-41b2-a7d2-ff02160c6679" path="/var/lib/kubelet/pods/d7a5e302-fb61-41b2-a7d2-ff02160c6679/volumes" Nov 21 14:40:23 crc kubenswrapper[4897]: I1121 14:40:23.043425 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-b8bf-account-create-x2vdt"] Nov 21 14:40:23 crc kubenswrapper[4897]: I1121 14:40:23.056547 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-376f-account-create-q6xsf"] Nov 21 14:40:23 crc kubenswrapper[4897]: I1121 14:40:23.071136 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-rkm52"] Nov 21 14:40:23 crc kubenswrapper[4897]: I1121 14:40:23.092847 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-b2f5-account-create-69ls6"] Nov 21 14:40:23 crc kubenswrapper[4897]: I1121 14:40:23.108481 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-b8bf-account-create-x2vdt"] Nov 21 14:40:23 crc kubenswrapper[4897]: I1121 14:40:23.128539 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-rkm52"] Nov 21 14:40:23 crc kubenswrapper[4897]: I1121 14:40:23.143591 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-376f-account-create-q6xsf"] Nov 21 14:40:23 crc kubenswrapper[4897]: I1121 14:40:23.154840 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-b2f5-account-create-69ls6"] Nov 21 14:40:23 crc kubenswrapper[4897]: I1121 14:40:23.458993 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Nov 21 14:40:23 crc kubenswrapper[4897]: I1121 14:40:23.732710 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Nov 21 14:40:24 crc kubenswrapper[4897]: I1121 14:40:24.101383 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="34a64a5d-5b9d-4aac-be3e-96edc721fa3b" path="/var/lib/kubelet/pods/34a64a5d-5b9d-4aac-be3e-96edc721fa3b/volumes" Nov 21 14:40:24 crc kubenswrapper[4897]: I1121 14:40:24.103901 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7501a167-0655-4ad4-bb04-c597a81359ba" path="/var/lib/kubelet/pods/7501a167-0655-4ad4-bb04-c597a81359ba/volumes" Nov 21 14:40:24 crc kubenswrapper[4897]: I1121 14:40:24.105552 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="980c0de3-d435-439f-8477-0dc68fbaf7f0" path="/var/lib/kubelet/pods/980c0de3-d435-439f-8477-0dc68fbaf7f0/volumes" Nov 21 14:40:24 crc kubenswrapper[4897]: I1121 14:40:24.107275 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="994c36f9-336e-49dc-b45d-0eb4bfa0b686" path="/var/lib/kubelet/pods/994c36f9-336e-49dc-b45d-0eb4bfa0b686/volumes" Nov 21 14:40:25 crc kubenswrapper[4897]: I1121 14:40:25.041662 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mysqld-exporter-openstack-db-create-z8nrd"] Nov 21 14:40:25 crc kubenswrapper[4897]: I1121 14:40:25.054674 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mysqld-exporter-95d1-account-create-7vtq6"] Nov 21 14:40:25 crc kubenswrapper[4897]: I1121 14:40:25.065728 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mysqld-exporter-openstack-db-create-z8nrd"] Nov 21 14:40:25 crc kubenswrapper[4897]: I1121 14:40:25.076473 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mysqld-exporter-95d1-account-create-7vtq6"] Nov 21 14:40:26 crc kubenswrapper[4897]: I1121 14:40:26.102011 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5d1468bb-2545-4906-a10c-c95ba9ef8693" path="/var/lib/kubelet/pods/5d1468bb-2545-4906-a10c-c95ba9ef8693/volumes" Nov 21 14:40:26 crc kubenswrapper[4897]: I1121 14:40:26.107151 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eb1688c2-d3d5-4935-b08b-32ef2ce7f6e1" path="/var/lib/kubelet/pods/eb1688c2-d3d5-4935-b08b-32ef2ce7f6e1/volumes" Nov 21 14:40:28 crc kubenswrapper[4897]: I1121 14:40:28.075327 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-d8nth" podUID="985eca1a-139e-46ce-a2a9-3e68d76b84d0" containerName="registry-server" probeResult="failure" output=< Nov 21 14:40:28 crc kubenswrapper[4897]: timeout: failed to connect service ":50051" within 1s Nov 21 14:40:28 crc kubenswrapper[4897]: > Nov 21 14:40:34 crc kubenswrapper[4897]: I1121 14:40:34.046841 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mysqld-exporter-openstack-cell1-db-create-jb9vc"] Nov 21 14:40:34 crc kubenswrapper[4897]: I1121 14:40:34.066642 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mysqld-exporter-openstack-cell1-db-create-jb9vc"] Nov 21 14:40:34 crc kubenswrapper[4897]: I1121 14:40:34.082444 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mysqld-exporter-b245-account-create-pprkv"] Nov 21 14:40:34 crc kubenswrapper[4897]: I1121 14:40:34.246858 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0722112c-c860-4cb0-ac80-9cf32bd12f39" path="/var/lib/kubelet/pods/0722112c-c860-4cb0-ac80-9cf32bd12f39/volumes" Nov 21 14:40:34 crc kubenswrapper[4897]: I1121 14:40:34.248014 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mysqld-exporter-b245-account-create-pprkv"] Nov 21 14:40:36 crc kubenswrapper[4897]: I1121 14:40:36.205966 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d56425f9-2cc4-4936-bda7-64520e84f5ab" path="/var/lib/kubelet/pods/d56425f9-2cc4-4936-bda7-64520e84f5ab/volumes" Nov 21 14:40:38 crc kubenswrapper[4897]: I1121 14:40:38.097247 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-d8nth" podUID="985eca1a-139e-46ce-a2a9-3e68d76b84d0" containerName="registry-server" probeResult="failure" output=< Nov 21 14:40:38 crc kubenswrapper[4897]: timeout: failed to connect service ":50051" within 1s Nov 21 14:40:38 crc kubenswrapper[4897]: > Nov 21 14:40:48 crc kubenswrapper[4897]: I1121 14:40:48.106818 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-d8nth" podUID="985eca1a-139e-46ce-a2a9-3e68d76b84d0" containerName="registry-server" probeResult="failure" output=< Nov 21 14:40:48 crc kubenswrapper[4897]: timeout: failed to connect service ":50051" within 1s Nov 21 14:40:48 crc kubenswrapper[4897]: > Nov 21 14:40:56 crc kubenswrapper[4897]: I1121 14:40:56.061761 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-dpnlm"] Nov 21 14:40:56 crc kubenswrapper[4897]: I1121 14:40:56.079897 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-cgbsk"] Nov 21 14:40:56 crc kubenswrapper[4897]: I1121 14:40:56.106423 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-cgbsk"] Nov 21 14:40:56 crc kubenswrapper[4897]: I1121 14:40:56.111355 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-5f6b-account-create-29rlc"] Nov 21 14:40:56 crc kubenswrapper[4897]: I1121 14:40:56.121574 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-hrqhb"] Nov 21 14:40:56 crc kubenswrapper[4897]: I1121 14:40:56.132108 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-c4c2-account-create-kzvn8"] Nov 21 14:40:56 crc kubenswrapper[4897]: I1121 14:40:56.144335 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-f2a9-account-create-rx44g"] Nov 21 14:40:56 crc kubenswrapper[4897]: I1121 14:40:56.155047 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-5f6b-account-create-29rlc"] Nov 21 14:40:56 crc kubenswrapper[4897]: I1121 14:40:56.165310 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-dpnlm"] Nov 21 14:40:56 crc kubenswrapper[4897]: I1121 14:40:56.176277 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-f2a9-account-create-rx44g"] Nov 21 14:40:56 crc kubenswrapper[4897]: I1121 14:40:56.185404 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-hrqhb"] Nov 21 14:40:56 crc kubenswrapper[4897]: I1121 14:40:56.195049 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-c4c2-account-create-kzvn8"] Nov 21 14:40:57 crc kubenswrapper[4897]: I1121 14:40:57.051017 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-create-hg4dh"] Nov 21 14:40:57 crc kubenswrapper[4897]: I1121 14:40:57.063609 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-a766-account-create-btlq4"] Nov 21 14:40:57 crc kubenswrapper[4897]: I1121 14:40:57.090379 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-create-hg4dh"] Nov 21 14:40:57 crc kubenswrapper[4897]: I1121 14:40:57.100959 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-a766-account-create-btlq4"] Nov 21 14:40:58 crc kubenswrapper[4897]: I1121 14:40:58.095109 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-d8nth" podUID="985eca1a-139e-46ce-a2a9-3e68d76b84d0" containerName="registry-server" probeResult="failure" output=< Nov 21 14:40:58 crc kubenswrapper[4897]: timeout: failed to connect service ":50051" within 1s Nov 21 14:40:58 crc kubenswrapper[4897]: > Nov 21 14:40:58 crc kubenswrapper[4897]: I1121 14:40:58.108024 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1fd99443-7153-4711-9575-6ebc4fbfcf42" path="/var/lib/kubelet/pods/1fd99443-7153-4711-9575-6ebc4fbfcf42/volumes" Nov 21 14:40:58 crc kubenswrapper[4897]: I1121 14:40:58.110789 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25643e36-c5eb-468b-85ce-e85cff1ad090" path="/var/lib/kubelet/pods/25643e36-c5eb-468b-85ce-e85cff1ad090/volumes" Nov 21 14:40:58 crc kubenswrapper[4897]: I1121 14:40:58.115007 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="56c583f0-0dba-4e94-a8b7-26d7b5c0e87f" path="/var/lib/kubelet/pods/56c583f0-0dba-4e94-a8b7-26d7b5c0e87f/volumes" Nov 21 14:40:58 crc kubenswrapper[4897]: I1121 14:40:58.117937 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="67dd6a02-c8f6-4035-ad18-c6bbea13c688" path="/var/lib/kubelet/pods/67dd6a02-c8f6-4035-ad18-c6bbea13c688/volumes" Nov 21 14:40:58 crc kubenswrapper[4897]: I1121 14:40:58.118556 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8bb39c02-70b6-470f-bf62-add1c2ebc02f" path="/var/lib/kubelet/pods/8bb39c02-70b6-470f-bf62-add1c2ebc02f/volumes" Nov 21 14:40:58 crc kubenswrapper[4897]: I1121 14:40:58.178582 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="982ba7b6-1d19-4bd4-bdea-c828fd678186" path="/var/lib/kubelet/pods/982ba7b6-1d19-4bd4-bdea-c828fd678186/volumes" Nov 21 14:40:58 crc kubenswrapper[4897]: I1121 14:40:58.186710 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a69b6579-ce5a-44d7-ab3d-5454dd5f85df" path="/var/lib/kubelet/pods/a69b6579-ce5a-44d7-ab3d-5454dd5f85df/volumes" Nov 21 14:40:58 crc kubenswrapper[4897]: I1121 14:40:58.187752 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee8a4380-421f-45b1-ba19-9832b69935fe" path="/var/lib/kubelet/pods/ee8a4380-421f-45b1-ba19-9832b69935fe/volumes" Nov 21 14:41:01 crc kubenswrapper[4897]: I1121 14:41:01.037952 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-227rw"] Nov 21 14:41:01 crc kubenswrapper[4897]: I1121 14:41:01.052207 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-227rw"] Nov 21 14:41:02 crc kubenswrapper[4897]: I1121 14:41:02.104233 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="208ac008-5901-4b46-983a-a73c9ed207f2" path="/var/lib/kubelet/pods/208ac008-5901-4b46-983a-a73c9ed207f2/volumes" Nov 21 14:41:02 crc kubenswrapper[4897]: I1121 14:41:02.698038 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 21 14:41:02 crc kubenswrapper[4897]: I1121 14:41:02.731869 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 21 14:41:05 crc kubenswrapper[4897]: I1121 14:41:05.201344 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-slsml" event={"ID":"f9422887-78a4-47c5-99ed-89f4ed9a97e9","Type":"ContainerStarted","Data":"f86960bf3fcbb0166df044017ccd3c080affb6d27e16d204771c87bffedbeb96"} Nov 21 14:41:05 crc kubenswrapper[4897]: I1121 14:41:05.203061 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ttw7z" event={"ID":"eea9b320-c0ba-4978-961a-a0e69803ffbf","Type":"ContainerStarted","Data":"a6a44b11937da00ead44bd0f6bb5dbb6cedb199f81c3cfdba70c148a915ad16c"} Nov 21 14:41:06 crc kubenswrapper[4897]: I1121 14:41:06.235497 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ttw7z" podStartSLOduration=21.035209789 podStartE2EDuration="1m39.235480774s" podCreationTimestamp="2025-11-21 14:39:27 +0000 UTC" firstStartedPulling="2025-11-21 14:39:44.492722577 +0000 UTC m=+1861.777316052" lastFinishedPulling="2025-11-21 14:41:02.692993562 +0000 UTC m=+1939.977587037" observedRunningTime="2025-11-21 14:41:06.232001592 +0000 UTC m=+1943.516595067" watchObservedRunningTime="2025-11-21 14:41:06.235480774 +0000 UTC m=+1943.520074249" Nov 21 14:41:06 crc kubenswrapper[4897]: I1121 14:41:06.264322 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-db-sync-slsml" podStartSLOduration=18.634953773 podStartE2EDuration="1m13.264280194s" podCreationTimestamp="2025-11-21 14:39:53 +0000 UTC" firstStartedPulling="2025-11-21 14:40:08.098381248 +0000 UTC m=+1885.382974723" lastFinishedPulling="2025-11-21 14:41:02.727707669 +0000 UTC m=+1940.012301144" observedRunningTime="2025-11-21 14:41:06.258944933 +0000 UTC m=+1943.543538418" watchObservedRunningTime="2025-11-21 14:41:06.264280194 +0000 UTC m=+1943.548873669" Nov 21 14:41:08 crc kubenswrapper[4897]: I1121 14:41:08.104905 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-d8nth" podUID="985eca1a-139e-46ce-a2a9-3e68d76b84d0" containerName="registry-server" probeResult="failure" output=< Nov 21 14:41:08 crc kubenswrapper[4897]: timeout: failed to connect service ":50051" within 1s Nov 21 14:41:08 crc kubenswrapper[4897]: > Nov 21 14:41:10 crc kubenswrapper[4897]: I1121 14:41:10.733907 4897 scope.go:117] "RemoveContainer" containerID="f611821b857e8dab1b5c82f556bf92d1d63333064f5fd2e8326c8cc267cfa952" Nov 21 14:41:12 crc kubenswrapper[4897]: I1121 14:41:12.621435 4897 scope.go:117] "RemoveContainer" containerID="f92ea24bf1f109608a14fbb251bce2cef26cc268df4c3daa8daeb16e4d2f599d" Nov 21 14:41:12 crc kubenswrapper[4897]: I1121 14:41:12.707396 4897 scope.go:117] "RemoveContainer" containerID="22185b922cd0f1f68e876cd66d956a126099e34ca5c3647d107d2adc78cb4383" Nov 21 14:41:12 crc kubenswrapper[4897]: I1121 14:41:12.852819 4897 scope.go:117] "RemoveContainer" containerID="848d3bf026757c3eac458bc2ae8f7730ed6a47c48b0f7cf0f0de11da3cf9f147" Nov 21 14:41:12 crc kubenswrapper[4897]: I1121 14:41:12.940256 4897 scope.go:117] "RemoveContainer" containerID="ede6a9437617ca4a9d2c0f1600f1fe7faedf4bcf189909eb8f3195b61567ed58" Nov 21 14:41:13 crc kubenswrapper[4897]: I1121 14:41:13.081859 4897 scope.go:117] "RemoveContainer" containerID="2a2f30c93a24730d081746bf1340d27054ff14feeb84ac94d2f98b5707eb77f6" Nov 21 14:41:13 crc kubenswrapper[4897]: I1121 14:41:13.234390 4897 scope.go:117] "RemoveContainer" containerID="940d3d57b3424e16b5ced6b938dfcf9dc5e5dff3244c38fa7a71385046318ac2" Nov 21 14:41:13 crc kubenswrapper[4897]: I1121 14:41:13.897860 4897 scope.go:117] "RemoveContainer" containerID="980de6ce7e9fd815f36367eb491b1f46149c4541db0944212ac301ea1c5c30c1" Nov 21 14:41:14 crc kubenswrapper[4897]: I1121 14:41:14.065785 4897 scope.go:117] "RemoveContainer" containerID="fce9d4dbbab86c5b7bd18a3bc4637c387571935a1eb7826370d6345f299fe008" Nov 21 14:41:14 crc kubenswrapper[4897]: I1121 14:41:14.280183 4897 scope.go:117] "RemoveContainer" containerID="a593f235df64892b4938a5931d426f32f36f7bee61dd522b3168c4919dce75a6" Nov 21 14:41:14 crc kubenswrapper[4897]: I1121 14:41:14.386251 4897 scope.go:117] "RemoveContainer" containerID="860f4995b77cd310712c0190183323bddec94b2b3f7a6bf682efd550850a9716" Nov 21 14:41:14 crc kubenswrapper[4897]: I1121 14:41:14.479143 4897 scope.go:117] "RemoveContainer" containerID="8781c24d75d3e6b5d39b1e57310980d15f0af48994ee70807841ba07c7ff9b3c" Nov 21 14:41:14 crc kubenswrapper[4897]: I1121 14:41:14.607899 4897 scope.go:117] "RemoveContainer" containerID="9522822e4d3514856f080fe94bb45b833c81336e3fd4d133f5c86ec83ec51d64" Nov 21 14:41:14 crc kubenswrapper[4897]: I1121 14:41:14.637822 4897 scope.go:117] "RemoveContainer" containerID="0cb696c4ee7260fa5bb8919890a7625680b2a0d45b07bff3285343319b1ea02f" Nov 21 14:41:14 crc kubenswrapper[4897]: I1121 14:41:14.771878 4897 scope.go:117] "RemoveContainer" containerID="8de5ff5ee2cf9cf2c61e0a6143f2ebc4b16a007274b3b322e3a275de8e4ce95a" Nov 21 14:41:14 crc kubenswrapper[4897]: I1121 14:41:14.825565 4897 scope.go:117] "RemoveContainer" containerID="0a09a0b4d2bf00b873dd3a26366c3cb500c70dfa74206e2efd43e17e17491e68" Nov 21 14:41:14 crc kubenswrapper[4897]: I1121 14:41:14.928779 4897 scope.go:117] "RemoveContainer" containerID="15cd2591cd330627ec6bc88cb3b35efc10d63e1d88ab6a3b6123dcf88f489ca8" Nov 21 14:41:14 crc kubenswrapper[4897]: I1121 14:41:14.998705 4897 scope.go:117] "RemoveContainer" containerID="ca081cb9216b8b6726dc8d2233dd7e385f478246a9c0246f412237ec6aa83a6a" Nov 21 14:41:15 crc kubenswrapper[4897]: I1121 14:41:15.441873 4897 scope.go:117] "RemoveContainer" containerID="587c7c12071ac71050fe033554b6b7a03bb00d1e6b792ed3be8c7fe6049d2e25" Nov 21 14:41:15 crc kubenswrapper[4897]: I1121 14:41:15.580769 4897 scope.go:117] "RemoveContainer" containerID="1675902aaf6ae6196f179d75e509846f5abd6d1664aa148b6578ceea932c4544" Nov 21 14:41:18 crc kubenswrapper[4897]: I1121 14:41:18.072926 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-d8nth" podUID="985eca1a-139e-46ce-a2a9-3e68d76b84d0" containerName="registry-server" probeResult="failure" output=< Nov 21 14:41:18 crc kubenswrapper[4897]: timeout: failed to connect service ":50051" within 1s Nov 21 14:41:18 crc kubenswrapper[4897]: > Nov 21 14:41:18 crc kubenswrapper[4897]: I1121 14:41:18.073492 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-d8nth" Nov 21 14:41:18 crc kubenswrapper[4897]: I1121 14:41:18.074364 4897 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="registry-server" containerStatusID={"Type":"cri-o","ID":"d687cc3a224dedde2763d0f27045e20eb89735f70f16352147e95af90a6720c2"} pod="openshift-marketplace/community-operators-d8nth" containerMessage="Container registry-server failed startup probe, will be restarted" Nov 21 14:41:18 crc kubenswrapper[4897]: I1121 14:41:18.074454 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-d8nth" podUID="985eca1a-139e-46ce-a2a9-3e68d76b84d0" containerName="registry-server" containerID="cri-o://d687cc3a224dedde2763d0f27045e20eb89735f70f16352147e95af90a6720c2" gracePeriod=30 Nov 21 14:41:28 crc kubenswrapper[4897]: I1121 14:41:28.839799 4897 generic.go:334] "Generic (PLEG): container finished" podID="eea9b320-c0ba-4978-961a-a0e69803ffbf" containerID="a6a44b11937da00ead44bd0f6bb5dbb6cedb199f81c3cfdba70c148a915ad16c" exitCode=0 Nov 21 14:41:28 crc kubenswrapper[4897]: I1121 14:41:28.839927 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ttw7z" event={"ID":"eea9b320-c0ba-4978-961a-a0e69803ffbf","Type":"ContainerDied","Data":"a6a44b11937da00ead44bd0f6bb5dbb6cedb199f81c3cfdba70c148a915ad16c"} Nov 21 14:41:36 crc kubenswrapper[4897]: I1121 14:41:36.795148 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ttw7z" Nov 21 14:41:36 crc kubenswrapper[4897]: I1121 14:41:36.919585 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/eea9b320-c0ba-4978-961a-a0e69803ffbf-inventory\") pod \"eea9b320-c0ba-4978-961a-a0e69803ffbf\" (UID: \"eea9b320-c0ba-4978-961a-a0e69803ffbf\") " Nov 21 14:41:36 crc kubenswrapper[4897]: I1121 14:41:36.919700 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/eea9b320-c0ba-4978-961a-a0e69803ffbf-ssh-key\") pod \"eea9b320-c0ba-4978-961a-a0e69803ffbf\" (UID: \"eea9b320-c0ba-4978-961a-a0e69803ffbf\") " Nov 21 14:41:36 crc kubenswrapper[4897]: I1121 14:41:36.919731 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jwlpq\" (UniqueName: \"kubernetes.io/projected/eea9b320-c0ba-4978-961a-a0e69803ffbf-kube-api-access-jwlpq\") pod \"eea9b320-c0ba-4978-961a-a0e69803ffbf\" (UID: \"eea9b320-c0ba-4978-961a-a0e69803ffbf\") " Nov 21 14:41:36 crc kubenswrapper[4897]: I1121 14:41:36.919898 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eea9b320-c0ba-4978-961a-a0e69803ffbf-repo-setup-combined-ca-bundle\") pod \"eea9b320-c0ba-4978-961a-a0e69803ffbf\" (UID: \"eea9b320-c0ba-4978-961a-a0e69803ffbf\") " Nov 21 14:41:36 crc kubenswrapper[4897]: I1121 14:41:36.922008 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ttw7z" event={"ID":"eea9b320-c0ba-4978-961a-a0e69803ffbf","Type":"ContainerDied","Data":"eb1f9642817079761f02315242af83860b2839c8bf0c50a266cc44a905ee84c4"} Nov 21 14:41:36 crc kubenswrapper[4897]: I1121 14:41:36.922045 4897 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eb1f9642817079761f02315242af83860b2839c8bf0c50a266cc44a905ee84c4" Nov 21 14:41:36 crc kubenswrapper[4897]: I1121 14:41:36.922098 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ttw7z" Nov 21 14:41:37 crc kubenswrapper[4897]: I1121 14:41:37.087840 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eea9b320-c0ba-4978-961a-a0e69803ffbf-kube-api-access-jwlpq" (OuterVolumeSpecName: "kube-api-access-jwlpq") pod "eea9b320-c0ba-4978-961a-a0e69803ffbf" (UID: "eea9b320-c0ba-4978-961a-a0e69803ffbf"). InnerVolumeSpecName "kube-api-access-jwlpq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:41:37 crc kubenswrapper[4897]: I1121 14:41:37.090649 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eea9b320-c0ba-4978-961a-a0e69803ffbf-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "eea9b320-c0ba-4978-961a-a0e69803ffbf" (UID: "eea9b320-c0ba-4978-961a-a0e69803ffbf"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:41:37 crc kubenswrapper[4897]: I1121 14:41:37.108783 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eea9b320-c0ba-4978-961a-a0e69803ffbf-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "eea9b320-c0ba-4978-961a-a0e69803ffbf" (UID: "eea9b320-c0ba-4978-961a-a0e69803ffbf"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:41:37 crc kubenswrapper[4897]: I1121 14:41:37.110497 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eea9b320-c0ba-4978-961a-a0e69803ffbf-inventory" (OuterVolumeSpecName: "inventory") pod "eea9b320-c0ba-4978-961a-a0e69803ffbf" (UID: "eea9b320-c0ba-4978-961a-a0e69803ffbf"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:41:37 crc kubenswrapper[4897]: I1121 14:41:37.125244 4897 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/eea9b320-c0ba-4978-961a-a0e69803ffbf-inventory\") on node \"crc\" DevicePath \"\"" Nov 21 14:41:37 crc kubenswrapper[4897]: I1121 14:41:37.125296 4897 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/eea9b320-c0ba-4978-961a-a0e69803ffbf-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 21 14:41:37 crc kubenswrapper[4897]: I1121 14:41:37.125307 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jwlpq\" (UniqueName: \"kubernetes.io/projected/eea9b320-c0ba-4978-961a-a0e69803ffbf-kube-api-access-jwlpq\") on node \"crc\" DevicePath \"\"" Nov 21 14:41:37 crc kubenswrapper[4897]: I1121 14:41:37.125319 4897 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eea9b320-c0ba-4978-961a-a0e69803ffbf-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 21 14:41:37 crc kubenswrapper[4897]: I1121 14:41:37.975631 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-d6gh8"] Nov 21 14:41:37 crc kubenswrapper[4897]: E1121 14:41:37.979117 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eea9b320-c0ba-4978-961a-a0e69803ffbf" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 21 14:41:37 crc kubenswrapper[4897]: I1121 14:41:37.979150 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="eea9b320-c0ba-4978-961a-a0e69803ffbf" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 21 14:41:37 crc kubenswrapper[4897]: I1121 14:41:37.979462 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="eea9b320-c0ba-4978-961a-a0e69803ffbf" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 21 14:41:37 crc kubenswrapper[4897]: I1121 14:41:37.980334 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-d6gh8" Nov 21 14:41:37 crc kubenswrapper[4897]: I1121 14:41:37.983480 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-dj88m" Nov 21 14:41:37 crc kubenswrapper[4897]: I1121 14:41:37.983534 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 21 14:41:37 crc kubenswrapper[4897]: I1121 14:41:37.983716 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 21 14:41:37 crc kubenswrapper[4897]: I1121 14:41:37.986743 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 21 14:41:37 crc kubenswrapper[4897]: I1121 14:41:37.992137 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-d6gh8"] Nov 21 14:41:38 crc kubenswrapper[4897]: I1121 14:41:38.150041 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jpkhw\" (UniqueName: \"kubernetes.io/projected/b68042aa-854b-441b-bf30-417b3da9baed-kube-api-access-jpkhw\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-d6gh8\" (UID: \"b68042aa-854b-441b-bf30-417b3da9baed\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-d6gh8" Nov 21 14:41:38 crc kubenswrapper[4897]: I1121 14:41:38.150240 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b68042aa-854b-441b-bf30-417b3da9baed-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-d6gh8\" (UID: \"b68042aa-854b-441b-bf30-417b3da9baed\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-d6gh8" Nov 21 14:41:38 crc kubenswrapper[4897]: I1121 14:41:38.150523 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b68042aa-854b-441b-bf30-417b3da9baed-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-d6gh8\" (UID: \"b68042aa-854b-441b-bf30-417b3da9baed\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-d6gh8" Nov 21 14:41:38 crc kubenswrapper[4897]: I1121 14:41:38.252251 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jpkhw\" (UniqueName: \"kubernetes.io/projected/b68042aa-854b-441b-bf30-417b3da9baed-kube-api-access-jpkhw\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-d6gh8\" (UID: \"b68042aa-854b-441b-bf30-417b3da9baed\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-d6gh8" Nov 21 14:41:38 crc kubenswrapper[4897]: I1121 14:41:38.252379 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b68042aa-854b-441b-bf30-417b3da9baed-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-d6gh8\" (UID: \"b68042aa-854b-441b-bf30-417b3da9baed\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-d6gh8" Nov 21 14:41:38 crc kubenswrapper[4897]: I1121 14:41:38.252479 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b68042aa-854b-441b-bf30-417b3da9baed-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-d6gh8\" (UID: \"b68042aa-854b-441b-bf30-417b3da9baed\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-d6gh8" Nov 21 14:41:38 crc kubenswrapper[4897]: I1121 14:41:38.257770 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b68042aa-854b-441b-bf30-417b3da9baed-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-d6gh8\" (UID: \"b68042aa-854b-441b-bf30-417b3da9baed\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-d6gh8" Nov 21 14:41:38 crc kubenswrapper[4897]: I1121 14:41:38.264023 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b68042aa-854b-441b-bf30-417b3da9baed-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-d6gh8\" (UID: \"b68042aa-854b-441b-bf30-417b3da9baed\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-d6gh8" Nov 21 14:41:38 crc kubenswrapper[4897]: I1121 14:41:38.269247 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jpkhw\" (UniqueName: \"kubernetes.io/projected/b68042aa-854b-441b-bf30-417b3da9baed-kube-api-access-jpkhw\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-d6gh8\" (UID: \"b68042aa-854b-441b-bf30-417b3da9baed\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-d6gh8" Nov 21 14:41:38 crc kubenswrapper[4897]: I1121 14:41:38.313298 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-d6gh8" Nov 21 14:41:48 crc kubenswrapper[4897]: E1121 14:41:48.262721 4897 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod985eca1a_139e_46ce_a2a9_3e68d76b84d0.slice/crio-conmon-d687cc3a224dedde2763d0f27045e20eb89735f70f16352147e95af90a6720c2.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod985eca1a_139e_46ce_a2a9_3e68d76b84d0.slice/crio-d687cc3a224dedde2763d0f27045e20eb89735f70f16352147e95af90a6720c2.scope\": RecentStats: unable to find data in memory cache]" Nov 21 14:41:48 crc kubenswrapper[4897]: E1121 14:41:48.262804 4897 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod985eca1a_139e_46ce_a2a9_3e68d76b84d0.slice/crio-conmon-d687cc3a224dedde2763d0f27045e20eb89735f70f16352147e95af90a6720c2.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod985eca1a_139e_46ce_a2a9_3e68d76b84d0.slice/crio-d687cc3a224dedde2763d0f27045e20eb89735f70f16352147e95af90a6720c2.scope\": RecentStats: unable to find data in memory cache]" Nov 21 14:41:48 crc kubenswrapper[4897]: I1121 14:41:48.341968 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-d6gh8"] Nov 21 14:41:49 crc kubenswrapper[4897]: I1121 14:41:49.044176 4897 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-d8nth_985eca1a-139e-46ce-a2a9-3e68d76b84d0/registry-server/0.log" Nov 21 14:41:49 crc kubenswrapper[4897]: I1121 14:41:49.044934 4897 generic.go:334] "Generic (PLEG): container finished" podID="985eca1a-139e-46ce-a2a9-3e68d76b84d0" containerID="d687cc3a224dedde2763d0f27045e20eb89735f70f16352147e95af90a6720c2" exitCode=137 Nov 21 14:41:49 crc kubenswrapper[4897]: I1121 14:41:49.045005 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d8nth" event={"ID":"985eca1a-139e-46ce-a2a9-3e68d76b84d0","Type":"ContainerDied","Data":"d687cc3a224dedde2763d0f27045e20eb89735f70f16352147e95af90a6720c2"} Nov 21 14:41:49 crc kubenswrapper[4897]: I1121 14:41:49.046711 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-d6gh8" event={"ID":"b68042aa-854b-441b-bf30-417b3da9baed","Type":"ContainerStarted","Data":"9bbdf75aaea80e157fa431ce33474d634c36e1fee91d57660f01a1ad46cbeb41"} Nov 21 14:41:49 crc kubenswrapper[4897]: I1121 14:41:49.425007 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 21 14:41:50 crc kubenswrapper[4897]: I1121 14:41:50.059190 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-d6gh8" event={"ID":"b68042aa-854b-441b-bf30-417b3da9baed","Type":"ContainerStarted","Data":"1259747b9d64b231d3759d5137b056404eef6bccaf69fd3943e6602498bdc6eb"} Nov 21 14:41:51 crc kubenswrapper[4897]: I1121 14:41:51.071335 4897 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-d8nth_985eca1a-139e-46ce-a2a9-3e68d76b84d0/registry-server/0.log" Nov 21 14:41:51 crc kubenswrapper[4897]: I1121 14:41:51.072571 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d8nth" event={"ID":"985eca1a-139e-46ce-a2a9-3e68d76b84d0","Type":"ContainerStarted","Data":"1ccf1e924cde0e2d7a1d25a5c7eb1b03623ad17b10a8a9291f26324d124b0f7d"} Nov 21 14:41:51 crc kubenswrapper[4897]: I1121 14:41:51.074549 4897 generic.go:334] "Generic (PLEG): container finished" podID="f9422887-78a4-47c5-99ed-89f4ed9a97e9" containerID="f86960bf3fcbb0166df044017ccd3c080affb6d27e16d204771c87bffedbeb96" exitCode=0 Nov 21 14:41:51 crc kubenswrapper[4897]: I1121 14:41:51.074631 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-slsml" event={"ID":"f9422887-78a4-47c5-99ed-89f4ed9a97e9","Type":"ContainerDied","Data":"f86960bf3fcbb0166df044017ccd3c080affb6d27e16d204771c87bffedbeb96"} Nov 21 14:41:51 crc kubenswrapper[4897]: I1121 14:41:51.132012 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-d6gh8" podStartSLOduration=13.052795426 podStartE2EDuration="14.13199231s" podCreationTimestamp="2025-11-21 14:41:37 +0000 UTC" firstStartedPulling="2025-11-21 14:41:48.340763243 +0000 UTC m=+1985.625356718" lastFinishedPulling="2025-11-21 14:41:49.419960117 +0000 UTC m=+1986.704553602" observedRunningTime="2025-11-21 14:41:51.131189319 +0000 UTC m=+1988.415782814" watchObservedRunningTime="2025-11-21 14:41:51.13199231 +0000 UTC m=+1988.416585785" Nov 21 14:41:52 crc kubenswrapper[4897]: I1121 14:41:52.510627 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-slsml" Nov 21 14:41:52 crc kubenswrapper[4897]: I1121 14:41:52.657475 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4q5kw\" (UniqueName: \"kubernetes.io/projected/f9422887-78a4-47c5-99ed-89f4ed9a97e9-kube-api-access-4q5kw\") pod \"f9422887-78a4-47c5-99ed-89f4ed9a97e9\" (UID: \"f9422887-78a4-47c5-99ed-89f4ed9a97e9\") " Nov 21 14:41:52 crc kubenswrapper[4897]: I1121 14:41:52.657620 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9422887-78a4-47c5-99ed-89f4ed9a97e9-combined-ca-bundle\") pod \"f9422887-78a4-47c5-99ed-89f4ed9a97e9\" (UID: \"f9422887-78a4-47c5-99ed-89f4ed9a97e9\") " Nov 21 14:41:52 crc kubenswrapper[4897]: I1121 14:41:52.657841 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9422887-78a4-47c5-99ed-89f4ed9a97e9-config-data\") pod \"f9422887-78a4-47c5-99ed-89f4ed9a97e9\" (UID: \"f9422887-78a4-47c5-99ed-89f4ed9a97e9\") " Nov 21 14:41:52 crc kubenswrapper[4897]: I1121 14:41:52.657895 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f9422887-78a4-47c5-99ed-89f4ed9a97e9-scripts\") pod \"f9422887-78a4-47c5-99ed-89f4ed9a97e9\" (UID: \"f9422887-78a4-47c5-99ed-89f4ed9a97e9\") " Nov 21 14:41:52 crc kubenswrapper[4897]: I1121 14:41:52.663596 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f9422887-78a4-47c5-99ed-89f4ed9a97e9-kube-api-access-4q5kw" (OuterVolumeSpecName: "kube-api-access-4q5kw") pod "f9422887-78a4-47c5-99ed-89f4ed9a97e9" (UID: "f9422887-78a4-47c5-99ed-89f4ed9a97e9"). InnerVolumeSpecName "kube-api-access-4q5kw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:41:52 crc kubenswrapper[4897]: I1121 14:41:52.664082 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9422887-78a4-47c5-99ed-89f4ed9a97e9-scripts" (OuterVolumeSpecName: "scripts") pod "f9422887-78a4-47c5-99ed-89f4ed9a97e9" (UID: "f9422887-78a4-47c5-99ed-89f4ed9a97e9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:41:52 crc kubenswrapper[4897]: I1121 14:41:52.691086 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9422887-78a4-47c5-99ed-89f4ed9a97e9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f9422887-78a4-47c5-99ed-89f4ed9a97e9" (UID: "f9422887-78a4-47c5-99ed-89f4ed9a97e9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:41:52 crc kubenswrapper[4897]: I1121 14:41:52.691711 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9422887-78a4-47c5-99ed-89f4ed9a97e9-config-data" (OuterVolumeSpecName: "config-data") pod "f9422887-78a4-47c5-99ed-89f4ed9a97e9" (UID: "f9422887-78a4-47c5-99ed-89f4ed9a97e9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:41:52 crc kubenswrapper[4897]: I1121 14:41:52.760361 4897 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9422887-78a4-47c5-99ed-89f4ed9a97e9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 21 14:41:52 crc kubenswrapper[4897]: I1121 14:41:52.760400 4897 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9422887-78a4-47c5-99ed-89f4ed9a97e9-config-data\") on node \"crc\" DevicePath \"\"" Nov 21 14:41:52 crc kubenswrapper[4897]: I1121 14:41:52.760410 4897 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f9422887-78a4-47c5-99ed-89f4ed9a97e9-scripts\") on node \"crc\" DevicePath \"\"" Nov 21 14:41:52 crc kubenswrapper[4897]: I1121 14:41:52.760419 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4q5kw\" (UniqueName: \"kubernetes.io/projected/f9422887-78a4-47c5-99ed-89f4ed9a97e9-kube-api-access-4q5kw\") on node \"crc\" DevicePath \"\"" Nov 21 14:41:53 crc kubenswrapper[4897]: I1121 14:41:53.097032 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-slsml" Nov 21 14:41:53 crc kubenswrapper[4897]: I1121 14:41:53.097063 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-slsml" event={"ID":"f9422887-78a4-47c5-99ed-89f4ed9a97e9","Type":"ContainerDied","Data":"b4572781fb394aa9fd7084435da75123f3dd969172f9f6a556bedc6dff9e871f"} Nov 21 14:41:53 crc kubenswrapper[4897]: I1121 14:41:53.097108 4897 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b4572781fb394aa9fd7084435da75123f3dd969172f9f6a556bedc6dff9e871f" Nov 21 14:41:53 crc kubenswrapper[4897]: I1121 14:41:53.098947 4897 generic.go:334] "Generic (PLEG): container finished" podID="b68042aa-854b-441b-bf30-417b3da9baed" containerID="1259747b9d64b231d3759d5137b056404eef6bccaf69fd3943e6602498bdc6eb" exitCode=0 Nov 21 14:41:53 crc kubenswrapper[4897]: I1121 14:41:53.098992 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-d6gh8" event={"ID":"b68042aa-854b-441b-bf30-417b3da9baed","Type":"ContainerDied","Data":"1259747b9d64b231d3759d5137b056404eef6bccaf69fd3943e6602498bdc6eb"} Nov 21 14:41:53 crc kubenswrapper[4897]: I1121 14:41:53.466642 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Nov 21 14:41:53 crc kubenswrapper[4897]: I1121 14:41:53.467051 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="1d5dac4b-9e6c-41d5-9d10-87015da49a79" containerName="aodh-api" containerID="cri-o://e65e756fbb3117c206622407abfc9990d581b756bb24310f36d0c24d49164dc0" gracePeriod=30 Nov 21 14:41:53 crc kubenswrapper[4897]: I1121 14:41:53.467144 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="1d5dac4b-9e6c-41d5-9d10-87015da49a79" containerName="aodh-listener" containerID="cri-o://715e768e1b2e43e144c6d93c0e95b4d92949f9b90ed622f285ee9f286c09f298" gracePeriod=30 Nov 21 14:41:53 crc kubenswrapper[4897]: I1121 14:41:53.467180 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="1d5dac4b-9e6c-41d5-9d10-87015da49a79" containerName="aodh-notifier" containerID="cri-o://88984a2b2c0a657d25b890e97a861f232ed8a051dddc4d79a706956451832fce" gracePeriod=30 Nov 21 14:41:53 crc kubenswrapper[4897]: I1121 14:41:53.467209 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="1d5dac4b-9e6c-41d5-9d10-87015da49a79" containerName="aodh-evaluator" containerID="cri-o://0b3e7824d93c1825d32ebe2895717aed9691091cdb0492f9fc2fed0b92d32dd9" gracePeriod=30 Nov 21 14:41:54 crc kubenswrapper[4897]: I1121 14:41:54.114033 4897 generic.go:334] "Generic (PLEG): container finished" podID="1d5dac4b-9e6c-41d5-9d10-87015da49a79" containerID="0b3e7824d93c1825d32ebe2895717aed9691091cdb0492f9fc2fed0b92d32dd9" exitCode=0 Nov 21 14:41:54 crc kubenswrapper[4897]: I1121 14:41:54.114067 4897 generic.go:334] "Generic (PLEG): container finished" podID="1d5dac4b-9e6c-41d5-9d10-87015da49a79" containerID="e65e756fbb3117c206622407abfc9990d581b756bb24310f36d0c24d49164dc0" exitCode=0 Nov 21 14:41:54 crc kubenswrapper[4897]: I1121 14:41:54.114248 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"1d5dac4b-9e6c-41d5-9d10-87015da49a79","Type":"ContainerDied","Data":"0b3e7824d93c1825d32ebe2895717aed9691091cdb0492f9fc2fed0b92d32dd9"} Nov 21 14:41:54 crc kubenswrapper[4897]: I1121 14:41:54.114273 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"1d5dac4b-9e6c-41d5-9d10-87015da49a79","Type":"ContainerDied","Data":"e65e756fbb3117c206622407abfc9990d581b756bb24310f36d0c24d49164dc0"} Nov 21 14:41:54 crc kubenswrapper[4897]: I1121 14:41:54.581835 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-d6gh8" Nov 21 14:41:54 crc kubenswrapper[4897]: I1121 14:41:54.709099 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b68042aa-854b-441b-bf30-417b3da9baed-inventory\") pod \"b68042aa-854b-441b-bf30-417b3da9baed\" (UID: \"b68042aa-854b-441b-bf30-417b3da9baed\") " Nov 21 14:41:54 crc kubenswrapper[4897]: I1121 14:41:54.709203 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jpkhw\" (UniqueName: \"kubernetes.io/projected/b68042aa-854b-441b-bf30-417b3da9baed-kube-api-access-jpkhw\") pod \"b68042aa-854b-441b-bf30-417b3da9baed\" (UID: \"b68042aa-854b-441b-bf30-417b3da9baed\") " Nov 21 14:41:54 crc kubenswrapper[4897]: I1121 14:41:54.709657 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b68042aa-854b-441b-bf30-417b3da9baed-ssh-key\") pod \"b68042aa-854b-441b-bf30-417b3da9baed\" (UID: \"b68042aa-854b-441b-bf30-417b3da9baed\") " Nov 21 14:41:54 crc kubenswrapper[4897]: I1121 14:41:54.718396 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b68042aa-854b-441b-bf30-417b3da9baed-kube-api-access-jpkhw" (OuterVolumeSpecName: "kube-api-access-jpkhw") pod "b68042aa-854b-441b-bf30-417b3da9baed" (UID: "b68042aa-854b-441b-bf30-417b3da9baed"). InnerVolumeSpecName "kube-api-access-jpkhw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:41:54 crc kubenswrapper[4897]: I1121 14:41:54.754383 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b68042aa-854b-441b-bf30-417b3da9baed-inventory" (OuterVolumeSpecName: "inventory") pod "b68042aa-854b-441b-bf30-417b3da9baed" (UID: "b68042aa-854b-441b-bf30-417b3da9baed"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:41:54 crc kubenswrapper[4897]: I1121 14:41:54.755009 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b68042aa-854b-441b-bf30-417b3da9baed-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b68042aa-854b-441b-bf30-417b3da9baed" (UID: "b68042aa-854b-441b-bf30-417b3da9baed"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:41:54 crc kubenswrapper[4897]: I1121 14:41:54.813260 4897 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b68042aa-854b-441b-bf30-417b3da9baed-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 21 14:41:54 crc kubenswrapper[4897]: I1121 14:41:54.813304 4897 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b68042aa-854b-441b-bf30-417b3da9baed-inventory\") on node \"crc\" DevicePath \"\"" Nov 21 14:41:54 crc kubenswrapper[4897]: I1121 14:41:54.813316 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jpkhw\" (UniqueName: \"kubernetes.io/projected/b68042aa-854b-441b-bf30-417b3da9baed-kube-api-access-jpkhw\") on node \"crc\" DevicePath \"\"" Nov 21 14:41:55 crc kubenswrapper[4897]: I1121 14:41:55.125998 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-d6gh8" event={"ID":"b68042aa-854b-441b-bf30-417b3da9baed","Type":"ContainerDied","Data":"9bbdf75aaea80e157fa431ce33474d634c36e1fee91d57660f01a1ad46cbeb41"} Nov 21 14:41:55 crc kubenswrapper[4897]: I1121 14:41:55.126268 4897 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9bbdf75aaea80e157fa431ce33474d634c36e1fee91d57660f01a1ad46cbeb41" Nov 21 14:41:55 crc kubenswrapper[4897]: I1121 14:41:55.126052 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-d6gh8" Nov 21 14:41:55 crc kubenswrapper[4897]: I1121 14:41:55.692905 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k27v5"] Nov 21 14:41:55 crc kubenswrapper[4897]: E1121 14:41:55.693800 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9422887-78a4-47c5-99ed-89f4ed9a97e9" containerName="aodh-db-sync" Nov 21 14:41:55 crc kubenswrapper[4897]: I1121 14:41:55.693832 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9422887-78a4-47c5-99ed-89f4ed9a97e9" containerName="aodh-db-sync" Nov 21 14:41:55 crc kubenswrapper[4897]: E1121 14:41:55.693925 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b68042aa-854b-441b-bf30-417b3da9baed" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Nov 21 14:41:55 crc kubenswrapper[4897]: I1121 14:41:55.693943 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="b68042aa-854b-441b-bf30-417b3da9baed" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Nov 21 14:41:55 crc kubenswrapper[4897]: I1121 14:41:55.694283 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9422887-78a4-47c5-99ed-89f4ed9a97e9" containerName="aodh-db-sync" Nov 21 14:41:55 crc kubenswrapper[4897]: I1121 14:41:55.694328 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="b68042aa-854b-441b-bf30-417b3da9baed" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Nov 21 14:41:55 crc kubenswrapper[4897]: I1121 14:41:55.695816 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k27v5" Nov 21 14:41:55 crc kubenswrapper[4897]: I1121 14:41:55.702190 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 21 14:41:55 crc kubenswrapper[4897]: I1121 14:41:55.702448 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 21 14:41:55 crc kubenswrapper[4897]: I1121 14:41:55.702715 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-dj88m" Nov 21 14:41:55 crc kubenswrapper[4897]: I1121 14:41:55.703312 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 21 14:41:55 crc kubenswrapper[4897]: I1121 14:41:55.715540 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k27v5"] Nov 21 14:41:55 crc kubenswrapper[4897]: I1121 14:41:55.834323 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/29e88709-d2ef-475a-85d2-1573f0719c8f-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k27v5\" (UID: \"29e88709-d2ef-475a-85d2-1573f0719c8f\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k27v5" Nov 21 14:41:55 crc kubenswrapper[4897]: I1121 14:41:55.834388 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2frpp\" (UniqueName: \"kubernetes.io/projected/29e88709-d2ef-475a-85d2-1573f0719c8f-kube-api-access-2frpp\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k27v5\" (UID: \"29e88709-d2ef-475a-85d2-1573f0719c8f\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k27v5" Nov 21 14:41:55 crc kubenswrapper[4897]: I1121 14:41:55.834418 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/29e88709-d2ef-475a-85d2-1573f0719c8f-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k27v5\" (UID: \"29e88709-d2ef-475a-85d2-1573f0719c8f\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k27v5" Nov 21 14:41:55 crc kubenswrapper[4897]: I1121 14:41:55.834576 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29e88709-d2ef-475a-85d2-1573f0719c8f-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k27v5\" (UID: \"29e88709-d2ef-475a-85d2-1573f0719c8f\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k27v5" Nov 21 14:41:55 crc kubenswrapper[4897]: I1121 14:41:55.937053 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/29e88709-d2ef-475a-85d2-1573f0719c8f-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k27v5\" (UID: \"29e88709-d2ef-475a-85d2-1573f0719c8f\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k27v5" Nov 21 14:41:55 crc kubenswrapper[4897]: I1121 14:41:55.937397 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2frpp\" (UniqueName: \"kubernetes.io/projected/29e88709-d2ef-475a-85d2-1573f0719c8f-kube-api-access-2frpp\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k27v5\" (UID: \"29e88709-d2ef-475a-85d2-1573f0719c8f\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k27v5" Nov 21 14:41:55 crc kubenswrapper[4897]: I1121 14:41:55.937647 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/29e88709-d2ef-475a-85d2-1573f0719c8f-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k27v5\" (UID: \"29e88709-d2ef-475a-85d2-1573f0719c8f\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k27v5" Nov 21 14:41:55 crc kubenswrapper[4897]: I1121 14:41:55.937761 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29e88709-d2ef-475a-85d2-1573f0719c8f-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k27v5\" (UID: \"29e88709-d2ef-475a-85d2-1573f0719c8f\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k27v5" Nov 21 14:41:55 crc kubenswrapper[4897]: I1121 14:41:55.942651 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/29e88709-d2ef-475a-85d2-1573f0719c8f-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k27v5\" (UID: \"29e88709-d2ef-475a-85d2-1573f0719c8f\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k27v5" Nov 21 14:41:55 crc kubenswrapper[4897]: I1121 14:41:55.942902 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/29e88709-d2ef-475a-85d2-1573f0719c8f-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k27v5\" (UID: \"29e88709-d2ef-475a-85d2-1573f0719c8f\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k27v5" Nov 21 14:41:55 crc kubenswrapper[4897]: I1121 14:41:55.951650 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29e88709-d2ef-475a-85d2-1573f0719c8f-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k27v5\" (UID: \"29e88709-d2ef-475a-85d2-1573f0719c8f\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k27v5" Nov 21 14:41:55 crc kubenswrapper[4897]: I1121 14:41:55.960723 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2frpp\" (UniqueName: \"kubernetes.io/projected/29e88709-d2ef-475a-85d2-1573f0719c8f-kube-api-access-2frpp\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k27v5\" (UID: \"29e88709-d2ef-475a-85d2-1573f0719c8f\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k27v5" Nov 21 14:41:56 crc kubenswrapper[4897]: I1121 14:41:56.089729 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k27v5" Nov 21 14:41:56 crc kubenswrapper[4897]: W1121 14:41:56.676780 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod29e88709_d2ef_475a_85d2_1573f0719c8f.slice/crio-5228a1c2999b65d7651dc1082c01d96d2b1f9028b50d001ed0bc4ca8c1475d93 WatchSource:0}: Error finding container 5228a1c2999b65d7651dc1082c01d96d2b1f9028b50d001ed0bc4ca8c1475d93: Status 404 returned error can't find the container with id 5228a1c2999b65d7651dc1082c01d96d2b1f9028b50d001ed0bc4ca8c1475d93 Nov 21 14:41:56 crc kubenswrapper[4897]: I1121 14:41:56.685519 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k27v5"] Nov 21 14:41:57 crc kubenswrapper[4897]: I1121 14:41:57.024037 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-d8nth" Nov 21 14:41:57 crc kubenswrapper[4897]: I1121 14:41:57.024093 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-d8nth" Nov 21 14:41:57 crc kubenswrapper[4897]: I1121 14:41:57.075347 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-d8nth" Nov 21 14:41:57 crc kubenswrapper[4897]: I1121 14:41:57.157457 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k27v5" event={"ID":"29e88709-d2ef-475a-85d2-1573f0719c8f","Type":"ContainerStarted","Data":"5228a1c2999b65d7651dc1082c01d96d2b1f9028b50d001ed0bc4ca8c1475d93"} Nov 21 14:41:57 crc kubenswrapper[4897]: I1121 14:41:57.210350 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-d8nth" Nov 21 14:41:57 crc kubenswrapper[4897]: I1121 14:41:57.983010 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-d8nth"] Nov 21 14:41:58 crc kubenswrapper[4897]: I1121 14:41:58.168636 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k27v5" event={"ID":"29e88709-d2ef-475a-85d2-1573f0719c8f","Type":"ContainerStarted","Data":"347a88ca18f62930f76018356446be838b94309ceb00131bdc44bb2afba978cf"} Nov 21 14:41:59 crc kubenswrapper[4897]: I1121 14:41:59.181065 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-d8nth" podUID="985eca1a-139e-46ce-a2a9-3e68d76b84d0" containerName="registry-server" containerID="cri-o://1ccf1e924cde0e2d7a1d25a5c7eb1b03623ad17b10a8a9291f26324d124b0f7d" gracePeriod=2 Nov 21 14:41:59 crc kubenswrapper[4897]: I1121 14:41:59.207346 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k27v5" podStartSLOduration=3.285479173 podStartE2EDuration="4.207325132s" podCreationTimestamp="2025-11-21 14:41:55 +0000 UTC" firstStartedPulling="2025-11-21 14:41:56.682097579 +0000 UTC m=+1993.966691054" lastFinishedPulling="2025-11-21 14:41:57.603943538 +0000 UTC m=+1994.888537013" observedRunningTime="2025-11-21 14:41:59.201322584 +0000 UTC m=+1996.485916069" watchObservedRunningTime="2025-11-21 14:41:59.207325132 +0000 UTC m=+1996.491918607" Nov 21 14:42:00 crc kubenswrapper[4897]: I1121 14:42:00.200528 4897 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-d8nth_985eca1a-139e-46ce-a2a9-3e68d76b84d0/registry-server/0.log" Nov 21 14:42:00 crc kubenswrapper[4897]: I1121 14:42:00.202687 4897 generic.go:334] "Generic (PLEG): container finished" podID="985eca1a-139e-46ce-a2a9-3e68d76b84d0" containerID="1ccf1e924cde0e2d7a1d25a5c7eb1b03623ad17b10a8a9291f26324d124b0f7d" exitCode=0 Nov 21 14:42:00 crc kubenswrapper[4897]: I1121 14:42:00.202733 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d8nth" event={"ID":"985eca1a-139e-46ce-a2a9-3e68d76b84d0","Type":"ContainerDied","Data":"1ccf1e924cde0e2d7a1d25a5c7eb1b03623ad17b10a8a9291f26324d124b0f7d"} Nov 21 14:42:00 crc kubenswrapper[4897]: I1121 14:42:00.202763 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d8nth" event={"ID":"985eca1a-139e-46ce-a2a9-3e68d76b84d0","Type":"ContainerDied","Data":"759a1e9e975e70246247f659c1fff33f80e45802f416023cd91a8765a8fde2b6"} Nov 21 14:42:00 crc kubenswrapper[4897]: I1121 14:42:00.202777 4897 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="759a1e9e975e70246247f659c1fff33f80e45802f416023cd91a8765a8fde2b6" Nov 21 14:42:00 crc kubenswrapper[4897]: I1121 14:42:00.202794 4897 scope.go:117] "RemoveContainer" containerID="d687cc3a224dedde2763d0f27045e20eb89735f70f16352147e95af90a6720c2" Nov 21 14:42:00 crc kubenswrapper[4897]: I1121 14:42:00.256770 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-d8nth" Nov 21 14:42:00 crc kubenswrapper[4897]: I1121 14:42:00.349792 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/985eca1a-139e-46ce-a2a9-3e68d76b84d0-catalog-content\") pod \"985eca1a-139e-46ce-a2a9-3e68d76b84d0\" (UID: \"985eca1a-139e-46ce-a2a9-3e68d76b84d0\") " Nov 21 14:42:00 crc kubenswrapper[4897]: I1121 14:42:00.349869 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/985eca1a-139e-46ce-a2a9-3e68d76b84d0-utilities\") pod \"985eca1a-139e-46ce-a2a9-3e68d76b84d0\" (UID: \"985eca1a-139e-46ce-a2a9-3e68d76b84d0\") " Nov 21 14:42:00 crc kubenswrapper[4897]: I1121 14:42:00.350190 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wqcw6\" (UniqueName: \"kubernetes.io/projected/985eca1a-139e-46ce-a2a9-3e68d76b84d0-kube-api-access-wqcw6\") pod \"985eca1a-139e-46ce-a2a9-3e68d76b84d0\" (UID: \"985eca1a-139e-46ce-a2a9-3e68d76b84d0\") " Nov 21 14:42:00 crc kubenswrapper[4897]: I1121 14:42:00.353707 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/985eca1a-139e-46ce-a2a9-3e68d76b84d0-utilities" (OuterVolumeSpecName: "utilities") pod "985eca1a-139e-46ce-a2a9-3e68d76b84d0" (UID: "985eca1a-139e-46ce-a2a9-3e68d76b84d0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:42:00 crc kubenswrapper[4897]: I1121 14:42:00.366915 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/985eca1a-139e-46ce-a2a9-3e68d76b84d0-kube-api-access-wqcw6" (OuterVolumeSpecName: "kube-api-access-wqcw6") pod "985eca1a-139e-46ce-a2a9-3e68d76b84d0" (UID: "985eca1a-139e-46ce-a2a9-3e68d76b84d0"). InnerVolumeSpecName "kube-api-access-wqcw6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:42:00 crc kubenswrapper[4897]: I1121 14:42:00.409690 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/985eca1a-139e-46ce-a2a9-3e68d76b84d0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "985eca1a-139e-46ce-a2a9-3e68d76b84d0" (UID: "985eca1a-139e-46ce-a2a9-3e68d76b84d0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:42:00 crc kubenswrapper[4897]: I1121 14:42:00.453305 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wqcw6\" (UniqueName: \"kubernetes.io/projected/985eca1a-139e-46ce-a2a9-3e68d76b84d0-kube-api-access-wqcw6\") on node \"crc\" DevicePath \"\"" Nov 21 14:42:00 crc kubenswrapper[4897]: I1121 14:42:00.453345 4897 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/985eca1a-139e-46ce-a2a9-3e68d76b84d0-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 21 14:42:00 crc kubenswrapper[4897]: I1121 14:42:00.453356 4897 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/985eca1a-139e-46ce-a2a9-3e68d76b84d0-utilities\") on node \"crc\" DevicePath \"\"" Nov 21 14:42:01 crc kubenswrapper[4897]: I1121 14:42:01.228576 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-d8nth" Nov 21 14:42:01 crc kubenswrapper[4897]: I1121 14:42:01.234924 4897 generic.go:334] "Generic (PLEG): container finished" podID="1d5dac4b-9e6c-41d5-9d10-87015da49a79" containerID="715e768e1b2e43e144c6d93c0e95b4d92949f9b90ed622f285ee9f286c09f298" exitCode=0 Nov 21 14:42:01 crc kubenswrapper[4897]: I1121 14:42:01.234970 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"1d5dac4b-9e6c-41d5-9d10-87015da49a79","Type":"ContainerDied","Data":"715e768e1b2e43e144c6d93c0e95b4d92949f9b90ed622f285ee9f286c09f298"} Nov 21 14:42:01 crc kubenswrapper[4897]: I1121 14:42:01.283844 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-d8nth"] Nov 21 14:42:01 crc kubenswrapper[4897]: I1121 14:42:01.293898 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-d8nth"] Nov 21 14:42:02 crc kubenswrapper[4897]: I1121 14:42:02.102866 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="985eca1a-139e-46ce-a2a9-3e68d76b84d0" path="/var/lib/kubelet/pods/985eca1a-139e-46ce-a2a9-3e68d76b84d0/volumes" Nov 21 14:42:02 crc kubenswrapper[4897]: I1121 14:42:02.260761 4897 generic.go:334] "Generic (PLEG): container finished" podID="1d5dac4b-9e6c-41d5-9d10-87015da49a79" containerID="88984a2b2c0a657d25b890e97a861f232ed8a051dddc4d79a706956451832fce" exitCode=0 Nov 21 14:42:02 crc kubenswrapper[4897]: I1121 14:42:02.260814 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"1d5dac4b-9e6c-41d5-9d10-87015da49a79","Type":"ContainerDied","Data":"88984a2b2c0a657d25b890e97a861f232ed8a051dddc4d79a706956451832fce"} Nov 21 14:42:02 crc kubenswrapper[4897]: I1121 14:42:02.260846 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"1d5dac4b-9e6c-41d5-9d10-87015da49a79","Type":"ContainerDied","Data":"d635a11d4cbbb9a1e3a5b53d856af24ad0a256da54c7d352f9f4735e521b150b"} Nov 21 14:42:02 crc kubenswrapper[4897]: I1121 14:42:02.260857 4897 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d635a11d4cbbb9a1e3a5b53d856af24ad0a256da54c7d352f9f4735e521b150b" Nov 21 14:42:02 crc kubenswrapper[4897]: I1121 14:42:02.266685 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Nov 21 14:42:02 crc kubenswrapper[4897]: I1121 14:42:02.394901 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d5dac4b-9e6c-41d5-9d10-87015da49a79-config-data\") pod \"1d5dac4b-9e6c-41d5-9d10-87015da49a79\" (UID: \"1d5dac4b-9e6c-41d5-9d10-87015da49a79\") " Nov 21 14:42:02 crc kubenswrapper[4897]: I1121 14:42:02.394959 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d5dac4b-9e6c-41d5-9d10-87015da49a79-scripts\") pod \"1d5dac4b-9e6c-41d5-9d10-87015da49a79\" (UID: \"1d5dac4b-9e6c-41d5-9d10-87015da49a79\") " Nov 21 14:42:02 crc kubenswrapper[4897]: I1121 14:42:02.395123 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6dmkw\" (UniqueName: \"kubernetes.io/projected/1d5dac4b-9e6c-41d5-9d10-87015da49a79-kube-api-access-6dmkw\") pod \"1d5dac4b-9e6c-41d5-9d10-87015da49a79\" (UID: \"1d5dac4b-9e6c-41d5-9d10-87015da49a79\") " Nov 21 14:42:02 crc kubenswrapper[4897]: I1121 14:42:02.395145 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d5dac4b-9e6c-41d5-9d10-87015da49a79-combined-ca-bundle\") pod \"1d5dac4b-9e6c-41d5-9d10-87015da49a79\" (UID: \"1d5dac4b-9e6c-41d5-9d10-87015da49a79\") " Nov 21 14:42:02 crc kubenswrapper[4897]: I1121 14:42:02.395291 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d5dac4b-9e6c-41d5-9d10-87015da49a79-public-tls-certs\") pod \"1d5dac4b-9e6c-41d5-9d10-87015da49a79\" (UID: \"1d5dac4b-9e6c-41d5-9d10-87015da49a79\") " Nov 21 14:42:02 crc kubenswrapper[4897]: I1121 14:42:02.395354 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d5dac4b-9e6c-41d5-9d10-87015da49a79-internal-tls-certs\") pod \"1d5dac4b-9e6c-41d5-9d10-87015da49a79\" (UID: \"1d5dac4b-9e6c-41d5-9d10-87015da49a79\") " Nov 21 14:42:02 crc kubenswrapper[4897]: I1121 14:42:02.406465 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d5dac4b-9e6c-41d5-9d10-87015da49a79-kube-api-access-6dmkw" (OuterVolumeSpecName: "kube-api-access-6dmkw") pod "1d5dac4b-9e6c-41d5-9d10-87015da49a79" (UID: "1d5dac4b-9e6c-41d5-9d10-87015da49a79"). InnerVolumeSpecName "kube-api-access-6dmkw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:42:02 crc kubenswrapper[4897]: I1121 14:42:02.408778 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d5dac4b-9e6c-41d5-9d10-87015da49a79-scripts" (OuterVolumeSpecName: "scripts") pod "1d5dac4b-9e6c-41d5-9d10-87015da49a79" (UID: "1d5dac4b-9e6c-41d5-9d10-87015da49a79"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:42:02 crc kubenswrapper[4897]: I1121 14:42:02.482154 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d5dac4b-9e6c-41d5-9d10-87015da49a79-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "1d5dac4b-9e6c-41d5-9d10-87015da49a79" (UID: "1d5dac4b-9e6c-41d5-9d10-87015da49a79"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:42:02 crc kubenswrapper[4897]: I1121 14:42:02.495871 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d5dac4b-9e6c-41d5-9d10-87015da49a79-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "1d5dac4b-9e6c-41d5-9d10-87015da49a79" (UID: "1d5dac4b-9e6c-41d5-9d10-87015da49a79"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:42:02 crc kubenswrapper[4897]: I1121 14:42:02.497910 4897 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d5dac4b-9e6c-41d5-9d10-87015da49a79-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 21 14:42:02 crc kubenswrapper[4897]: I1121 14:42:02.497939 4897 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d5dac4b-9e6c-41d5-9d10-87015da49a79-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 21 14:42:02 crc kubenswrapper[4897]: I1121 14:42:02.497952 4897 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d5dac4b-9e6c-41d5-9d10-87015da49a79-scripts\") on node \"crc\" DevicePath \"\"" Nov 21 14:42:02 crc kubenswrapper[4897]: I1121 14:42:02.497965 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6dmkw\" (UniqueName: \"kubernetes.io/projected/1d5dac4b-9e6c-41d5-9d10-87015da49a79-kube-api-access-6dmkw\") on node \"crc\" DevicePath \"\"" Nov 21 14:42:02 crc kubenswrapper[4897]: I1121 14:42:02.560347 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d5dac4b-9e6c-41d5-9d10-87015da49a79-config-data" (OuterVolumeSpecName: "config-data") pod "1d5dac4b-9e6c-41d5-9d10-87015da49a79" (UID: "1d5dac4b-9e6c-41d5-9d10-87015da49a79"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:42:02 crc kubenswrapper[4897]: I1121 14:42:02.578761 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d5dac4b-9e6c-41d5-9d10-87015da49a79-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1d5dac4b-9e6c-41d5-9d10-87015da49a79" (UID: "1d5dac4b-9e6c-41d5-9d10-87015da49a79"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:42:02 crc kubenswrapper[4897]: I1121 14:42:02.602038 4897 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d5dac4b-9e6c-41d5-9d10-87015da49a79-config-data\") on node \"crc\" DevicePath \"\"" Nov 21 14:42:02 crc kubenswrapper[4897]: I1121 14:42:02.602345 4897 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d5dac4b-9e6c-41d5-9d10-87015da49a79-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 21 14:42:03 crc kubenswrapper[4897]: I1121 14:42:03.271752 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Nov 21 14:42:03 crc kubenswrapper[4897]: I1121 14:42:03.350742 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Nov 21 14:42:03 crc kubenswrapper[4897]: I1121 14:42:03.368582 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-0"] Nov 21 14:42:03 crc kubenswrapper[4897]: I1121 14:42:03.379979 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Nov 21 14:42:03 crc kubenswrapper[4897]: E1121 14:42:03.380523 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="985eca1a-139e-46ce-a2a9-3e68d76b84d0" containerName="registry-server" Nov 21 14:42:03 crc kubenswrapper[4897]: I1121 14:42:03.380548 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="985eca1a-139e-46ce-a2a9-3e68d76b84d0" containerName="registry-server" Nov 21 14:42:03 crc kubenswrapper[4897]: E1121 14:42:03.380564 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="985eca1a-139e-46ce-a2a9-3e68d76b84d0" containerName="registry-server" Nov 21 14:42:03 crc kubenswrapper[4897]: I1121 14:42:03.380571 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="985eca1a-139e-46ce-a2a9-3e68d76b84d0" containerName="registry-server" Nov 21 14:42:03 crc kubenswrapper[4897]: E1121 14:42:03.380608 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d5dac4b-9e6c-41d5-9d10-87015da49a79" containerName="aodh-listener" Nov 21 14:42:03 crc kubenswrapper[4897]: I1121 14:42:03.380616 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d5dac4b-9e6c-41d5-9d10-87015da49a79" containerName="aodh-listener" Nov 21 14:42:03 crc kubenswrapper[4897]: E1121 14:42:03.380633 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="985eca1a-139e-46ce-a2a9-3e68d76b84d0" containerName="extract-content" Nov 21 14:42:03 crc kubenswrapper[4897]: I1121 14:42:03.380640 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="985eca1a-139e-46ce-a2a9-3e68d76b84d0" containerName="extract-content" Nov 21 14:42:03 crc kubenswrapper[4897]: E1121 14:42:03.380653 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d5dac4b-9e6c-41d5-9d10-87015da49a79" containerName="aodh-evaluator" Nov 21 14:42:03 crc kubenswrapper[4897]: I1121 14:42:03.380658 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d5dac4b-9e6c-41d5-9d10-87015da49a79" containerName="aodh-evaluator" Nov 21 14:42:03 crc kubenswrapper[4897]: E1121 14:42:03.380671 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d5dac4b-9e6c-41d5-9d10-87015da49a79" containerName="aodh-notifier" Nov 21 14:42:03 crc kubenswrapper[4897]: I1121 14:42:03.380678 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d5dac4b-9e6c-41d5-9d10-87015da49a79" containerName="aodh-notifier" Nov 21 14:42:03 crc kubenswrapper[4897]: E1121 14:42:03.380694 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="985eca1a-139e-46ce-a2a9-3e68d76b84d0" containerName="extract-utilities" Nov 21 14:42:03 crc kubenswrapper[4897]: I1121 14:42:03.380700 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="985eca1a-139e-46ce-a2a9-3e68d76b84d0" containerName="extract-utilities" Nov 21 14:42:03 crc kubenswrapper[4897]: E1121 14:42:03.380727 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d5dac4b-9e6c-41d5-9d10-87015da49a79" containerName="aodh-api" Nov 21 14:42:03 crc kubenswrapper[4897]: I1121 14:42:03.380733 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d5dac4b-9e6c-41d5-9d10-87015da49a79" containerName="aodh-api" Nov 21 14:42:03 crc kubenswrapper[4897]: I1121 14:42:03.380950 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d5dac4b-9e6c-41d5-9d10-87015da49a79" containerName="aodh-evaluator" Nov 21 14:42:03 crc kubenswrapper[4897]: I1121 14:42:03.380963 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="985eca1a-139e-46ce-a2a9-3e68d76b84d0" containerName="registry-server" Nov 21 14:42:03 crc kubenswrapper[4897]: I1121 14:42:03.380987 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d5dac4b-9e6c-41d5-9d10-87015da49a79" containerName="aodh-notifier" Nov 21 14:42:03 crc kubenswrapper[4897]: I1121 14:42:03.381001 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d5dac4b-9e6c-41d5-9d10-87015da49a79" containerName="aodh-listener" Nov 21 14:42:03 crc kubenswrapper[4897]: I1121 14:42:03.381017 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d5dac4b-9e6c-41d5-9d10-87015da49a79" containerName="aodh-api" Nov 21 14:42:03 crc kubenswrapper[4897]: I1121 14:42:03.381032 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="985eca1a-139e-46ce-a2a9-3e68d76b84d0" containerName="registry-server" Nov 21 14:42:03 crc kubenswrapper[4897]: I1121 14:42:03.383268 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Nov 21 14:42:03 crc kubenswrapper[4897]: I1121 14:42:03.390350 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-public-svc" Nov 21 14:42:03 crc kubenswrapper[4897]: I1121 14:42:03.390584 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Nov 21 14:42:03 crc kubenswrapper[4897]: I1121 14:42:03.390710 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-internal-svc" Nov 21 14:42:03 crc kubenswrapper[4897]: I1121 14:42:03.390895 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-m57c8" Nov 21 14:42:03 crc kubenswrapper[4897]: I1121 14:42:03.391001 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Nov 21 14:42:03 crc kubenswrapper[4897]: I1121 14:42:03.399157 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Nov 21 14:42:03 crc kubenswrapper[4897]: I1121 14:42:03.423962 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6263fa87-27b1-4c34-8165-61e41c29d92a-internal-tls-certs\") pod \"aodh-0\" (UID: \"6263fa87-27b1-4c34-8165-61e41c29d92a\") " pod="openstack/aodh-0" Nov 21 14:42:03 crc kubenswrapper[4897]: I1121 14:42:03.424028 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6263fa87-27b1-4c34-8165-61e41c29d92a-combined-ca-bundle\") pod \"aodh-0\" (UID: \"6263fa87-27b1-4c34-8165-61e41c29d92a\") " pod="openstack/aodh-0" Nov 21 14:42:03 crc kubenswrapper[4897]: I1121 14:42:03.424083 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jkbk4\" (UniqueName: \"kubernetes.io/projected/6263fa87-27b1-4c34-8165-61e41c29d92a-kube-api-access-jkbk4\") pod \"aodh-0\" (UID: \"6263fa87-27b1-4c34-8165-61e41c29d92a\") " pod="openstack/aodh-0" Nov 21 14:42:03 crc kubenswrapper[4897]: I1121 14:42:03.424147 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6263fa87-27b1-4c34-8165-61e41c29d92a-scripts\") pod \"aodh-0\" (UID: \"6263fa87-27b1-4c34-8165-61e41c29d92a\") " pod="openstack/aodh-0" Nov 21 14:42:03 crc kubenswrapper[4897]: I1121 14:42:03.424316 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6263fa87-27b1-4c34-8165-61e41c29d92a-public-tls-certs\") pod \"aodh-0\" (UID: \"6263fa87-27b1-4c34-8165-61e41c29d92a\") " pod="openstack/aodh-0" Nov 21 14:42:03 crc kubenswrapper[4897]: I1121 14:42:03.424350 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6263fa87-27b1-4c34-8165-61e41c29d92a-config-data\") pod \"aodh-0\" (UID: \"6263fa87-27b1-4c34-8165-61e41c29d92a\") " pod="openstack/aodh-0" Nov 21 14:42:03 crc kubenswrapper[4897]: I1121 14:42:03.526119 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6263fa87-27b1-4c34-8165-61e41c29d92a-combined-ca-bundle\") pod \"aodh-0\" (UID: \"6263fa87-27b1-4c34-8165-61e41c29d92a\") " pod="openstack/aodh-0" Nov 21 14:42:03 crc kubenswrapper[4897]: I1121 14:42:03.526206 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jkbk4\" (UniqueName: \"kubernetes.io/projected/6263fa87-27b1-4c34-8165-61e41c29d92a-kube-api-access-jkbk4\") pod \"aodh-0\" (UID: \"6263fa87-27b1-4c34-8165-61e41c29d92a\") " pod="openstack/aodh-0" Nov 21 14:42:03 crc kubenswrapper[4897]: I1121 14:42:03.526290 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6263fa87-27b1-4c34-8165-61e41c29d92a-scripts\") pod \"aodh-0\" (UID: \"6263fa87-27b1-4c34-8165-61e41c29d92a\") " pod="openstack/aodh-0" Nov 21 14:42:03 crc kubenswrapper[4897]: I1121 14:42:03.526558 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6263fa87-27b1-4c34-8165-61e41c29d92a-public-tls-certs\") pod \"aodh-0\" (UID: \"6263fa87-27b1-4c34-8165-61e41c29d92a\") " pod="openstack/aodh-0" Nov 21 14:42:03 crc kubenswrapper[4897]: I1121 14:42:03.526624 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6263fa87-27b1-4c34-8165-61e41c29d92a-config-data\") pod \"aodh-0\" (UID: \"6263fa87-27b1-4c34-8165-61e41c29d92a\") " pod="openstack/aodh-0" Nov 21 14:42:03 crc kubenswrapper[4897]: I1121 14:42:03.526683 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6263fa87-27b1-4c34-8165-61e41c29d92a-internal-tls-certs\") pod \"aodh-0\" (UID: \"6263fa87-27b1-4c34-8165-61e41c29d92a\") " pod="openstack/aodh-0" Nov 21 14:42:03 crc kubenswrapper[4897]: I1121 14:42:03.531556 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6263fa87-27b1-4c34-8165-61e41c29d92a-combined-ca-bundle\") pod \"aodh-0\" (UID: \"6263fa87-27b1-4c34-8165-61e41c29d92a\") " pod="openstack/aodh-0" Nov 21 14:42:03 crc kubenswrapper[4897]: I1121 14:42:03.531819 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6263fa87-27b1-4c34-8165-61e41c29d92a-scripts\") pod \"aodh-0\" (UID: \"6263fa87-27b1-4c34-8165-61e41c29d92a\") " pod="openstack/aodh-0" Nov 21 14:42:03 crc kubenswrapper[4897]: I1121 14:42:03.532702 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6263fa87-27b1-4c34-8165-61e41c29d92a-config-data\") pod \"aodh-0\" (UID: \"6263fa87-27b1-4c34-8165-61e41c29d92a\") " pod="openstack/aodh-0" Nov 21 14:42:03 crc kubenswrapper[4897]: I1121 14:42:03.532966 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6263fa87-27b1-4c34-8165-61e41c29d92a-public-tls-certs\") pod \"aodh-0\" (UID: \"6263fa87-27b1-4c34-8165-61e41c29d92a\") " pod="openstack/aodh-0" Nov 21 14:42:03 crc kubenswrapper[4897]: I1121 14:42:03.535299 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6263fa87-27b1-4c34-8165-61e41c29d92a-internal-tls-certs\") pod \"aodh-0\" (UID: \"6263fa87-27b1-4c34-8165-61e41c29d92a\") " pod="openstack/aodh-0" Nov 21 14:42:03 crc kubenswrapper[4897]: I1121 14:42:03.546945 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jkbk4\" (UniqueName: \"kubernetes.io/projected/6263fa87-27b1-4c34-8165-61e41c29d92a-kube-api-access-jkbk4\") pod \"aodh-0\" (UID: \"6263fa87-27b1-4c34-8165-61e41c29d92a\") " pod="openstack/aodh-0" Nov 21 14:42:03 crc kubenswrapper[4897]: I1121 14:42:03.702965 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Nov 21 14:42:04 crc kubenswrapper[4897]: I1121 14:42:04.108632 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d5dac4b-9e6c-41d5-9d10-87015da49a79" path="/var/lib/kubelet/pods/1d5dac4b-9e6c-41d5-9d10-87015da49a79/volumes" Nov 21 14:42:04 crc kubenswrapper[4897]: I1121 14:42:04.330330 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Nov 21 14:42:05 crc kubenswrapper[4897]: I1121 14:42:05.296737 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"6263fa87-27b1-4c34-8165-61e41c29d92a","Type":"ContainerStarted","Data":"2a53fb6fdf41af8f396fcf32bdc1df6ce6a426036870dbd1850875d6e2f7b665"} Nov 21 14:42:07 crc kubenswrapper[4897]: I1121 14:42:07.321196 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"6263fa87-27b1-4c34-8165-61e41c29d92a","Type":"ContainerStarted","Data":"f7fe1e6efdd9f22ddd168e2fec8181ac15293bd057e12f14aadb6cc63942f44a"} Nov 21 14:42:08 crc kubenswrapper[4897]: I1121 14:42:08.049837 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-dm94g"] Nov 21 14:42:08 crc kubenswrapper[4897]: I1121 14:42:08.072445 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-dm94g"] Nov 21 14:42:08 crc kubenswrapper[4897]: I1121 14:42:08.105789 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="de087e28-b2d8-4686-b758-f4cd38443f25" path="/var/lib/kubelet/pods/de087e28-b2d8-4686-b758-f4cd38443f25/volumes" Nov 21 14:42:09 crc kubenswrapper[4897]: I1121 14:42:09.033932 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-bgs55"] Nov 21 14:42:09 crc kubenswrapper[4897]: I1121 14:42:09.043969 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-xj48w"] Nov 21 14:42:09 crc kubenswrapper[4897]: I1121 14:42:09.057677 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-bgs55"] Nov 21 14:42:09 crc kubenswrapper[4897]: I1121 14:42:09.069080 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-xj48w"] Nov 21 14:42:09 crc kubenswrapper[4897]: I1121 14:42:09.353120 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"6263fa87-27b1-4c34-8165-61e41c29d92a","Type":"ContainerStarted","Data":"edf2ef27c77fecc1120a48f7485e97e7c596d4dc8fe776bf1cc0a9381b9bddf2"} Nov 21 14:42:10 crc kubenswrapper[4897]: I1121 14:42:10.038685 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-zz6kx"] Nov 21 14:42:10 crc kubenswrapper[4897]: I1121 14:42:10.059417 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-zz6kx"] Nov 21 14:42:10 crc kubenswrapper[4897]: I1121 14:42:10.109191 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7f7487cb-deab-4a72-ad73-5d805b4e25cd" path="/var/lib/kubelet/pods/7f7487cb-deab-4a72-ad73-5d805b4e25cd/volumes" Nov 21 14:42:10 crc kubenswrapper[4897]: I1121 14:42:10.113142 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c08d8b3-d052-47b7-ad42-7b0f9fbe1920" path="/var/lib/kubelet/pods/9c08d8b3-d052-47b7-ad42-7b0f9fbe1920/volumes" Nov 21 14:42:10 crc kubenswrapper[4897]: I1121 14:42:10.114352 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b0c8cf90-d5b4-43b0-9037-4b579f26c74b" path="/var/lib/kubelet/pods/b0c8cf90-d5b4-43b0-9037-4b579f26c74b/volumes" Nov 21 14:42:10 crc kubenswrapper[4897]: I1121 14:42:10.367237 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"6263fa87-27b1-4c34-8165-61e41c29d92a","Type":"ContainerStarted","Data":"720c48bf4023fdac3cde420c2b406d2058182caa0611e94aec479fba883a4399"} Nov 21 14:42:13 crc kubenswrapper[4897]: I1121 14:42:13.401735 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"6263fa87-27b1-4c34-8165-61e41c29d92a","Type":"ContainerStarted","Data":"8b2be56641465a2d2badea84521add3129afe5f7d81e50cf884d3cc7c65b5b06"} Nov 21 14:42:13 crc kubenswrapper[4897]: I1121 14:42:13.471329 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=2.469405111 podStartE2EDuration="10.471303034s" podCreationTimestamp="2025-11-21 14:42:03 +0000 UTC" firstStartedPulling="2025-11-21 14:42:04.336130378 +0000 UTC m=+2001.620723843" lastFinishedPulling="2025-11-21 14:42:12.338028291 +0000 UTC m=+2009.622621766" observedRunningTime="2025-11-21 14:42:13.425807712 +0000 UTC m=+2010.710401187" watchObservedRunningTime="2025-11-21 14:42:13.471303034 +0000 UTC m=+2010.755896509" Nov 21 14:42:19 crc kubenswrapper[4897]: I1121 14:42:19.342478 4897 scope.go:117] "RemoveContainer" containerID="9d873a09028f196e71827f91a3a2045d109b5c5614da0e1247c853f713bf9839" Nov 21 14:42:19 crc kubenswrapper[4897]: I1121 14:42:19.372919 4897 scope.go:117] "RemoveContainer" containerID="0b3e7824d93c1825d32ebe2895717aed9691091cdb0492f9fc2fed0b92d32dd9" Nov 21 14:42:19 crc kubenswrapper[4897]: I1121 14:42:19.436657 4897 scope.go:117] "RemoveContainer" containerID="715e768e1b2e43e144c6d93c0e95b4d92949f9b90ed622f285ee9f286c09f298" Nov 21 14:42:19 crc kubenswrapper[4897]: I1121 14:42:19.474740 4897 scope.go:117] "RemoveContainer" containerID="d2be4b278d182e43defddab52a77b6052ea8408529ad229d4d7dbe3734a7b4c4" Nov 21 14:42:19 crc kubenswrapper[4897]: I1121 14:42:19.502027 4897 scope.go:117] "RemoveContainer" containerID="a4982453c7f834bedde18c2a269072073f1527ae983e08fe842d8d47389611db" Nov 21 14:42:19 crc kubenswrapper[4897]: I1121 14:42:19.539064 4897 scope.go:117] "RemoveContainer" containerID="88984a2b2c0a657d25b890e97a861f232ed8a051dddc4d79a706956451832fce" Nov 21 14:42:19 crc kubenswrapper[4897]: I1121 14:42:19.577684 4897 scope.go:117] "RemoveContainer" containerID="442b3e7ff2a0cf9fd50378ea8e061b6f8a89f463d1baacbad2a4188c2f931c88" Nov 21 14:42:19 crc kubenswrapper[4897]: I1121 14:42:19.690849 4897 scope.go:117] "RemoveContainer" containerID="e65e756fbb3117c206622407abfc9990d581b756bb24310f36d0c24d49164dc0" Nov 21 14:42:19 crc kubenswrapper[4897]: I1121 14:42:19.727349 4897 scope.go:117] "RemoveContainer" containerID="f494f8ffa1945123f3045599a6855553d8ce35e55d6ce7171e3c0d5f9ed36b69" Nov 21 14:42:19 crc kubenswrapper[4897]: I1121 14:42:19.772143 4897 scope.go:117] "RemoveContainer" containerID="f0b5d22717e8bf9ca37f66f5b7cbefcb84e12420967595ff9e141eebb59aaa33" Nov 21 14:42:34 crc kubenswrapper[4897]: I1121 14:42:34.043639 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-7dvt9"] Nov 21 14:42:34 crc kubenswrapper[4897]: I1121 14:42:34.055332 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-7dvt9"] Nov 21 14:42:34 crc kubenswrapper[4897]: I1121 14:42:34.109864 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="55644d8b-b8a3-463e-8045-b7391e9c2419" path="/var/lib/kubelet/pods/55644d8b-b8a3-463e-8045-b7391e9c2419/volumes" Nov 21 14:42:34 crc kubenswrapper[4897]: I1121 14:42:34.370439 4897 patch_prober.go:28] interesting pod/machine-config-daemon-krv5b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 21 14:42:34 crc kubenswrapper[4897]: I1121 14:42:34.370523 4897 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 21 14:42:50 crc kubenswrapper[4897]: I1121 14:42:50.051218 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-mrctr"] Nov 21 14:42:50 crc kubenswrapper[4897]: I1121 14:42:50.063315 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-mrctr"] Nov 21 14:42:50 crc kubenswrapper[4897]: I1121 14:42:50.104771 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4a683589-a940-405b-afe1-1838f34dac9a" path="/var/lib/kubelet/pods/4a683589-a940-405b-afe1-1838f34dac9a/volumes" Nov 21 14:43:04 crc kubenswrapper[4897]: I1121 14:43:04.371093 4897 patch_prober.go:28] interesting pod/machine-config-daemon-krv5b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 21 14:43:04 crc kubenswrapper[4897]: I1121 14:43:04.371976 4897 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 21 14:43:20 crc kubenswrapper[4897]: I1121 14:43:20.048173 4897 scope.go:117] "RemoveContainer" containerID="b3b649aa1cd6febd3c35ff713d0112d349fcbe8d37fb13176de414dfeca6b26c" Nov 21 14:43:20 crc kubenswrapper[4897]: I1121 14:43:20.101872 4897 scope.go:117] "RemoveContainer" containerID="e7ac0248e2b28f323c61ff8d85c2fd0aaf30199090d2984738889239f670d176" Nov 21 14:43:34 crc kubenswrapper[4897]: I1121 14:43:34.370272 4897 patch_prober.go:28] interesting pod/machine-config-daemon-krv5b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 21 14:43:34 crc kubenswrapper[4897]: I1121 14:43:34.370906 4897 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 21 14:43:34 crc kubenswrapper[4897]: I1121 14:43:34.370957 4897 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" Nov 21 14:43:34 crc kubenswrapper[4897]: I1121 14:43:34.371806 4897 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a4d7f2b343d4dace3a7a2ec8aeabc1074996aaa7dc843ecf94e43c56479ac344"} pod="openshift-machine-config-operator/machine-config-daemon-krv5b" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 21 14:43:34 crc kubenswrapper[4897]: I1121 14:43:34.371858 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" containerName="machine-config-daemon" containerID="cri-o://a4d7f2b343d4dace3a7a2ec8aeabc1074996aaa7dc843ecf94e43c56479ac344" gracePeriod=600 Nov 21 14:43:35 crc kubenswrapper[4897]: I1121 14:43:35.406151 4897 generic.go:334] "Generic (PLEG): container finished" podID="e7670227-d280-4847-b882-754429f56b0f" containerID="a4d7f2b343d4dace3a7a2ec8aeabc1074996aaa7dc843ecf94e43c56479ac344" exitCode=0 Nov 21 14:43:35 crc kubenswrapper[4897]: I1121 14:43:35.406225 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" event={"ID":"e7670227-d280-4847-b882-754429f56b0f","Type":"ContainerDied","Data":"a4d7f2b343d4dace3a7a2ec8aeabc1074996aaa7dc843ecf94e43c56479ac344"} Nov 21 14:43:35 crc kubenswrapper[4897]: I1121 14:43:35.406849 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" event={"ID":"e7670227-d280-4847-b882-754429f56b0f","Type":"ContainerStarted","Data":"dc3c0a9877f70b7d5c55314a92df879b827a4103351566cc755cfbf916773aa2"} Nov 21 14:43:35 crc kubenswrapper[4897]: I1121 14:43:35.406888 4897 scope.go:117] "RemoveContainer" containerID="c68f1b003d12b4a2964169a85a04e3f0272d86c891f52e6bebd52f3781520d1a" Nov 21 14:43:54 crc kubenswrapper[4897]: I1121 14:43:54.058488 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-7qv9b"] Nov 21 14:43:54 crc kubenswrapper[4897]: I1121 14:43:54.070724 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-7qv9b"] Nov 21 14:43:54 crc kubenswrapper[4897]: I1121 14:43:54.118957 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0d0c865b-9c9b-49cf-ba2f-730f98adce04" path="/var/lib/kubelet/pods/0d0c865b-9c9b-49cf-ba2f-730f98adce04/volumes" Nov 21 14:43:56 crc kubenswrapper[4897]: I1121 14:43:56.029750 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-5e1d-account-create-wkbm2"] Nov 21 14:43:56 crc kubenswrapper[4897]: I1121 14:43:56.042358 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-5e1d-account-create-wkbm2"] Nov 21 14:43:56 crc kubenswrapper[4897]: I1121 14:43:56.116097 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="548045ab-1444-4b7a-a663-3fa90b7fdb22" path="/var/lib/kubelet/pods/548045ab-1444-4b7a-a663-3fa90b7fdb22/volumes" Nov 21 14:43:58 crc kubenswrapper[4897]: I1121 14:43:58.050130 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-1c5d-account-create-lx2qd"] Nov 21 14:43:58 crc kubenswrapper[4897]: I1121 14:43:58.063182 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-vs69w"] Nov 21 14:43:58 crc kubenswrapper[4897]: I1121 14:43:58.075992 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-5f35-account-create-v8krn"] Nov 21 14:43:58 crc kubenswrapper[4897]: I1121 14:43:58.086961 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-vs69w"] Nov 21 14:43:58 crc kubenswrapper[4897]: I1121 14:43:58.104099 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6f516d79-7775-40cc-9d50-4167d9b1afbb" path="/var/lib/kubelet/pods/6f516d79-7775-40cc-9d50-4167d9b1afbb/volumes" Nov 21 14:43:58 crc kubenswrapper[4897]: I1121 14:43:58.104815 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-5f35-account-create-v8krn"] Nov 21 14:43:58 crc kubenswrapper[4897]: I1121 14:43:58.104850 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-1c5d-account-create-lx2qd"] Nov 21 14:43:58 crc kubenswrapper[4897]: I1121 14:43:58.116046 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-v4ws9"] Nov 21 14:43:58 crc kubenswrapper[4897]: I1121 14:43:58.128457 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-v4ws9"] Nov 21 14:44:00 crc kubenswrapper[4897]: I1121 14:44:00.105580 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="56d6af89-a8f0-4b64-9426-04536dedf90c" path="/var/lib/kubelet/pods/56d6af89-a8f0-4b64-9426-04536dedf90c/volumes" Nov 21 14:44:00 crc kubenswrapper[4897]: I1121 14:44:00.107730 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9b100fec-a276-44ac-b14d-febed2079580" path="/var/lib/kubelet/pods/9b100fec-a276-44ac-b14d-febed2079580/volumes" Nov 21 14:44:00 crc kubenswrapper[4897]: I1121 14:44:00.108385 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f5440e5f-8e45-4348-ba61-e9d3e56832c6" path="/var/lib/kubelet/pods/f5440e5f-8e45-4348-ba61-e9d3e56832c6/volumes" Nov 21 14:44:02 crc kubenswrapper[4897]: I1121 14:44:02.248737 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-pgq9f"] Nov 21 14:44:02 crc kubenswrapper[4897]: I1121 14:44:02.254826 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pgq9f" Nov 21 14:44:02 crc kubenswrapper[4897]: I1121 14:44:02.272408 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-pgq9f"] Nov 21 14:44:02 crc kubenswrapper[4897]: I1121 14:44:02.398862 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a62940fe-e1a9-42c5-b524-c269aa42d1a5-utilities\") pod \"certified-operators-pgq9f\" (UID: \"a62940fe-e1a9-42c5-b524-c269aa42d1a5\") " pod="openshift-marketplace/certified-operators-pgq9f" Nov 21 14:44:02 crc kubenswrapper[4897]: I1121 14:44:02.398935 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a62940fe-e1a9-42c5-b524-c269aa42d1a5-catalog-content\") pod \"certified-operators-pgq9f\" (UID: \"a62940fe-e1a9-42c5-b524-c269aa42d1a5\") " pod="openshift-marketplace/certified-operators-pgq9f" Nov 21 14:44:02 crc kubenswrapper[4897]: I1121 14:44:02.399628 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bs2mb\" (UniqueName: \"kubernetes.io/projected/a62940fe-e1a9-42c5-b524-c269aa42d1a5-kube-api-access-bs2mb\") pod \"certified-operators-pgq9f\" (UID: \"a62940fe-e1a9-42c5-b524-c269aa42d1a5\") " pod="openshift-marketplace/certified-operators-pgq9f" Nov 21 14:44:02 crc kubenswrapper[4897]: I1121 14:44:02.501999 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a62940fe-e1a9-42c5-b524-c269aa42d1a5-utilities\") pod \"certified-operators-pgq9f\" (UID: \"a62940fe-e1a9-42c5-b524-c269aa42d1a5\") " pod="openshift-marketplace/certified-operators-pgq9f" Nov 21 14:44:02 crc kubenswrapper[4897]: I1121 14:44:02.502058 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a62940fe-e1a9-42c5-b524-c269aa42d1a5-catalog-content\") pod \"certified-operators-pgq9f\" (UID: \"a62940fe-e1a9-42c5-b524-c269aa42d1a5\") " pod="openshift-marketplace/certified-operators-pgq9f" Nov 21 14:44:02 crc kubenswrapper[4897]: I1121 14:44:02.502225 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bs2mb\" (UniqueName: \"kubernetes.io/projected/a62940fe-e1a9-42c5-b524-c269aa42d1a5-kube-api-access-bs2mb\") pod \"certified-operators-pgq9f\" (UID: \"a62940fe-e1a9-42c5-b524-c269aa42d1a5\") " pod="openshift-marketplace/certified-operators-pgq9f" Nov 21 14:44:02 crc kubenswrapper[4897]: I1121 14:44:02.503041 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a62940fe-e1a9-42c5-b524-c269aa42d1a5-utilities\") pod \"certified-operators-pgq9f\" (UID: \"a62940fe-e1a9-42c5-b524-c269aa42d1a5\") " pod="openshift-marketplace/certified-operators-pgq9f" Nov 21 14:44:02 crc kubenswrapper[4897]: I1121 14:44:02.503251 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a62940fe-e1a9-42c5-b524-c269aa42d1a5-catalog-content\") pod \"certified-operators-pgq9f\" (UID: \"a62940fe-e1a9-42c5-b524-c269aa42d1a5\") " pod="openshift-marketplace/certified-operators-pgq9f" Nov 21 14:44:02 crc kubenswrapper[4897]: I1121 14:44:02.537361 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bs2mb\" (UniqueName: \"kubernetes.io/projected/a62940fe-e1a9-42c5-b524-c269aa42d1a5-kube-api-access-bs2mb\") pod \"certified-operators-pgq9f\" (UID: \"a62940fe-e1a9-42c5-b524-c269aa42d1a5\") " pod="openshift-marketplace/certified-operators-pgq9f" Nov 21 14:44:02 crc kubenswrapper[4897]: I1121 14:44:02.588799 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pgq9f" Nov 21 14:44:03 crc kubenswrapper[4897]: I1121 14:44:03.144797 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-pgq9f"] Nov 21 14:44:03 crc kubenswrapper[4897]: I1121 14:44:03.746904 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pgq9f" event={"ID":"a62940fe-e1a9-42c5-b524-c269aa42d1a5","Type":"ContainerStarted","Data":"d91ed7adff5b9773b78de6eefc30522022213fff6248f725a1896ecf2df76c0b"} Nov 21 14:44:04 crc kubenswrapper[4897]: I1121 14:44:04.760530 4897 generic.go:334] "Generic (PLEG): container finished" podID="a62940fe-e1a9-42c5-b524-c269aa42d1a5" containerID="aced14f00c6d363cc94a4666b24fe09255b379d9023fd9f2ccbc0ea4ff3b1abb" exitCode=0 Nov 21 14:44:04 crc kubenswrapper[4897]: I1121 14:44:04.760579 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pgq9f" event={"ID":"a62940fe-e1a9-42c5-b524-c269aa42d1a5","Type":"ContainerDied","Data":"aced14f00c6d363cc94a4666b24fe09255b379d9023fd9f2ccbc0ea4ff3b1abb"} Nov 21 14:44:08 crc kubenswrapper[4897]: I1121 14:44:08.810315 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pgq9f" event={"ID":"a62940fe-e1a9-42c5-b524-c269aa42d1a5","Type":"ContainerStarted","Data":"7be51a0b56a85e75f8bc95d8dc537be42b24d1260ca44b247b5f2e8fe9542ac0"} Nov 21 14:44:19 crc kubenswrapper[4897]: I1121 14:44:19.925619 4897 generic.go:334] "Generic (PLEG): container finished" podID="a62940fe-e1a9-42c5-b524-c269aa42d1a5" containerID="7be51a0b56a85e75f8bc95d8dc537be42b24d1260ca44b247b5f2e8fe9542ac0" exitCode=0 Nov 21 14:44:19 crc kubenswrapper[4897]: I1121 14:44:19.925690 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pgq9f" event={"ID":"a62940fe-e1a9-42c5-b524-c269aa42d1a5","Type":"ContainerDied","Data":"7be51a0b56a85e75f8bc95d8dc537be42b24d1260ca44b247b5f2e8fe9542ac0"} Nov 21 14:44:20 crc kubenswrapper[4897]: I1121 14:44:20.222379 4897 scope.go:117] "RemoveContainer" containerID="f5690c0c986bc84e69fdd7b3c70b9bd8e55ba08bd1c98561ae2c687488b4cd32" Nov 21 14:44:20 crc kubenswrapper[4897]: I1121 14:44:20.328247 4897 scope.go:117] "RemoveContainer" containerID="08ac180150001e919e507fa3ef316521456df000f2841292a09dafc0dc32f60f" Nov 21 14:44:20 crc kubenswrapper[4897]: I1121 14:44:20.361546 4897 scope.go:117] "RemoveContainer" containerID="c0b618bf5ae4ce49aef42bd5f9fd1db2a9749f5add75cb36a9b75ad9708b75ca" Nov 21 14:44:20 crc kubenswrapper[4897]: I1121 14:44:20.420734 4897 scope.go:117] "RemoveContainer" containerID="92f1d630304a62d838c8406625e9d7c47364f3e598d34992f5a62a5858181229" Nov 21 14:44:20 crc kubenswrapper[4897]: I1121 14:44:20.516095 4897 scope.go:117] "RemoveContainer" containerID="ff653de54f2f14431bfbdca9aaceb1a64697a7c40805a7f359cb44ee5c297a09" Nov 21 14:44:20 crc kubenswrapper[4897]: I1121 14:44:20.589209 4897 scope.go:117] "RemoveContainer" containerID="98c4e1657f5fe0d5be7978356a94be35f57ab572a6d8aaf9109027d63a7df60a" Nov 21 14:44:28 crc kubenswrapper[4897]: I1121 14:44:28.015340 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pgq9f" event={"ID":"a62940fe-e1a9-42c5-b524-c269aa42d1a5","Type":"ContainerStarted","Data":"d6a1b86964b7ccfd3c92c59c8407e4a4fadfa153be512b40644482084d69e135"} Nov 21 14:44:32 crc kubenswrapper[4897]: I1121 14:44:32.589660 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-pgq9f" Nov 21 14:44:32 crc kubenswrapper[4897]: I1121 14:44:32.590329 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-pgq9f" Nov 21 14:44:32 crc kubenswrapper[4897]: I1121 14:44:32.654464 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-pgq9f" Nov 21 14:44:32 crc kubenswrapper[4897]: I1121 14:44:32.684602 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-pgq9f" podStartSLOduration=11.34928481 podStartE2EDuration="30.684580436s" podCreationTimestamp="2025-11-21 14:44:02 +0000 UTC" firstStartedPulling="2025-11-21 14:44:04.763423348 +0000 UTC m=+2122.048016833" lastFinishedPulling="2025-11-21 14:44:24.098718964 +0000 UTC m=+2141.383312459" observedRunningTime="2025-11-21 14:44:30.05953368 +0000 UTC m=+2147.344127165" watchObservedRunningTime="2025-11-21 14:44:32.684580436 +0000 UTC m=+2149.969173921" Nov 21 14:44:33 crc kubenswrapper[4897]: I1121 14:44:33.135160 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-pgq9f" Nov 21 14:44:33 crc kubenswrapper[4897]: I1121 14:44:33.449997 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-pgq9f"] Nov 21 14:44:35 crc kubenswrapper[4897]: I1121 14:44:35.115788 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-pgq9f" podUID="a62940fe-e1a9-42c5-b524-c269aa42d1a5" containerName="registry-server" containerID="cri-o://d6a1b86964b7ccfd3c92c59c8407e4a4fadfa153be512b40644482084d69e135" gracePeriod=2 Nov 21 14:44:35 crc kubenswrapper[4897]: I1121 14:44:35.703079 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pgq9f" Nov 21 14:44:35 crc kubenswrapper[4897]: I1121 14:44:35.796355 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a62940fe-e1a9-42c5-b524-c269aa42d1a5-utilities\") pod \"a62940fe-e1a9-42c5-b524-c269aa42d1a5\" (UID: \"a62940fe-e1a9-42c5-b524-c269aa42d1a5\") " Nov 21 14:44:35 crc kubenswrapper[4897]: I1121 14:44:35.796442 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bs2mb\" (UniqueName: \"kubernetes.io/projected/a62940fe-e1a9-42c5-b524-c269aa42d1a5-kube-api-access-bs2mb\") pod \"a62940fe-e1a9-42c5-b524-c269aa42d1a5\" (UID: \"a62940fe-e1a9-42c5-b524-c269aa42d1a5\") " Nov 21 14:44:35 crc kubenswrapper[4897]: I1121 14:44:35.796521 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a62940fe-e1a9-42c5-b524-c269aa42d1a5-catalog-content\") pod \"a62940fe-e1a9-42c5-b524-c269aa42d1a5\" (UID: \"a62940fe-e1a9-42c5-b524-c269aa42d1a5\") " Nov 21 14:44:35 crc kubenswrapper[4897]: I1121 14:44:35.797013 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a62940fe-e1a9-42c5-b524-c269aa42d1a5-utilities" (OuterVolumeSpecName: "utilities") pod "a62940fe-e1a9-42c5-b524-c269aa42d1a5" (UID: "a62940fe-e1a9-42c5-b524-c269aa42d1a5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:44:35 crc kubenswrapper[4897]: I1121 14:44:35.801876 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a62940fe-e1a9-42c5-b524-c269aa42d1a5-kube-api-access-bs2mb" (OuterVolumeSpecName: "kube-api-access-bs2mb") pod "a62940fe-e1a9-42c5-b524-c269aa42d1a5" (UID: "a62940fe-e1a9-42c5-b524-c269aa42d1a5"). InnerVolumeSpecName "kube-api-access-bs2mb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:44:35 crc kubenswrapper[4897]: I1121 14:44:35.840307 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a62940fe-e1a9-42c5-b524-c269aa42d1a5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a62940fe-e1a9-42c5-b524-c269aa42d1a5" (UID: "a62940fe-e1a9-42c5-b524-c269aa42d1a5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:44:35 crc kubenswrapper[4897]: I1121 14:44:35.898699 4897 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a62940fe-e1a9-42c5-b524-c269aa42d1a5-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 21 14:44:35 crc kubenswrapper[4897]: I1121 14:44:35.898749 4897 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a62940fe-e1a9-42c5-b524-c269aa42d1a5-utilities\") on node \"crc\" DevicePath \"\"" Nov 21 14:44:35 crc kubenswrapper[4897]: I1121 14:44:35.898769 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bs2mb\" (UniqueName: \"kubernetes.io/projected/a62940fe-e1a9-42c5-b524-c269aa42d1a5-kube-api-access-bs2mb\") on node \"crc\" DevicePath \"\"" Nov 21 14:44:36 crc kubenswrapper[4897]: I1121 14:44:36.054565 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0f93-account-create-54bcj"] Nov 21 14:44:36 crc kubenswrapper[4897]: I1121 14:44:36.066240 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-create-jjdjs"] Nov 21 14:44:36 crc kubenswrapper[4897]: I1121 14:44:36.075180 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-0f93-account-create-54bcj"] Nov 21 14:44:36 crc kubenswrapper[4897]: I1121 14:44:36.083545 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-create-jjdjs"] Nov 21 14:44:36 crc kubenswrapper[4897]: I1121 14:44:36.109923 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="407be862-21b7-4f8d-9e62-81e6e8e40ac5" path="/var/lib/kubelet/pods/407be862-21b7-4f8d-9e62-81e6e8e40ac5/volumes" Nov 21 14:44:36 crc kubenswrapper[4897]: I1121 14:44:36.111034 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f2251225-86fd-4e3e-87fe-d4ac1db2ee8e" path="/var/lib/kubelet/pods/f2251225-86fd-4e3e-87fe-d4ac1db2ee8e/volumes" Nov 21 14:44:36 crc kubenswrapper[4897]: I1121 14:44:36.128821 4897 generic.go:334] "Generic (PLEG): container finished" podID="a62940fe-e1a9-42c5-b524-c269aa42d1a5" containerID="d6a1b86964b7ccfd3c92c59c8407e4a4fadfa153be512b40644482084d69e135" exitCode=0 Nov 21 14:44:36 crc kubenswrapper[4897]: I1121 14:44:36.128873 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pgq9f" event={"ID":"a62940fe-e1a9-42c5-b524-c269aa42d1a5","Type":"ContainerDied","Data":"d6a1b86964b7ccfd3c92c59c8407e4a4fadfa153be512b40644482084d69e135"} Nov 21 14:44:36 crc kubenswrapper[4897]: I1121 14:44:36.128907 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pgq9f" event={"ID":"a62940fe-e1a9-42c5-b524-c269aa42d1a5","Type":"ContainerDied","Data":"d91ed7adff5b9773b78de6eefc30522022213fff6248f725a1896ecf2df76c0b"} Nov 21 14:44:36 crc kubenswrapper[4897]: I1121 14:44:36.128929 4897 scope.go:117] "RemoveContainer" containerID="d6a1b86964b7ccfd3c92c59c8407e4a4fadfa153be512b40644482084d69e135" Nov 21 14:44:36 crc kubenswrapper[4897]: I1121 14:44:36.129634 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pgq9f" Nov 21 14:44:36 crc kubenswrapper[4897]: I1121 14:44:36.155597 4897 scope.go:117] "RemoveContainer" containerID="7be51a0b56a85e75f8bc95d8dc537be42b24d1260ca44b247b5f2e8fe9542ac0" Nov 21 14:44:36 crc kubenswrapper[4897]: I1121 14:44:36.164640 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-pgq9f"] Nov 21 14:44:36 crc kubenswrapper[4897]: I1121 14:44:36.175167 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-pgq9f"] Nov 21 14:44:36 crc kubenswrapper[4897]: I1121 14:44:36.178879 4897 scope.go:117] "RemoveContainer" containerID="aced14f00c6d363cc94a4666b24fe09255b379d9023fd9f2ccbc0ea4ff3b1abb" Nov 21 14:44:36 crc kubenswrapper[4897]: I1121 14:44:36.233470 4897 scope.go:117] "RemoveContainer" containerID="d6a1b86964b7ccfd3c92c59c8407e4a4fadfa153be512b40644482084d69e135" Nov 21 14:44:36 crc kubenswrapper[4897]: E1121 14:44:36.233960 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d6a1b86964b7ccfd3c92c59c8407e4a4fadfa153be512b40644482084d69e135\": container with ID starting with d6a1b86964b7ccfd3c92c59c8407e4a4fadfa153be512b40644482084d69e135 not found: ID does not exist" containerID="d6a1b86964b7ccfd3c92c59c8407e4a4fadfa153be512b40644482084d69e135" Nov 21 14:44:36 crc kubenswrapper[4897]: I1121 14:44:36.234078 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d6a1b86964b7ccfd3c92c59c8407e4a4fadfa153be512b40644482084d69e135"} err="failed to get container status \"d6a1b86964b7ccfd3c92c59c8407e4a4fadfa153be512b40644482084d69e135\": rpc error: code = NotFound desc = could not find container \"d6a1b86964b7ccfd3c92c59c8407e4a4fadfa153be512b40644482084d69e135\": container with ID starting with d6a1b86964b7ccfd3c92c59c8407e4a4fadfa153be512b40644482084d69e135 not found: ID does not exist" Nov 21 14:44:36 crc kubenswrapper[4897]: I1121 14:44:36.234161 4897 scope.go:117] "RemoveContainer" containerID="7be51a0b56a85e75f8bc95d8dc537be42b24d1260ca44b247b5f2e8fe9542ac0" Nov 21 14:44:36 crc kubenswrapper[4897]: E1121 14:44:36.234487 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7be51a0b56a85e75f8bc95d8dc537be42b24d1260ca44b247b5f2e8fe9542ac0\": container with ID starting with 7be51a0b56a85e75f8bc95d8dc537be42b24d1260ca44b247b5f2e8fe9542ac0 not found: ID does not exist" containerID="7be51a0b56a85e75f8bc95d8dc537be42b24d1260ca44b247b5f2e8fe9542ac0" Nov 21 14:44:36 crc kubenswrapper[4897]: I1121 14:44:36.234594 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7be51a0b56a85e75f8bc95d8dc537be42b24d1260ca44b247b5f2e8fe9542ac0"} err="failed to get container status \"7be51a0b56a85e75f8bc95d8dc537be42b24d1260ca44b247b5f2e8fe9542ac0\": rpc error: code = NotFound desc = could not find container \"7be51a0b56a85e75f8bc95d8dc537be42b24d1260ca44b247b5f2e8fe9542ac0\": container with ID starting with 7be51a0b56a85e75f8bc95d8dc537be42b24d1260ca44b247b5f2e8fe9542ac0 not found: ID does not exist" Nov 21 14:44:36 crc kubenswrapper[4897]: I1121 14:44:36.234628 4897 scope.go:117] "RemoveContainer" containerID="aced14f00c6d363cc94a4666b24fe09255b379d9023fd9f2ccbc0ea4ff3b1abb" Nov 21 14:44:36 crc kubenswrapper[4897]: E1121 14:44:36.234982 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aced14f00c6d363cc94a4666b24fe09255b379d9023fd9f2ccbc0ea4ff3b1abb\": container with ID starting with aced14f00c6d363cc94a4666b24fe09255b379d9023fd9f2ccbc0ea4ff3b1abb not found: ID does not exist" containerID="aced14f00c6d363cc94a4666b24fe09255b379d9023fd9f2ccbc0ea4ff3b1abb" Nov 21 14:44:36 crc kubenswrapper[4897]: I1121 14:44:36.235063 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aced14f00c6d363cc94a4666b24fe09255b379d9023fd9f2ccbc0ea4ff3b1abb"} err="failed to get container status \"aced14f00c6d363cc94a4666b24fe09255b379d9023fd9f2ccbc0ea4ff3b1abb\": rpc error: code = NotFound desc = could not find container \"aced14f00c6d363cc94a4666b24fe09255b379d9023fd9f2ccbc0ea4ff3b1abb\": container with ID starting with aced14f00c6d363cc94a4666b24fe09255b379d9023fd9f2ccbc0ea4ff3b1abb not found: ID does not exist" Nov 21 14:44:38 crc kubenswrapper[4897]: I1121 14:44:38.101389 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a62940fe-e1a9-42c5-b524-c269aa42d1a5" path="/var/lib/kubelet/pods/a62940fe-e1a9-42c5-b524-c269aa42d1a5/volumes" Nov 21 14:45:00 crc kubenswrapper[4897]: I1121 14:45:00.149495 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29395605-gqtnp"] Nov 21 14:45:00 crc kubenswrapper[4897]: E1121 14:45:00.150706 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a62940fe-e1a9-42c5-b524-c269aa42d1a5" containerName="extract-content" Nov 21 14:45:00 crc kubenswrapper[4897]: I1121 14:45:00.150722 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="a62940fe-e1a9-42c5-b524-c269aa42d1a5" containerName="extract-content" Nov 21 14:45:00 crc kubenswrapper[4897]: E1121 14:45:00.150770 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a62940fe-e1a9-42c5-b524-c269aa42d1a5" containerName="registry-server" Nov 21 14:45:00 crc kubenswrapper[4897]: I1121 14:45:00.150776 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="a62940fe-e1a9-42c5-b524-c269aa42d1a5" containerName="registry-server" Nov 21 14:45:00 crc kubenswrapper[4897]: E1121 14:45:00.150785 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a62940fe-e1a9-42c5-b524-c269aa42d1a5" containerName="extract-utilities" Nov 21 14:45:00 crc kubenswrapper[4897]: I1121 14:45:00.150791 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="a62940fe-e1a9-42c5-b524-c269aa42d1a5" containerName="extract-utilities" Nov 21 14:45:00 crc kubenswrapper[4897]: I1121 14:45:00.151015 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="a62940fe-e1a9-42c5-b524-c269aa42d1a5" containerName="registry-server" Nov 21 14:45:00 crc kubenswrapper[4897]: I1121 14:45:00.151939 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29395605-gqtnp" Nov 21 14:45:00 crc kubenswrapper[4897]: I1121 14:45:00.154124 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 21 14:45:00 crc kubenswrapper[4897]: I1121 14:45:00.155243 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 21 14:45:00 crc kubenswrapper[4897]: I1121 14:45:00.178921 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29395605-gqtnp"] Nov 21 14:45:00 crc kubenswrapper[4897]: I1121 14:45:00.279663 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/52dcab2f-b253-4bb2-a479-518edcb7801a-config-volume\") pod \"collect-profiles-29395605-gqtnp\" (UID: \"52dcab2f-b253-4bb2-a479-518edcb7801a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29395605-gqtnp" Nov 21 14:45:00 crc kubenswrapper[4897]: I1121 14:45:00.279980 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/52dcab2f-b253-4bb2-a479-518edcb7801a-secret-volume\") pod \"collect-profiles-29395605-gqtnp\" (UID: \"52dcab2f-b253-4bb2-a479-518edcb7801a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29395605-gqtnp" Nov 21 14:45:00 crc kubenswrapper[4897]: I1121 14:45:00.280096 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5sdt4\" (UniqueName: \"kubernetes.io/projected/52dcab2f-b253-4bb2-a479-518edcb7801a-kube-api-access-5sdt4\") pod \"collect-profiles-29395605-gqtnp\" (UID: \"52dcab2f-b253-4bb2-a479-518edcb7801a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29395605-gqtnp" Nov 21 14:45:00 crc kubenswrapper[4897]: I1121 14:45:00.382126 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/52dcab2f-b253-4bb2-a479-518edcb7801a-secret-volume\") pod \"collect-profiles-29395605-gqtnp\" (UID: \"52dcab2f-b253-4bb2-a479-518edcb7801a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29395605-gqtnp" Nov 21 14:45:00 crc kubenswrapper[4897]: I1121 14:45:00.382225 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5sdt4\" (UniqueName: \"kubernetes.io/projected/52dcab2f-b253-4bb2-a479-518edcb7801a-kube-api-access-5sdt4\") pod \"collect-profiles-29395605-gqtnp\" (UID: \"52dcab2f-b253-4bb2-a479-518edcb7801a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29395605-gqtnp" Nov 21 14:45:00 crc kubenswrapper[4897]: I1121 14:45:00.382421 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/52dcab2f-b253-4bb2-a479-518edcb7801a-config-volume\") pod \"collect-profiles-29395605-gqtnp\" (UID: \"52dcab2f-b253-4bb2-a479-518edcb7801a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29395605-gqtnp" Nov 21 14:45:00 crc kubenswrapper[4897]: I1121 14:45:00.383264 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/52dcab2f-b253-4bb2-a479-518edcb7801a-config-volume\") pod \"collect-profiles-29395605-gqtnp\" (UID: \"52dcab2f-b253-4bb2-a479-518edcb7801a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29395605-gqtnp" Nov 21 14:45:00 crc kubenswrapper[4897]: I1121 14:45:00.388383 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/52dcab2f-b253-4bb2-a479-518edcb7801a-secret-volume\") pod \"collect-profiles-29395605-gqtnp\" (UID: \"52dcab2f-b253-4bb2-a479-518edcb7801a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29395605-gqtnp" Nov 21 14:45:00 crc kubenswrapper[4897]: I1121 14:45:00.398691 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5sdt4\" (UniqueName: \"kubernetes.io/projected/52dcab2f-b253-4bb2-a479-518edcb7801a-kube-api-access-5sdt4\") pod \"collect-profiles-29395605-gqtnp\" (UID: \"52dcab2f-b253-4bb2-a479-518edcb7801a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29395605-gqtnp" Nov 21 14:45:00 crc kubenswrapper[4897]: I1121 14:45:00.536829 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29395605-gqtnp" Nov 21 14:45:01 crc kubenswrapper[4897]: I1121 14:45:01.017394 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29395605-gqtnp"] Nov 21 14:45:01 crc kubenswrapper[4897]: I1121 14:45:01.407788 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29395605-gqtnp" event={"ID":"52dcab2f-b253-4bb2-a479-518edcb7801a","Type":"ContainerStarted","Data":"1098b67859440bd4a11d413ae31321bc4824e29a052fc8467fe213efa49a6d66"} Nov 21 14:45:01 crc kubenswrapper[4897]: I1121 14:45:01.408107 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29395605-gqtnp" event={"ID":"52dcab2f-b253-4bb2-a479-518edcb7801a","Type":"ContainerStarted","Data":"c065b4406ede1f5cc8f202f29b95811e4f0a4b2ed8de5633c1d45b91ecb881f8"} Nov 21 14:45:01 crc kubenswrapper[4897]: I1121 14:45:01.435783 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29395605-gqtnp" podStartSLOduration=1.435763487 podStartE2EDuration="1.435763487s" podCreationTimestamp="2025-11-21 14:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 14:45:01.427972809 +0000 UTC m=+2178.712566294" watchObservedRunningTime="2025-11-21 14:45:01.435763487 +0000 UTC m=+2178.720356962" Nov 21 14:45:02 crc kubenswrapper[4897]: I1121 14:45:02.418342 4897 generic.go:334] "Generic (PLEG): container finished" podID="52dcab2f-b253-4bb2-a479-518edcb7801a" containerID="1098b67859440bd4a11d413ae31321bc4824e29a052fc8467fe213efa49a6d66" exitCode=0 Nov 21 14:45:02 crc kubenswrapper[4897]: I1121 14:45:02.418433 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29395605-gqtnp" event={"ID":"52dcab2f-b253-4bb2-a479-518edcb7801a","Type":"ContainerDied","Data":"1098b67859440bd4a11d413ae31321bc4824e29a052fc8467fe213efa49a6d66"} Nov 21 14:45:03 crc kubenswrapper[4897]: I1121 14:45:03.853767 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29395605-gqtnp" Nov 21 14:45:03 crc kubenswrapper[4897]: I1121 14:45:03.968049 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/52dcab2f-b253-4bb2-a479-518edcb7801a-secret-volume\") pod \"52dcab2f-b253-4bb2-a479-518edcb7801a\" (UID: \"52dcab2f-b253-4bb2-a479-518edcb7801a\") " Nov 21 14:45:03 crc kubenswrapper[4897]: I1121 14:45:03.968258 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/52dcab2f-b253-4bb2-a479-518edcb7801a-config-volume\") pod \"52dcab2f-b253-4bb2-a479-518edcb7801a\" (UID: \"52dcab2f-b253-4bb2-a479-518edcb7801a\") " Nov 21 14:45:03 crc kubenswrapper[4897]: I1121 14:45:03.968450 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5sdt4\" (UniqueName: \"kubernetes.io/projected/52dcab2f-b253-4bb2-a479-518edcb7801a-kube-api-access-5sdt4\") pod \"52dcab2f-b253-4bb2-a479-518edcb7801a\" (UID: \"52dcab2f-b253-4bb2-a479-518edcb7801a\") " Nov 21 14:45:03 crc kubenswrapper[4897]: I1121 14:45:03.969115 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/52dcab2f-b253-4bb2-a479-518edcb7801a-config-volume" (OuterVolumeSpecName: "config-volume") pod "52dcab2f-b253-4bb2-a479-518edcb7801a" (UID: "52dcab2f-b253-4bb2-a479-518edcb7801a"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:45:03 crc kubenswrapper[4897]: I1121 14:45:03.969380 4897 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/52dcab2f-b253-4bb2-a479-518edcb7801a-config-volume\") on node \"crc\" DevicePath \"\"" Nov 21 14:45:03 crc kubenswrapper[4897]: I1121 14:45:03.974425 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/52dcab2f-b253-4bb2-a479-518edcb7801a-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "52dcab2f-b253-4bb2-a479-518edcb7801a" (UID: "52dcab2f-b253-4bb2-a479-518edcb7801a"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:45:03 crc kubenswrapper[4897]: I1121 14:45:03.986911 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/52dcab2f-b253-4bb2-a479-518edcb7801a-kube-api-access-5sdt4" (OuterVolumeSpecName: "kube-api-access-5sdt4") pod "52dcab2f-b253-4bb2-a479-518edcb7801a" (UID: "52dcab2f-b253-4bb2-a479-518edcb7801a"). InnerVolumeSpecName "kube-api-access-5sdt4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:45:04 crc kubenswrapper[4897]: I1121 14:45:04.073337 4897 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/52dcab2f-b253-4bb2-a479-518edcb7801a-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 21 14:45:04 crc kubenswrapper[4897]: I1121 14:45:04.073388 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5sdt4\" (UniqueName: \"kubernetes.io/projected/52dcab2f-b253-4bb2-a479-518edcb7801a-kube-api-access-5sdt4\") on node \"crc\" DevicePath \"\"" Nov 21 14:45:04 crc kubenswrapper[4897]: I1121 14:45:04.441668 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29395605-gqtnp" event={"ID":"52dcab2f-b253-4bb2-a479-518edcb7801a","Type":"ContainerDied","Data":"c065b4406ede1f5cc8f202f29b95811e4f0a4b2ed8de5633c1d45b91ecb881f8"} Nov 21 14:45:04 crc kubenswrapper[4897]: I1121 14:45:04.441985 4897 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c065b4406ede1f5cc8f202f29b95811e4f0a4b2ed8de5633c1d45b91ecb881f8" Nov 21 14:45:04 crc kubenswrapper[4897]: I1121 14:45:04.441764 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29395605-gqtnp" Nov 21 14:45:04 crc kubenswrapper[4897]: I1121 14:45:04.925108 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29395560-pq7jf"] Nov 21 14:45:04 crc kubenswrapper[4897]: I1121 14:45:04.936225 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29395560-pq7jf"] Nov 21 14:45:06 crc kubenswrapper[4897]: I1121 14:45:06.103860 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c14a1c93-8be4-42b7-aae1-9c31d53a6bfc" path="/var/lib/kubelet/pods/c14a1c93-8be4-42b7-aae1-9c31d53a6bfc/volumes" Nov 21 14:45:20 crc kubenswrapper[4897]: I1121 14:45:20.977726 4897 scope.go:117] "RemoveContainer" containerID="e91c668b832ac90da6a2db93f34a74105ac8f124d7251d1e057949d2b4fc96c6" Nov 21 14:45:21 crc kubenswrapper[4897]: I1121 14:45:21.023935 4897 scope.go:117] "RemoveContainer" containerID="3723470e7b3e96c5e7a84947a48a7a6b9cf8050e5366b390cb2a6236ee20e58c" Nov 21 14:45:21 crc kubenswrapper[4897]: I1121 14:45:21.052887 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-cv5pw"] Nov 21 14:45:21 crc kubenswrapper[4897]: I1121 14:45:21.070099 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-cv5pw"] Nov 21 14:45:21 crc kubenswrapper[4897]: I1121 14:45:21.106565 4897 scope.go:117] "RemoveContainer" containerID="37527d0435969a408457ca6e56d44e25c335f7b2875e57166c3d10f6c123e1c7" Nov 21 14:45:21 crc kubenswrapper[4897]: I1121 14:45:21.209680 4897 scope.go:117] "RemoveContainer" containerID="7c71e97295b9fe2b2d30a0502ffb2637c6dd199e5c02c6709abf1958f64afa5a" Nov 21 14:45:21 crc kubenswrapper[4897]: I1121 14:45:21.277720 4897 scope.go:117] "RemoveContainer" containerID="fe610035df51ba9b7ec1dd836f736c0f477e2aaee4d035bbc2b9dd97f160dd34" Nov 21 14:45:22 crc kubenswrapper[4897]: I1121 14:45:22.129975 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c6e90788-4f97-45c0-aacc-cfa848ac0f2f" path="/var/lib/kubelet/pods/c6e90788-4f97-45c0-aacc-cfa848ac0f2f/volumes" Nov 21 14:45:34 crc kubenswrapper[4897]: I1121 14:45:34.370563 4897 patch_prober.go:28] interesting pod/machine-config-daemon-krv5b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 21 14:45:34 crc kubenswrapper[4897]: I1121 14:45:34.371218 4897 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 21 14:45:46 crc kubenswrapper[4897]: I1121 14:45:46.049392 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-x8ttp"] Nov 21 14:45:46 crc kubenswrapper[4897]: I1121 14:45:46.064040 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-x8ttp"] Nov 21 14:45:46 crc kubenswrapper[4897]: I1121 14:45:46.102296 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="518b64a4-d073-462f-ada5-20209e14ff27" path="/var/lib/kubelet/pods/518b64a4-d073-462f-ada5-20209e14ff27/volumes" Nov 21 14:45:50 crc kubenswrapper[4897]: I1121 14:45:50.922424 4897 generic.go:334] "Generic (PLEG): container finished" podID="29e88709-d2ef-475a-85d2-1573f0719c8f" containerID="347a88ca18f62930f76018356446be838b94309ceb00131bdc44bb2afba978cf" exitCode=0 Nov 21 14:45:50 crc kubenswrapper[4897]: I1121 14:45:50.922565 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k27v5" event={"ID":"29e88709-d2ef-475a-85d2-1573f0719c8f","Type":"ContainerDied","Data":"347a88ca18f62930f76018356446be838b94309ceb00131bdc44bb2afba978cf"} Nov 21 14:45:52 crc kubenswrapper[4897]: I1121 14:45:52.412126 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k27v5" Nov 21 14:45:52 crc kubenswrapper[4897]: I1121 14:45:52.476218 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/29e88709-d2ef-475a-85d2-1573f0719c8f-inventory\") pod \"29e88709-d2ef-475a-85d2-1573f0719c8f\" (UID: \"29e88709-d2ef-475a-85d2-1573f0719c8f\") " Nov 21 14:45:52 crc kubenswrapper[4897]: I1121 14:45:52.476573 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29e88709-d2ef-475a-85d2-1573f0719c8f-bootstrap-combined-ca-bundle\") pod \"29e88709-d2ef-475a-85d2-1573f0719c8f\" (UID: \"29e88709-d2ef-475a-85d2-1573f0719c8f\") " Nov 21 14:45:52 crc kubenswrapper[4897]: I1121 14:45:52.476609 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/29e88709-d2ef-475a-85d2-1573f0719c8f-ssh-key\") pod \"29e88709-d2ef-475a-85d2-1573f0719c8f\" (UID: \"29e88709-d2ef-475a-85d2-1573f0719c8f\") " Nov 21 14:45:52 crc kubenswrapper[4897]: I1121 14:45:52.476701 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2frpp\" (UniqueName: \"kubernetes.io/projected/29e88709-d2ef-475a-85d2-1573f0719c8f-kube-api-access-2frpp\") pod \"29e88709-d2ef-475a-85d2-1573f0719c8f\" (UID: \"29e88709-d2ef-475a-85d2-1573f0719c8f\") " Nov 21 14:45:52 crc kubenswrapper[4897]: I1121 14:45:52.483098 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/29e88709-d2ef-475a-85d2-1573f0719c8f-kube-api-access-2frpp" (OuterVolumeSpecName: "kube-api-access-2frpp") pod "29e88709-d2ef-475a-85d2-1573f0719c8f" (UID: "29e88709-d2ef-475a-85d2-1573f0719c8f"). InnerVolumeSpecName "kube-api-access-2frpp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:45:52 crc kubenswrapper[4897]: I1121 14:45:52.486017 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29e88709-d2ef-475a-85d2-1573f0719c8f-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "29e88709-d2ef-475a-85d2-1573f0719c8f" (UID: "29e88709-d2ef-475a-85d2-1573f0719c8f"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:45:52 crc kubenswrapper[4897]: I1121 14:45:52.531674 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29e88709-d2ef-475a-85d2-1573f0719c8f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "29e88709-d2ef-475a-85d2-1573f0719c8f" (UID: "29e88709-d2ef-475a-85d2-1573f0719c8f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:45:52 crc kubenswrapper[4897]: I1121 14:45:52.583521 4897 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29e88709-d2ef-475a-85d2-1573f0719c8f-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 21 14:45:52 crc kubenswrapper[4897]: I1121 14:45:52.583572 4897 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/29e88709-d2ef-475a-85d2-1573f0719c8f-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 21 14:45:52 crc kubenswrapper[4897]: I1121 14:45:52.583585 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2frpp\" (UniqueName: \"kubernetes.io/projected/29e88709-d2ef-475a-85d2-1573f0719c8f-kube-api-access-2frpp\") on node \"crc\" DevicePath \"\"" Nov 21 14:45:52 crc kubenswrapper[4897]: I1121 14:45:52.625303 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29e88709-d2ef-475a-85d2-1573f0719c8f-inventory" (OuterVolumeSpecName: "inventory") pod "29e88709-d2ef-475a-85d2-1573f0719c8f" (UID: "29e88709-d2ef-475a-85d2-1573f0719c8f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:45:52 crc kubenswrapper[4897]: I1121 14:45:52.685735 4897 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/29e88709-d2ef-475a-85d2-1573f0719c8f-inventory\") on node \"crc\" DevicePath \"\"" Nov 21 14:45:52 crc kubenswrapper[4897]: I1121 14:45:52.943742 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k27v5" event={"ID":"29e88709-d2ef-475a-85d2-1573f0719c8f","Type":"ContainerDied","Data":"5228a1c2999b65d7651dc1082c01d96d2b1f9028b50d001ed0bc4ca8c1475d93"} Nov 21 14:45:52 crc kubenswrapper[4897]: I1121 14:45:52.943781 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k27v5" Nov 21 14:45:52 crc kubenswrapper[4897]: I1121 14:45:52.943792 4897 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5228a1c2999b65d7651dc1082c01d96d2b1f9028b50d001ed0bc4ca8c1475d93" Nov 21 14:45:53 crc kubenswrapper[4897]: I1121 14:45:53.027382 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bqj2z"] Nov 21 14:45:53 crc kubenswrapper[4897]: E1121 14:45:53.027913 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52dcab2f-b253-4bb2-a479-518edcb7801a" containerName="collect-profiles" Nov 21 14:45:53 crc kubenswrapper[4897]: I1121 14:45:53.027933 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="52dcab2f-b253-4bb2-a479-518edcb7801a" containerName="collect-profiles" Nov 21 14:45:53 crc kubenswrapper[4897]: E1121 14:45:53.027950 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29e88709-d2ef-475a-85d2-1573f0719c8f" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 21 14:45:53 crc kubenswrapper[4897]: I1121 14:45:53.027957 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="29e88709-d2ef-475a-85d2-1573f0719c8f" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 21 14:45:53 crc kubenswrapper[4897]: I1121 14:45:53.028182 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="29e88709-d2ef-475a-85d2-1573f0719c8f" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 21 14:45:53 crc kubenswrapper[4897]: I1121 14:45:53.028206 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="52dcab2f-b253-4bb2-a479-518edcb7801a" containerName="collect-profiles" Nov 21 14:45:53 crc kubenswrapper[4897]: I1121 14:45:53.029044 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bqj2z" Nov 21 14:45:53 crc kubenswrapper[4897]: I1121 14:45:53.031930 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 21 14:45:53 crc kubenswrapper[4897]: I1121 14:45:53.032137 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 21 14:45:53 crc kubenswrapper[4897]: I1121 14:45:53.032229 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 21 14:45:53 crc kubenswrapper[4897]: I1121 14:45:53.032861 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-dj88m" Nov 21 14:45:53 crc kubenswrapper[4897]: I1121 14:45:53.062255 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bqj2z"] Nov 21 14:45:53 crc kubenswrapper[4897]: I1121 14:45:53.098364 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d1a63d47-990b-40df-a70b-bdd3467bb792-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-bqj2z\" (UID: \"d1a63d47-990b-40df-a70b-bdd3467bb792\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bqj2z" Nov 21 14:45:53 crc kubenswrapper[4897]: I1121 14:45:53.098627 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-96bjq\" (UniqueName: \"kubernetes.io/projected/d1a63d47-990b-40df-a70b-bdd3467bb792-kube-api-access-96bjq\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-bqj2z\" (UID: \"d1a63d47-990b-40df-a70b-bdd3467bb792\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bqj2z" Nov 21 14:45:53 crc kubenswrapper[4897]: I1121 14:45:53.098806 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d1a63d47-990b-40df-a70b-bdd3467bb792-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-bqj2z\" (UID: \"d1a63d47-990b-40df-a70b-bdd3467bb792\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bqj2z" Nov 21 14:45:53 crc kubenswrapper[4897]: I1121 14:45:53.200607 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d1a63d47-990b-40df-a70b-bdd3467bb792-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-bqj2z\" (UID: \"d1a63d47-990b-40df-a70b-bdd3467bb792\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bqj2z" Nov 21 14:45:53 crc kubenswrapper[4897]: I1121 14:45:53.200769 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d1a63d47-990b-40df-a70b-bdd3467bb792-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-bqj2z\" (UID: \"d1a63d47-990b-40df-a70b-bdd3467bb792\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bqj2z" Nov 21 14:45:53 crc kubenswrapper[4897]: I1121 14:45:53.200841 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-96bjq\" (UniqueName: \"kubernetes.io/projected/d1a63d47-990b-40df-a70b-bdd3467bb792-kube-api-access-96bjq\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-bqj2z\" (UID: \"d1a63d47-990b-40df-a70b-bdd3467bb792\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bqj2z" Nov 21 14:45:53 crc kubenswrapper[4897]: I1121 14:45:53.204674 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d1a63d47-990b-40df-a70b-bdd3467bb792-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-bqj2z\" (UID: \"d1a63d47-990b-40df-a70b-bdd3467bb792\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bqj2z" Nov 21 14:45:53 crc kubenswrapper[4897]: I1121 14:45:53.210291 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d1a63d47-990b-40df-a70b-bdd3467bb792-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-bqj2z\" (UID: \"d1a63d47-990b-40df-a70b-bdd3467bb792\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bqj2z" Nov 21 14:45:53 crc kubenswrapper[4897]: I1121 14:45:53.222776 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-96bjq\" (UniqueName: \"kubernetes.io/projected/d1a63d47-990b-40df-a70b-bdd3467bb792-kube-api-access-96bjq\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-bqj2z\" (UID: \"d1a63d47-990b-40df-a70b-bdd3467bb792\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bqj2z" Nov 21 14:45:53 crc kubenswrapper[4897]: I1121 14:45:53.354035 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bqj2z" Nov 21 14:45:53 crc kubenswrapper[4897]: I1121 14:45:53.922118 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bqj2z"] Nov 21 14:45:53 crc kubenswrapper[4897]: I1121 14:45:53.927265 4897 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 21 14:45:53 crc kubenswrapper[4897]: I1121 14:45:53.954608 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bqj2z" event={"ID":"d1a63d47-990b-40df-a70b-bdd3467bb792","Type":"ContainerStarted","Data":"720c26537e69685d649602f1f28869ce05983204c91ff385fbb1c52527bc29f6"} Nov 21 14:45:55 crc kubenswrapper[4897]: I1121 14:45:55.977633 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bqj2z" event={"ID":"d1a63d47-990b-40df-a70b-bdd3467bb792","Type":"ContainerStarted","Data":"9c4d0eb2399e2b76a1aef7e0060ff496f9f9c654f866a7c1a13909deaad6e399"} Nov 21 14:45:55 crc kubenswrapper[4897]: I1121 14:45:55.996275 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bqj2z" podStartSLOduration=1.958665584 podStartE2EDuration="2.996258892s" podCreationTimestamp="2025-11-21 14:45:53 +0000 UTC" firstStartedPulling="2025-11-21 14:45:53.927076868 +0000 UTC m=+2231.211670343" lastFinishedPulling="2025-11-21 14:45:54.964670176 +0000 UTC m=+2232.249263651" observedRunningTime="2025-11-21 14:45:55.994785483 +0000 UTC m=+2233.279378968" watchObservedRunningTime="2025-11-21 14:45:55.996258892 +0000 UTC m=+2233.280852357" Nov 21 14:46:00 crc kubenswrapper[4897]: I1121 14:46:00.033374 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-bpgdn"] Nov 21 14:46:00 crc kubenswrapper[4897]: I1121 14:46:00.043669 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-bpgdn"] Nov 21 14:46:00 crc kubenswrapper[4897]: I1121 14:46:00.102618 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a558cde2-95bc-4f8d-9daf-90fe592ab9ff" path="/var/lib/kubelet/pods/a558cde2-95bc-4f8d-9daf-90fe592ab9ff/volumes" Nov 21 14:46:04 crc kubenswrapper[4897]: I1121 14:46:04.370598 4897 patch_prober.go:28] interesting pod/machine-config-daemon-krv5b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 21 14:46:04 crc kubenswrapper[4897]: I1121 14:46:04.372216 4897 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 21 14:46:21 crc kubenswrapper[4897]: I1121 14:46:21.472075 4897 scope.go:117] "RemoveContainer" containerID="b85b61c6134535cb1d4f859af67cc0a923a31c62fbd636fe98341334044c0bd0" Nov 21 14:46:21 crc kubenswrapper[4897]: I1121 14:46:21.506907 4897 scope.go:117] "RemoveContainer" containerID="03e6cd518a468591cf4ed2a936e891b6a325fffb7bd905bc50948a197be7a6cb" Nov 21 14:46:21 crc kubenswrapper[4897]: I1121 14:46:21.568396 4897 scope.go:117] "RemoveContainer" containerID="5d3558195d06f9c88281f03204b0c131aad447a4332e1b80f7f41f8bf27b346b" Nov 21 14:46:30 crc kubenswrapper[4897]: I1121 14:46:30.194203 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-7q2jl"] Nov 21 14:46:30 crc kubenswrapper[4897]: I1121 14:46:30.197881 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7q2jl" Nov 21 14:46:30 crc kubenswrapper[4897]: I1121 14:46:30.214671 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7q2jl"] Nov 21 14:46:30 crc kubenswrapper[4897]: I1121 14:46:30.260704 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa6bb2ce-428b-4104-8823-99402d5540eb-catalog-content\") pod \"redhat-operators-7q2jl\" (UID: \"fa6bb2ce-428b-4104-8823-99402d5540eb\") " pod="openshift-marketplace/redhat-operators-7q2jl" Nov 21 14:46:30 crc kubenswrapper[4897]: I1121 14:46:30.260894 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa6bb2ce-428b-4104-8823-99402d5540eb-utilities\") pod \"redhat-operators-7q2jl\" (UID: \"fa6bb2ce-428b-4104-8823-99402d5540eb\") " pod="openshift-marketplace/redhat-operators-7q2jl" Nov 21 14:46:30 crc kubenswrapper[4897]: I1121 14:46:30.260932 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-98vsv\" (UniqueName: \"kubernetes.io/projected/fa6bb2ce-428b-4104-8823-99402d5540eb-kube-api-access-98vsv\") pod \"redhat-operators-7q2jl\" (UID: \"fa6bb2ce-428b-4104-8823-99402d5540eb\") " pod="openshift-marketplace/redhat-operators-7q2jl" Nov 21 14:46:30 crc kubenswrapper[4897]: I1121 14:46:30.362878 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa6bb2ce-428b-4104-8823-99402d5540eb-utilities\") pod \"redhat-operators-7q2jl\" (UID: \"fa6bb2ce-428b-4104-8823-99402d5540eb\") " pod="openshift-marketplace/redhat-operators-7q2jl" Nov 21 14:46:30 crc kubenswrapper[4897]: I1121 14:46:30.362945 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-98vsv\" (UniqueName: \"kubernetes.io/projected/fa6bb2ce-428b-4104-8823-99402d5540eb-kube-api-access-98vsv\") pod \"redhat-operators-7q2jl\" (UID: \"fa6bb2ce-428b-4104-8823-99402d5540eb\") " pod="openshift-marketplace/redhat-operators-7q2jl" Nov 21 14:46:30 crc kubenswrapper[4897]: I1121 14:46:30.363105 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa6bb2ce-428b-4104-8823-99402d5540eb-catalog-content\") pod \"redhat-operators-7q2jl\" (UID: \"fa6bb2ce-428b-4104-8823-99402d5540eb\") " pod="openshift-marketplace/redhat-operators-7q2jl" Nov 21 14:46:30 crc kubenswrapper[4897]: I1121 14:46:30.363470 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa6bb2ce-428b-4104-8823-99402d5540eb-utilities\") pod \"redhat-operators-7q2jl\" (UID: \"fa6bb2ce-428b-4104-8823-99402d5540eb\") " pod="openshift-marketplace/redhat-operators-7q2jl" Nov 21 14:46:30 crc kubenswrapper[4897]: I1121 14:46:30.363575 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa6bb2ce-428b-4104-8823-99402d5540eb-catalog-content\") pod \"redhat-operators-7q2jl\" (UID: \"fa6bb2ce-428b-4104-8823-99402d5540eb\") " pod="openshift-marketplace/redhat-operators-7q2jl" Nov 21 14:46:30 crc kubenswrapper[4897]: I1121 14:46:30.384892 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-98vsv\" (UniqueName: \"kubernetes.io/projected/fa6bb2ce-428b-4104-8823-99402d5540eb-kube-api-access-98vsv\") pod \"redhat-operators-7q2jl\" (UID: \"fa6bb2ce-428b-4104-8823-99402d5540eb\") " pod="openshift-marketplace/redhat-operators-7q2jl" Nov 21 14:46:30 crc kubenswrapper[4897]: I1121 14:46:30.533077 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7q2jl" Nov 21 14:46:31 crc kubenswrapper[4897]: I1121 14:46:31.007553 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7q2jl"] Nov 21 14:46:31 crc kubenswrapper[4897]: I1121 14:46:31.337645 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7q2jl" event={"ID":"fa6bb2ce-428b-4104-8823-99402d5540eb","Type":"ContainerStarted","Data":"7f4fd3857db49627de17fa20bc8870281caef7afb27b1355eab2e8764797422d"} Nov 21 14:46:32 crc kubenswrapper[4897]: I1121 14:46:32.045616 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-dgr4b"] Nov 21 14:46:32 crc kubenswrapper[4897]: I1121 14:46:32.060245 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-dgr4b"] Nov 21 14:46:32 crc kubenswrapper[4897]: I1121 14:46:32.105735 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="522a19f3-71a6-4609-89cd-6142f5f0d375" path="/var/lib/kubelet/pods/522a19f3-71a6-4609-89cd-6142f5f0d375/volumes" Nov 21 14:46:32 crc kubenswrapper[4897]: I1121 14:46:32.351466 4897 generic.go:334] "Generic (PLEG): container finished" podID="fa6bb2ce-428b-4104-8823-99402d5540eb" containerID="459ac624f4ce8df735dc72685519998622a1ca3c2028b3c5a6c26460f27b5371" exitCode=0 Nov 21 14:46:32 crc kubenswrapper[4897]: I1121 14:46:32.351558 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7q2jl" event={"ID":"fa6bb2ce-428b-4104-8823-99402d5540eb","Type":"ContainerDied","Data":"459ac624f4ce8df735dc72685519998622a1ca3c2028b3c5a6c26460f27b5371"} Nov 21 14:46:34 crc kubenswrapper[4897]: I1121 14:46:34.371367 4897 patch_prober.go:28] interesting pod/machine-config-daemon-krv5b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 21 14:46:34 crc kubenswrapper[4897]: I1121 14:46:34.371434 4897 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 21 14:46:34 crc kubenswrapper[4897]: I1121 14:46:34.371483 4897 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" Nov 21 14:46:34 crc kubenswrapper[4897]: I1121 14:46:34.372089 4897 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"dc3c0a9877f70b7d5c55314a92df879b827a4103351566cc755cfbf916773aa2"} pod="openshift-machine-config-operator/machine-config-daemon-krv5b" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 21 14:46:34 crc kubenswrapper[4897]: I1121 14:46:34.372163 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" containerName="machine-config-daemon" containerID="cri-o://dc3c0a9877f70b7d5c55314a92df879b827a4103351566cc755cfbf916773aa2" gracePeriod=600 Nov 21 14:46:34 crc kubenswrapper[4897]: E1121 14:46:34.580354 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 14:46:34 crc kubenswrapper[4897]: E1121 14:46:34.760307 4897 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode7670227_d280_4847_b882_754429f56b0f.slice/crio-conmon-dc3c0a9877f70b7d5c55314a92df879b827a4103351566cc755cfbf916773aa2.scope\": RecentStats: unable to find data in memory cache]" Nov 21 14:46:35 crc kubenswrapper[4897]: I1121 14:46:35.400764 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7q2jl" event={"ID":"fa6bb2ce-428b-4104-8823-99402d5540eb","Type":"ContainerStarted","Data":"402be7df4c6824fef386bd48ed434a263ba1ea7d55a4818f231d1f575da47f98"} Nov 21 14:46:35 crc kubenswrapper[4897]: I1121 14:46:35.402942 4897 generic.go:334] "Generic (PLEG): container finished" podID="e7670227-d280-4847-b882-754429f56b0f" containerID="dc3c0a9877f70b7d5c55314a92df879b827a4103351566cc755cfbf916773aa2" exitCode=0 Nov 21 14:46:35 crc kubenswrapper[4897]: I1121 14:46:35.402984 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" event={"ID":"e7670227-d280-4847-b882-754429f56b0f","Type":"ContainerDied","Data":"dc3c0a9877f70b7d5c55314a92df879b827a4103351566cc755cfbf916773aa2"} Nov 21 14:46:35 crc kubenswrapper[4897]: I1121 14:46:35.403016 4897 scope.go:117] "RemoveContainer" containerID="a4d7f2b343d4dace3a7a2ec8aeabc1074996aaa7dc843ecf94e43c56479ac344" Nov 21 14:46:35 crc kubenswrapper[4897]: I1121 14:46:35.403864 4897 scope.go:117] "RemoveContainer" containerID="dc3c0a9877f70b7d5c55314a92df879b827a4103351566cc755cfbf916773aa2" Nov 21 14:46:35 crc kubenswrapper[4897]: E1121 14:46:35.404139 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 14:46:50 crc kubenswrapper[4897]: I1121 14:46:50.089665 4897 scope.go:117] "RemoveContainer" containerID="dc3c0a9877f70b7d5c55314a92df879b827a4103351566cc755cfbf916773aa2" Nov 21 14:46:50 crc kubenswrapper[4897]: E1121 14:46:50.090554 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 14:46:57 crc kubenswrapper[4897]: I1121 14:46:57.642964 4897 generic.go:334] "Generic (PLEG): container finished" podID="fa6bb2ce-428b-4104-8823-99402d5540eb" containerID="402be7df4c6824fef386bd48ed434a263ba1ea7d55a4818f231d1f575da47f98" exitCode=0 Nov 21 14:46:57 crc kubenswrapper[4897]: I1121 14:46:57.643045 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7q2jl" event={"ID":"fa6bb2ce-428b-4104-8823-99402d5540eb","Type":"ContainerDied","Data":"402be7df4c6824fef386bd48ed434a263ba1ea7d55a4818f231d1f575da47f98"} Nov 21 14:46:59 crc kubenswrapper[4897]: I1121 14:46:59.675465 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7q2jl" event={"ID":"fa6bb2ce-428b-4104-8823-99402d5540eb","Type":"ContainerStarted","Data":"89d4cb1cf846fb322bc3d686ed55a4664f1ad9f506becc41c215ec82e10b82ea"} Nov 21 14:46:59 crc kubenswrapper[4897]: I1121 14:46:59.713053 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-7q2jl" podStartSLOduration=3.206772203 podStartE2EDuration="29.71303403s" podCreationTimestamp="2025-11-21 14:46:30 +0000 UTC" firstStartedPulling="2025-11-21 14:46:32.354041312 +0000 UTC m=+2269.638634787" lastFinishedPulling="2025-11-21 14:46:58.860303139 +0000 UTC m=+2296.144896614" observedRunningTime="2025-11-21 14:46:59.69546836 +0000 UTC m=+2296.980061835" watchObservedRunningTime="2025-11-21 14:46:59.71303403 +0000 UTC m=+2296.997627505" Nov 21 14:47:00 crc kubenswrapper[4897]: I1121 14:47:00.533520 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-7q2jl" Nov 21 14:47:00 crc kubenswrapper[4897]: I1121 14:47:00.533880 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-7q2jl" Nov 21 14:47:01 crc kubenswrapper[4897]: I1121 14:47:01.585273 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-7q2jl" podUID="fa6bb2ce-428b-4104-8823-99402d5540eb" containerName="registry-server" probeResult="failure" output=< Nov 21 14:47:01 crc kubenswrapper[4897]: timeout: failed to connect service ":50051" within 1s Nov 21 14:47:01 crc kubenswrapper[4897]: > Nov 21 14:47:05 crc kubenswrapper[4897]: I1121 14:47:05.089283 4897 scope.go:117] "RemoveContainer" containerID="dc3c0a9877f70b7d5c55314a92df879b827a4103351566cc755cfbf916773aa2" Nov 21 14:47:05 crc kubenswrapper[4897]: E1121 14:47:05.089976 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 14:47:11 crc kubenswrapper[4897]: I1121 14:47:11.583689 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-7q2jl" podUID="fa6bb2ce-428b-4104-8823-99402d5540eb" containerName="registry-server" probeResult="failure" output=< Nov 21 14:47:11 crc kubenswrapper[4897]: timeout: failed to connect service ":50051" within 1s Nov 21 14:47:11 crc kubenswrapper[4897]: > Nov 21 14:47:19 crc kubenswrapper[4897]: I1121 14:47:19.089822 4897 scope.go:117] "RemoveContainer" containerID="dc3c0a9877f70b7d5c55314a92df879b827a4103351566cc755cfbf916773aa2" Nov 21 14:47:19 crc kubenswrapper[4897]: E1121 14:47:19.090680 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 14:47:20 crc kubenswrapper[4897]: I1121 14:47:20.592334 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-7q2jl" Nov 21 14:47:20 crc kubenswrapper[4897]: I1121 14:47:20.653439 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-7q2jl" Nov 21 14:47:20 crc kubenswrapper[4897]: I1121 14:47:20.851012 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7q2jl"] Nov 21 14:47:21 crc kubenswrapper[4897]: I1121 14:47:21.685158 4897 scope.go:117] "RemoveContainer" containerID="8f7788e6ea517fa5675fbd9e32ba8202c7b3ea53d2a83293a84c4235cad59e54" Nov 21 14:47:21 crc kubenswrapper[4897]: I1121 14:47:21.934455 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-7q2jl" podUID="fa6bb2ce-428b-4104-8823-99402d5540eb" containerName="registry-server" containerID="cri-o://89d4cb1cf846fb322bc3d686ed55a4664f1ad9f506becc41c215ec82e10b82ea" gracePeriod=2 Nov 21 14:47:22 crc kubenswrapper[4897]: I1121 14:47:22.945262 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7q2jl" Nov 21 14:47:22 crc kubenswrapper[4897]: I1121 14:47:22.945287 4897 generic.go:334] "Generic (PLEG): container finished" podID="fa6bb2ce-428b-4104-8823-99402d5540eb" containerID="89d4cb1cf846fb322bc3d686ed55a4664f1ad9f506becc41c215ec82e10b82ea" exitCode=0 Nov 21 14:47:22 crc kubenswrapper[4897]: I1121 14:47:22.945315 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7q2jl" event={"ID":"fa6bb2ce-428b-4104-8823-99402d5540eb","Type":"ContainerDied","Data":"89d4cb1cf846fb322bc3d686ed55a4664f1ad9f506becc41c215ec82e10b82ea"} Nov 21 14:47:22 crc kubenswrapper[4897]: I1121 14:47:22.945724 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7q2jl" event={"ID":"fa6bb2ce-428b-4104-8823-99402d5540eb","Type":"ContainerDied","Data":"7f4fd3857db49627de17fa20bc8870281caef7afb27b1355eab2e8764797422d"} Nov 21 14:47:22 crc kubenswrapper[4897]: I1121 14:47:22.945768 4897 scope.go:117] "RemoveContainer" containerID="89d4cb1cf846fb322bc3d686ed55a4664f1ad9f506becc41c215ec82e10b82ea" Nov 21 14:47:22 crc kubenswrapper[4897]: I1121 14:47:22.974222 4897 scope.go:117] "RemoveContainer" containerID="402be7df4c6824fef386bd48ed434a263ba1ea7d55a4818f231d1f575da47f98" Nov 21 14:47:23 crc kubenswrapper[4897]: I1121 14:47:23.012410 4897 scope.go:117] "RemoveContainer" containerID="459ac624f4ce8df735dc72685519998622a1ca3c2028b3c5a6c26460f27b5371" Nov 21 14:47:23 crc kubenswrapper[4897]: I1121 14:47:23.059562 4897 scope.go:117] "RemoveContainer" containerID="89d4cb1cf846fb322bc3d686ed55a4664f1ad9f506becc41c215ec82e10b82ea" Nov 21 14:47:23 crc kubenswrapper[4897]: E1121 14:47:23.059937 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"89d4cb1cf846fb322bc3d686ed55a4664f1ad9f506becc41c215ec82e10b82ea\": container with ID starting with 89d4cb1cf846fb322bc3d686ed55a4664f1ad9f506becc41c215ec82e10b82ea not found: ID does not exist" containerID="89d4cb1cf846fb322bc3d686ed55a4664f1ad9f506becc41c215ec82e10b82ea" Nov 21 14:47:23 crc kubenswrapper[4897]: I1121 14:47:23.059998 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"89d4cb1cf846fb322bc3d686ed55a4664f1ad9f506becc41c215ec82e10b82ea"} err="failed to get container status \"89d4cb1cf846fb322bc3d686ed55a4664f1ad9f506becc41c215ec82e10b82ea\": rpc error: code = NotFound desc = could not find container \"89d4cb1cf846fb322bc3d686ed55a4664f1ad9f506becc41c215ec82e10b82ea\": container with ID starting with 89d4cb1cf846fb322bc3d686ed55a4664f1ad9f506becc41c215ec82e10b82ea not found: ID does not exist" Nov 21 14:47:23 crc kubenswrapper[4897]: I1121 14:47:23.060031 4897 scope.go:117] "RemoveContainer" containerID="402be7df4c6824fef386bd48ed434a263ba1ea7d55a4818f231d1f575da47f98" Nov 21 14:47:23 crc kubenswrapper[4897]: E1121 14:47:23.060838 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"402be7df4c6824fef386bd48ed434a263ba1ea7d55a4818f231d1f575da47f98\": container with ID starting with 402be7df4c6824fef386bd48ed434a263ba1ea7d55a4818f231d1f575da47f98 not found: ID does not exist" containerID="402be7df4c6824fef386bd48ed434a263ba1ea7d55a4818f231d1f575da47f98" Nov 21 14:47:23 crc kubenswrapper[4897]: I1121 14:47:23.060874 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"402be7df4c6824fef386bd48ed434a263ba1ea7d55a4818f231d1f575da47f98"} err="failed to get container status \"402be7df4c6824fef386bd48ed434a263ba1ea7d55a4818f231d1f575da47f98\": rpc error: code = NotFound desc = could not find container \"402be7df4c6824fef386bd48ed434a263ba1ea7d55a4818f231d1f575da47f98\": container with ID starting with 402be7df4c6824fef386bd48ed434a263ba1ea7d55a4818f231d1f575da47f98 not found: ID does not exist" Nov 21 14:47:23 crc kubenswrapper[4897]: I1121 14:47:23.060891 4897 scope.go:117] "RemoveContainer" containerID="459ac624f4ce8df735dc72685519998622a1ca3c2028b3c5a6c26460f27b5371" Nov 21 14:47:23 crc kubenswrapper[4897]: E1121 14:47:23.061106 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"459ac624f4ce8df735dc72685519998622a1ca3c2028b3c5a6c26460f27b5371\": container with ID starting with 459ac624f4ce8df735dc72685519998622a1ca3c2028b3c5a6c26460f27b5371 not found: ID does not exist" containerID="459ac624f4ce8df735dc72685519998622a1ca3c2028b3c5a6c26460f27b5371" Nov 21 14:47:23 crc kubenswrapper[4897]: I1121 14:47:23.061129 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"459ac624f4ce8df735dc72685519998622a1ca3c2028b3c5a6c26460f27b5371"} err="failed to get container status \"459ac624f4ce8df735dc72685519998622a1ca3c2028b3c5a6c26460f27b5371\": rpc error: code = NotFound desc = could not find container \"459ac624f4ce8df735dc72685519998622a1ca3c2028b3c5a6c26460f27b5371\": container with ID starting with 459ac624f4ce8df735dc72685519998622a1ca3c2028b3c5a6c26460f27b5371 not found: ID does not exist" Nov 21 14:47:23 crc kubenswrapper[4897]: I1121 14:47:23.087816 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa6bb2ce-428b-4104-8823-99402d5540eb-catalog-content\") pod \"fa6bb2ce-428b-4104-8823-99402d5540eb\" (UID: \"fa6bb2ce-428b-4104-8823-99402d5540eb\") " Nov 21 14:47:23 crc kubenswrapper[4897]: I1121 14:47:23.087911 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-98vsv\" (UniqueName: \"kubernetes.io/projected/fa6bb2ce-428b-4104-8823-99402d5540eb-kube-api-access-98vsv\") pod \"fa6bb2ce-428b-4104-8823-99402d5540eb\" (UID: \"fa6bb2ce-428b-4104-8823-99402d5540eb\") " Nov 21 14:47:23 crc kubenswrapper[4897]: I1121 14:47:23.089685 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa6bb2ce-428b-4104-8823-99402d5540eb-utilities\") pod \"fa6bb2ce-428b-4104-8823-99402d5540eb\" (UID: \"fa6bb2ce-428b-4104-8823-99402d5540eb\") " Nov 21 14:47:23 crc kubenswrapper[4897]: I1121 14:47:23.090670 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fa6bb2ce-428b-4104-8823-99402d5540eb-utilities" (OuterVolumeSpecName: "utilities") pod "fa6bb2ce-428b-4104-8823-99402d5540eb" (UID: "fa6bb2ce-428b-4104-8823-99402d5540eb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:47:23 crc kubenswrapper[4897]: I1121 14:47:23.094980 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa6bb2ce-428b-4104-8823-99402d5540eb-kube-api-access-98vsv" (OuterVolumeSpecName: "kube-api-access-98vsv") pod "fa6bb2ce-428b-4104-8823-99402d5540eb" (UID: "fa6bb2ce-428b-4104-8823-99402d5540eb"). InnerVolumeSpecName "kube-api-access-98vsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:47:23 crc kubenswrapper[4897]: I1121 14:47:23.194341 4897 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa6bb2ce-428b-4104-8823-99402d5540eb-utilities\") on node \"crc\" DevicePath \"\"" Nov 21 14:47:23 crc kubenswrapper[4897]: I1121 14:47:23.194395 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-98vsv\" (UniqueName: \"kubernetes.io/projected/fa6bb2ce-428b-4104-8823-99402d5540eb-kube-api-access-98vsv\") on node \"crc\" DevicePath \"\"" Nov 21 14:47:23 crc kubenswrapper[4897]: I1121 14:47:23.239302 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fa6bb2ce-428b-4104-8823-99402d5540eb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fa6bb2ce-428b-4104-8823-99402d5540eb" (UID: "fa6bb2ce-428b-4104-8823-99402d5540eb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:47:23 crc kubenswrapper[4897]: I1121 14:47:23.296742 4897 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa6bb2ce-428b-4104-8823-99402d5540eb-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 21 14:47:23 crc kubenswrapper[4897]: I1121 14:47:23.956488 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7q2jl" Nov 21 14:47:23 crc kubenswrapper[4897]: I1121 14:47:23.983865 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7q2jl"] Nov 21 14:47:23 crc kubenswrapper[4897]: I1121 14:47:23.992363 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-7q2jl"] Nov 21 14:47:24 crc kubenswrapper[4897]: I1121 14:47:24.101518 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fa6bb2ce-428b-4104-8823-99402d5540eb" path="/var/lib/kubelet/pods/fa6bb2ce-428b-4104-8823-99402d5540eb/volumes" Nov 21 14:47:32 crc kubenswrapper[4897]: I1121 14:47:32.089355 4897 scope.go:117] "RemoveContainer" containerID="dc3c0a9877f70b7d5c55314a92df879b827a4103351566cc755cfbf916773aa2" Nov 21 14:47:32 crc kubenswrapper[4897]: E1121 14:47:32.090340 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 14:47:47 crc kubenswrapper[4897]: I1121 14:47:47.089447 4897 scope.go:117] "RemoveContainer" containerID="dc3c0a9877f70b7d5c55314a92df879b827a4103351566cc755cfbf916773aa2" Nov 21 14:47:47 crc kubenswrapper[4897]: E1121 14:47:47.094536 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 14:47:58 crc kubenswrapper[4897]: I1121 14:47:58.089335 4897 scope.go:117] "RemoveContainer" containerID="dc3c0a9877f70b7d5c55314a92df879b827a4103351566cc755cfbf916773aa2" Nov 21 14:47:58 crc kubenswrapper[4897]: E1121 14:47:58.090293 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 14:48:10 crc kubenswrapper[4897]: I1121 14:48:10.173793 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/speaker-hqq7g" podUID="89b1034c-5696-4975-a683-fdd9c163801f" containerName="speaker" probeResult="failure" output="Get \"http://localhost:29150/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 21 14:48:13 crc kubenswrapper[4897]: I1121 14:48:13.089241 4897 scope.go:117] "RemoveContainer" containerID="dc3c0a9877f70b7d5c55314a92df879b827a4103351566cc755cfbf916773aa2" Nov 21 14:48:13 crc kubenswrapper[4897]: E1121 14:48:13.090167 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 14:48:21 crc kubenswrapper[4897]: I1121 14:48:21.751342 4897 scope.go:117] "RemoveContainer" containerID="1ccf1e924cde0e2d7a1d25a5c7eb1b03623ad17b10a8a9291f26324d124b0f7d" Nov 21 14:48:27 crc kubenswrapper[4897]: I1121 14:48:27.089440 4897 scope.go:117] "RemoveContainer" containerID="dc3c0a9877f70b7d5c55314a92df879b827a4103351566cc755cfbf916773aa2" Nov 21 14:48:27 crc kubenswrapper[4897]: E1121 14:48:27.090421 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 14:48:36 crc kubenswrapper[4897]: I1121 14:48:36.044676 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-sync-m55sf"] Nov 21 14:48:36 crc kubenswrapper[4897]: I1121 14:48:36.055445 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-sync-m55sf"] Nov 21 14:48:36 crc kubenswrapper[4897]: I1121 14:48:36.102524 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c4c479c7-7728-41fc-bf98-ce2dcf800840" path="/var/lib/kubelet/pods/c4c479c7-7728-41fc-bf98-ce2dcf800840/volumes" Nov 21 14:48:40 crc kubenswrapper[4897]: I1121 14:48:40.089109 4897 scope.go:117] "RemoveContainer" containerID="dc3c0a9877f70b7d5c55314a92df879b827a4103351566cc755cfbf916773aa2" Nov 21 14:48:40 crc kubenswrapper[4897]: E1121 14:48:40.090213 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 14:48:51 crc kubenswrapper[4897]: I1121 14:48:51.089863 4897 scope.go:117] "RemoveContainer" containerID="dc3c0a9877f70b7d5c55314a92df879b827a4103351566cc755cfbf916773aa2" Nov 21 14:48:51 crc kubenswrapper[4897]: E1121 14:48:51.090695 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 14:49:06 crc kubenswrapper[4897]: I1121 14:49:06.089486 4897 scope.go:117] "RemoveContainer" containerID="dc3c0a9877f70b7d5c55314a92df879b827a4103351566cc755cfbf916773aa2" Nov 21 14:49:06 crc kubenswrapper[4897]: E1121 14:49:06.090356 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 14:49:20 crc kubenswrapper[4897]: I1121 14:49:20.089272 4897 scope.go:117] "RemoveContainer" containerID="dc3c0a9877f70b7d5c55314a92df879b827a4103351566cc755cfbf916773aa2" Nov 21 14:49:20 crc kubenswrapper[4897]: E1121 14:49:20.091312 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 14:49:21 crc kubenswrapper[4897]: I1121 14:49:21.867451 4897 scope.go:117] "RemoveContainer" containerID="c43eba824c1aabd68653c0fb2ebca80d3784c2763791bff3f6d120f3daaefcb6" Nov 21 14:49:31 crc kubenswrapper[4897]: I1121 14:49:31.089202 4897 scope.go:117] "RemoveContainer" containerID="dc3c0a9877f70b7d5c55314a92df879b827a4103351566cc755cfbf916773aa2" Nov 21 14:49:31 crc kubenswrapper[4897]: E1121 14:49:31.090078 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 14:49:43 crc kubenswrapper[4897]: I1121 14:49:43.090206 4897 scope.go:117] "RemoveContainer" containerID="dc3c0a9877f70b7d5c55314a92df879b827a4103351566cc755cfbf916773aa2" Nov 21 14:49:43 crc kubenswrapper[4897]: E1121 14:49:43.091540 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 14:49:57 crc kubenswrapper[4897]: I1121 14:49:57.089300 4897 scope.go:117] "RemoveContainer" containerID="dc3c0a9877f70b7d5c55314a92df879b827a4103351566cc755cfbf916773aa2" Nov 21 14:49:57 crc kubenswrapper[4897]: E1121 14:49:57.090323 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 14:50:11 crc kubenswrapper[4897]: I1121 14:50:11.089821 4897 scope.go:117] "RemoveContainer" containerID="dc3c0a9877f70b7d5c55314a92df879b827a4103351566cc755cfbf916773aa2" Nov 21 14:50:11 crc kubenswrapper[4897]: E1121 14:50:11.090942 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 14:50:26 crc kubenswrapper[4897]: I1121 14:50:26.089721 4897 scope.go:117] "RemoveContainer" containerID="dc3c0a9877f70b7d5c55314a92df879b827a4103351566cc755cfbf916773aa2" Nov 21 14:50:26 crc kubenswrapper[4897]: E1121 14:50:26.090545 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 14:50:38 crc kubenswrapper[4897]: I1121 14:50:38.089588 4897 scope.go:117] "RemoveContainer" containerID="dc3c0a9877f70b7d5c55314a92df879b827a4103351566cc755cfbf916773aa2" Nov 21 14:50:38 crc kubenswrapper[4897]: E1121 14:50:38.091522 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 14:50:39 crc kubenswrapper[4897]: I1121 14:50:39.076376 4897 generic.go:334] "Generic (PLEG): container finished" podID="d1a63d47-990b-40df-a70b-bdd3467bb792" containerID="9c4d0eb2399e2b76a1aef7e0060ff496f9f9c654f866a7c1a13909deaad6e399" exitCode=0 Nov 21 14:50:39 crc kubenswrapper[4897]: I1121 14:50:39.076472 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bqj2z" event={"ID":"d1a63d47-990b-40df-a70b-bdd3467bb792","Type":"ContainerDied","Data":"9c4d0eb2399e2b76a1aef7e0060ff496f9f9c654f866a7c1a13909deaad6e399"} Nov 21 14:50:40 crc kubenswrapper[4897]: I1121 14:50:40.614213 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bqj2z" Nov 21 14:50:40 crc kubenswrapper[4897]: I1121 14:50:40.650446 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d1a63d47-990b-40df-a70b-bdd3467bb792-inventory\") pod \"d1a63d47-990b-40df-a70b-bdd3467bb792\" (UID: \"d1a63d47-990b-40df-a70b-bdd3467bb792\") " Nov 21 14:50:40 crc kubenswrapper[4897]: I1121 14:50:40.650894 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d1a63d47-990b-40df-a70b-bdd3467bb792-ssh-key\") pod \"d1a63d47-990b-40df-a70b-bdd3467bb792\" (UID: \"d1a63d47-990b-40df-a70b-bdd3467bb792\") " Nov 21 14:50:40 crc kubenswrapper[4897]: I1121 14:50:40.651042 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-96bjq\" (UniqueName: \"kubernetes.io/projected/d1a63d47-990b-40df-a70b-bdd3467bb792-kube-api-access-96bjq\") pod \"d1a63d47-990b-40df-a70b-bdd3467bb792\" (UID: \"d1a63d47-990b-40df-a70b-bdd3467bb792\") " Nov 21 14:50:40 crc kubenswrapper[4897]: I1121 14:50:40.656619 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1a63d47-990b-40df-a70b-bdd3467bb792-kube-api-access-96bjq" (OuterVolumeSpecName: "kube-api-access-96bjq") pod "d1a63d47-990b-40df-a70b-bdd3467bb792" (UID: "d1a63d47-990b-40df-a70b-bdd3467bb792"). InnerVolumeSpecName "kube-api-access-96bjq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:50:40 crc kubenswrapper[4897]: I1121 14:50:40.685389 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1a63d47-990b-40df-a70b-bdd3467bb792-inventory" (OuterVolumeSpecName: "inventory") pod "d1a63d47-990b-40df-a70b-bdd3467bb792" (UID: "d1a63d47-990b-40df-a70b-bdd3467bb792"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:50:40 crc kubenswrapper[4897]: I1121 14:50:40.689748 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1a63d47-990b-40df-a70b-bdd3467bb792-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d1a63d47-990b-40df-a70b-bdd3467bb792" (UID: "d1a63d47-990b-40df-a70b-bdd3467bb792"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:50:40 crc kubenswrapper[4897]: I1121 14:50:40.753026 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-96bjq\" (UniqueName: \"kubernetes.io/projected/d1a63d47-990b-40df-a70b-bdd3467bb792-kube-api-access-96bjq\") on node \"crc\" DevicePath \"\"" Nov 21 14:50:40 crc kubenswrapper[4897]: I1121 14:50:40.753354 4897 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d1a63d47-990b-40df-a70b-bdd3467bb792-inventory\") on node \"crc\" DevicePath \"\"" Nov 21 14:50:40 crc kubenswrapper[4897]: I1121 14:50:40.753368 4897 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d1a63d47-990b-40df-a70b-bdd3467bb792-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 21 14:50:41 crc kubenswrapper[4897]: I1121 14:50:41.099233 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bqj2z" event={"ID":"d1a63d47-990b-40df-a70b-bdd3467bb792","Type":"ContainerDied","Data":"720c26537e69685d649602f1f28869ce05983204c91ff385fbb1c52527bc29f6"} Nov 21 14:50:41 crc kubenswrapper[4897]: I1121 14:50:41.099272 4897 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="720c26537e69685d649602f1f28869ce05983204c91ff385fbb1c52527bc29f6" Nov 21 14:50:41 crc kubenswrapper[4897]: I1121 14:50:41.099920 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bqj2z" Nov 21 14:50:41 crc kubenswrapper[4897]: I1121 14:50:41.476088 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2prl7"] Nov 21 14:50:41 crc kubenswrapper[4897]: E1121 14:50:41.476641 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa6bb2ce-428b-4104-8823-99402d5540eb" containerName="extract-content" Nov 21 14:50:41 crc kubenswrapper[4897]: I1121 14:50:41.476662 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa6bb2ce-428b-4104-8823-99402d5540eb" containerName="extract-content" Nov 21 14:50:41 crc kubenswrapper[4897]: E1121 14:50:41.476689 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1a63d47-990b-40df-a70b-bdd3467bb792" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Nov 21 14:50:41 crc kubenswrapper[4897]: I1121 14:50:41.476696 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1a63d47-990b-40df-a70b-bdd3467bb792" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Nov 21 14:50:41 crc kubenswrapper[4897]: E1121 14:50:41.476708 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa6bb2ce-428b-4104-8823-99402d5540eb" containerName="extract-utilities" Nov 21 14:50:41 crc kubenswrapper[4897]: I1121 14:50:41.476714 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa6bb2ce-428b-4104-8823-99402d5540eb" containerName="extract-utilities" Nov 21 14:50:41 crc kubenswrapper[4897]: E1121 14:50:41.476725 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa6bb2ce-428b-4104-8823-99402d5540eb" containerName="registry-server" Nov 21 14:50:41 crc kubenswrapper[4897]: I1121 14:50:41.476731 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa6bb2ce-428b-4104-8823-99402d5540eb" containerName="registry-server" Nov 21 14:50:41 crc kubenswrapper[4897]: I1121 14:50:41.476957 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa6bb2ce-428b-4104-8823-99402d5540eb" containerName="registry-server" Nov 21 14:50:41 crc kubenswrapper[4897]: I1121 14:50:41.476989 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1a63d47-990b-40df-a70b-bdd3467bb792" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Nov 21 14:50:41 crc kubenswrapper[4897]: I1121 14:50:41.477834 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2prl7" Nov 21 14:50:41 crc kubenswrapper[4897]: I1121 14:50:41.483609 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 21 14:50:41 crc kubenswrapper[4897]: I1121 14:50:41.483801 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-dj88m" Nov 21 14:50:41 crc kubenswrapper[4897]: I1121 14:50:41.483907 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 21 14:50:41 crc kubenswrapper[4897]: I1121 14:50:41.484068 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 21 14:50:41 crc kubenswrapper[4897]: I1121 14:50:41.487752 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2prl7"] Nov 21 14:50:41 crc kubenswrapper[4897]: I1121 14:50:41.674493 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b3fe592e-d817-4d38-8082-fe800be92147-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-2prl7\" (UID: \"b3fe592e-d817-4d38-8082-fe800be92147\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2prl7" Nov 21 14:50:41 crc kubenswrapper[4897]: I1121 14:50:41.674765 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sq9st\" (UniqueName: \"kubernetes.io/projected/b3fe592e-d817-4d38-8082-fe800be92147-kube-api-access-sq9st\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-2prl7\" (UID: \"b3fe592e-d817-4d38-8082-fe800be92147\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2prl7" Nov 21 14:50:41 crc kubenswrapper[4897]: I1121 14:50:41.675240 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b3fe592e-d817-4d38-8082-fe800be92147-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-2prl7\" (UID: \"b3fe592e-d817-4d38-8082-fe800be92147\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2prl7" Nov 21 14:50:41 crc kubenswrapper[4897]: I1121 14:50:41.779826 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b3fe592e-d817-4d38-8082-fe800be92147-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-2prl7\" (UID: \"b3fe592e-d817-4d38-8082-fe800be92147\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2prl7" Nov 21 14:50:41 crc kubenswrapper[4897]: I1121 14:50:41.780494 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b3fe592e-d817-4d38-8082-fe800be92147-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-2prl7\" (UID: \"b3fe592e-d817-4d38-8082-fe800be92147\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2prl7" Nov 21 14:50:41 crc kubenswrapper[4897]: I1121 14:50:41.780601 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sq9st\" (UniqueName: \"kubernetes.io/projected/b3fe592e-d817-4d38-8082-fe800be92147-kube-api-access-sq9st\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-2prl7\" (UID: \"b3fe592e-d817-4d38-8082-fe800be92147\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2prl7" Nov 21 14:50:41 crc kubenswrapper[4897]: I1121 14:50:41.785150 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b3fe592e-d817-4d38-8082-fe800be92147-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-2prl7\" (UID: \"b3fe592e-d817-4d38-8082-fe800be92147\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2prl7" Nov 21 14:50:41 crc kubenswrapper[4897]: I1121 14:50:41.785282 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b3fe592e-d817-4d38-8082-fe800be92147-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-2prl7\" (UID: \"b3fe592e-d817-4d38-8082-fe800be92147\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2prl7" Nov 21 14:50:41 crc kubenswrapper[4897]: I1121 14:50:41.803576 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sq9st\" (UniqueName: \"kubernetes.io/projected/b3fe592e-d817-4d38-8082-fe800be92147-kube-api-access-sq9st\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-2prl7\" (UID: \"b3fe592e-d817-4d38-8082-fe800be92147\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2prl7" Nov 21 14:50:41 crc kubenswrapper[4897]: I1121 14:50:41.837465 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2prl7" Nov 21 14:50:43 crc kubenswrapper[4897]: I1121 14:50:43.066018 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2prl7"] Nov 21 14:50:43 crc kubenswrapper[4897]: W1121 14:50:43.076615 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb3fe592e_d817_4d38_8082_fe800be92147.slice/crio-0ecd1b3dd69349a80ead581f50fc3d4cee8bd58a3091855c85ac0bf61a1ec830 WatchSource:0}: Error finding container 0ecd1b3dd69349a80ead581f50fc3d4cee8bd58a3091855c85ac0bf61a1ec830: Status 404 returned error can't find the container with id 0ecd1b3dd69349a80ead581f50fc3d4cee8bd58a3091855c85ac0bf61a1ec830 Nov 21 14:50:43 crc kubenswrapper[4897]: I1121 14:50:43.127706 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2prl7" event={"ID":"b3fe592e-d817-4d38-8082-fe800be92147","Type":"ContainerStarted","Data":"0ecd1b3dd69349a80ead581f50fc3d4cee8bd58a3091855c85ac0bf61a1ec830"} Nov 21 14:50:46 crc kubenswrapper[4897]: I1121 14:50:46.328848 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 21 14:50:47 crc kubenswrapper[4897]: I1121 14:50:47.172229 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2prl7" event={"ID":"b3fe592e-d817-4d38-8082-fe800be92147","Type":"ContainerStarted","Data":"11c606772bd58f64703e6275d28a02bc9e6c6ff3a42f24aba7761fb9771a1e38"} Nov 21 14:50:47 crc kubenswrapper[4897]: I1121 14:50:47.198134 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2prl7" podStartSLOduration=2.954629406 podStartE2EDuration="6.198119406s" podCreationTimestamp="2025-11-21 14:50:41 +0000 UTC" firstStartedPulling="2025-11-21 14:50:43.081857317 +0000 UTC m=+2520.366450802" lastFinishedPulling="2025-11-21 14:50:46.325347317 +0000 UTC m=+2523.609940802" observedRunningTime="2025-11-21 14:50:47.197233361 +0000 UTC m=+2524.481826836" watchObservedRunningTime="2025-11-21 14:50:47.198119406 +0000 UTC m=+2524.482712881" Nov 21 14:50:49 crc kubenswrapper[4897]: I1121 14:50:49.089465 4897 scope.go:117] "RemoveContainer" containerID="dc3c0a9877f70b7d5c55314a92df879b827a4103351566cc755cfbf916773aa2" Nov 21 14:50:49 crc kubenswrapper[4897]: E1121 14:50:49.090628 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 14:50:52 crc kubenswrapper[4897]: I1121 14:50:52.123841 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-kstjm"] Nov 21 14:50:52 crc kubenswrapper[4897]: I1121 14:50:52.129168 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kstjm" Nov 21 14:50:52 crc kubenswrapper[4897]: I1121 14:50:52.136241 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kstjm"] Nov 21 14:50:52 crc kubenswrapper[4897]: I1121 14:50:52.186280 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3de8f392-aca8-4f49-bd5c-b789ec0c7a0a-catalog-content\") pod \"redhat-marketplace-kstjm\" (UID: \"3de8f392-aca8-4f49-bd5c-b789ec0c7a0a\") " pod="openshift-marketplace/redhat-marketplace-kstjm" Nov 21 14:50:52 crc kubenswrapper[4897]: I1121 14:50:52.186345 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3de8f392-aca8-4f49-bd5c-b789ec0c7a0a-utilities\") pod \"redhat-marketplace-kstjm\" (UID: \"3de8f392-aca8-4f49-bd5c-b789ec0c7a0a\") " pod="openshift-marketplace/redhat-marketplace-kstjm" Nov 21 14:50:52 crc kubenswrapper[4897]: I1121 14:50:52.187895 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ldnxr\" (UniqueName: \"kubernetes.io/projected/3de8f392-aca8-4f49-bd5c-b789ec0c7a0a-kube-api-access-ldnxr\") pod \"redhat-marketplace-kstjm\" (UID: \"3de8f392-aca8-4f49-bd5c-b789ec0c7a0a\") " pod="openshift-marketplace/redhat-marketplace-kstjm" Nov 21 14:50:52 crc kubenswrapper[4897]: I1121 14:50:52.290793 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ldnxr\" (UniqueName: \"kubernetes.io/projected/3de8f392-aca8-4f49-bd5c-b789ec0c7a0a-kube-api-access-ldnxr\") pod \"redhat-marketplace-kstjm\" (UID: \"3de8f392-aca8-4f49-bd5c-b789ec0c7a0a\") " pod="openshift-marketplace/redhat-marketplace-kstjm" Nov 21 14:50:52 crc kubenswrapper[4897]: I1121 14:50:52.290934 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3de8f392-aca8-4f49-bd5c-b789ec0c7a0a-catalog-content\") pod \"redhat-marketplace-kstjm\" (UID: \"3de8f392-aca8-4f49-bd5c-b789ec0c7a0a\") " pod="openshift-marketplace/redhat-marketplace-kstjm" Nov 21 14:50:52 crc kubenswrapper[4897]: I1121 14:50:52.291005 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3de8f392-aca8-4f49-bd5c-b789ec0c7a0a-utilities\") pod \"redhat-marketplace-kstjm\" (UID: \"3de8f392-aca8-4f49-bd5c-b789ec0c7a0a\") " pod="openshift-marketplace/redhat-marketplace-kstjm" Nov 21 14:50:52 crc kubenswrapper[4897]: I1121 14:50:52.291437 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3de8f392-aca8-4f49-bd5c-b789ec0c7a0a-catalog-content\") pod \"redhat-marketplace-kstjm\" (UID: \"3de8f392-aca8-4f49-bd5c-b789ec0c7a0a\") " pod="openshift-marketplace/redhat-marketplace-kstjm" Nov 21 14:50:52 crc kubenswrapper[4897]: I1121 14:50:52.291604 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3de8f392-aca8-4f49-bd5c-b789ec0c7a0a-utilities\") pod \"redhat-marketplace-kstjm\" (UID: \"3de8f392-aca8-4f49-bd5c-b789ec0c7a0a\") " pod="openshift-marketplace/redhat-marketplace-kstjm" Nov 21 14:50:52 crc kubenswrapper[4897]: I1121 14:50:52.310754 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ldnxr\" (UniqueName: \"kubernetes.io/projected/3de8f392-aca8-4f49-bd5c-b789ec0c7a0a-kube-api-access-ldnxr\") pod \"redhat-marketplace-kstjm\" (UID: \"3de8f392-aca8-4f49-bd5c-b789ec0c7a0a\") " pod="openshift-marketplace/redhat-marketplace-kstjm" Nov 21 14:50:52 crc kubenswrapper[4897]: I1121 14:50:52.452617 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kstjm" Nov 21 14:50:52 crc kubenswrapper[4897]: I1121 14:50:52.911969 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kstjm"] Nov 21 14:50:53 crc kubenswrapper[4897]: I1121 14:50:53.240071 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kstjm" event={"ID":"3de8f392-aca8-4f49-bd5c-b789ec0c7a0a","Type":"ContainerStarted","Data":"9470c8dbf3e937dc9db0d4f050aef89ff547e399c4de8cbb4a9997d7fa49d632"} Nov 21 14:50:53 crc kubenswrapper[4897]: I1121 14:50:53.240413 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kstjm" event={"ID":"3de8f392-aca8-4f49-bd5c-b789ec0c7a0a","Type":"ContainerStarted","Data":"ef19f10b72c6bc2d456c4400d0dca970bdcc7bc63d890aa13184675006c132b4"} Nov 21 14:50:54 crc kubenswrapper[4897]: I1121 14:50:54.251087 4897 generic.go:334] "Generic (PLEG): container finished" podID="3de8f392-aca8-4f49-bd5c-b789ec0c7a0a" containerID="9470c8dbf3e937dc9db0d4f050aef89ff547e399c4de8cbb4a9997d7fa49d632" exitCode=0 Nov 21 14:50:54 crc kubenswrapper[4897]: I1121 14:50:54.251134 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kstjm" event={"ID":"3de8f392-aca8-4f49-bd5c-b789ec0c7a0a","Type":"ContainerDied","Data":"9470c8dbf3e937dc9db0d4f050aef89ff547e399c4de8cbb4a9997d7fa49d632"} Nov 21 14:50:54 crc kubenswrapper[4897]: I1121 14:50:54.253932 4897 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 21 14:50:57 crc kubenswrapper[4897]: I1121 14:50:57.284622 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kstjm" event={"ID":"3de8f392-aca8-4f49-bd5c-b789ec0c7a0a","Type":"ContainerStarted","Data":"0350f76a174d96729d18a4de3bff7b243203b8351e149e83b86222fd17dca0ed"} Nov 21 14:50:59 crc kubenswrapper[4897]: I1121 14:50:59.311628 4897 generic.go:334] "Generic (PLEG): container finished" podID="3de8f392-aca8-4f49-bd5c-b789ec0c7a0a" containerID="0350f76a174d96729d18a4de3bff7b243203b8351e149e83b86222fd17dca0ed" exitCode=0 Nov 21 14:50:59 crc kubenswrapper[4897]: I1121 14:50:59.311690 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kstjm" event={"ID":"3de8f392-aca8-4f49-bd5c-b789ec0c7a0a","Type":"ContainerDied","Data":"0350f76a174d96729d18a4de3bff7b243203b8351e149e83b86222fd17dca0ed"} Nov 21 14:51:01 crc kubenswrapper[4897]: I1121 14:51:01.337858 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kstjm" event={"ID":"3de8f392-aca8-4f49-bd5c-b789ec0c7a0a","Type":"ContainerStarted","Data":"6384a723f6fc4dff1af702909de34b8bf5ec504fb07aefa07ce73817caa71f00"} Nov 21 14:51:01 crc kubenswrapper[4897]: I1121 14:51:01.366582 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-kstjm" podStartSLOduration=3.258911276 podStartE2EDuration="9.366563233s" podCreationTimestamp="2025-11-21 14:50:52 +0000 UTC" firstStartedPulling="2025-11-21 14:50:54.253659393 +0000 UTC m=+2531.538252868" lastFinishedPulling="2025-11-21 14:51:00.36131134 +0000 UTC m=+2537.645904825" observedRunningTime="2025-11-21 14:51:01.357661312 +0000 UTC m=+2538.642254797" watchObservedRunningTime="2025-11-21 14:51:01.366563233 +0000 UTC m=+2538.651156708" Nov 21 14:51:02 crc kubenswrapper[4897]: I1121 14:51:02.452981 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-kstjm" Nov 21 14:51:02 crc kubenswrapper[4897]: I1121 14:51:02.453368 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-kstjm" Nov 21 14:51:02 crc kubenswrapper[4897]: I1121 14:51:02.527382 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-kstjm" Nov 21 14:51:04 crc kubenswrapper[4897]: I1121 14:51:04.101920 4897 scope.go:117] "RemoveContainer" containerID="dc3c0a9877f70b7d5c55314a92df879b827a4103351566cc755cfbf916773aa2" Nov 21 14:51:04 crc kubenswrapper[4897]: E1121 14:51:04.102769 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 14:51:12 crc kubenswrapper[4897]: I1121 14:51:12.511414 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-kstjm" Nov 21 14:51:12 crc kubenswrapper[4897]: I1121 14:51:12.566622 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-kstjm"] Nov 21 14:51:13 crc kubenswrapper[4897]: I1121 14:51:13.465709 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-kstjm" podUID="3de8f392-aca8-4f49-bd5c-b789ec0c7a0a" containerName="registry-server" containerID="cri-o://6384a723f6fc4dff1af702909de34b8bf5ec504fb07aefa07ce73817caa71f00" gracePeriod=2 Nov 21 14:51:14 crc kubenswrapper[4897]: I1121 14:51:14.022867 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kstjm" Nov 21 14:51:14 crc kubenswrapper[4897]: I1121 14:51:14.232041 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ldnxr\" (UniqueName: \"kubernetes.io/projected/3de8f392-aca8-4f49-bd5c-b789ec0c7a0a-kube-api-access-ldnxr\") pod \"3de8f392-aca8-4f49-bd5c-b789ec0c7a0a\" (UID: \"3de8f392-aca8-4f49-bd5c-b789ec0c7a0a\") " Nov 21 14:51:14 crc kubenswrapper[4897]: I1121 14:51:14.232227 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3de8f392-aca8-4f49-bd5c-b789ec0c7a0a-utilities\") pod \"3de8f392-aca8-4f49-bd5c-b789ec0c7a0a\" (UID: \"3de8f392-aca8-4f49-bd5c-b789ec0c7a0a\") " Nov 21 14:51:14 crc kubenswrapper[4897]: I1121 14:51:14.232316 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3de8f392-aca8-4f49-bd5c-b789ec0c7a0a-catalog-content\") pod \"3de8f392-aca8-4f49-bd5c-b789ec0c7a0a\" (UID: \"3de8f392-aca8-4f49-bd5c-b789ec0c7a0a\") " Nov 21 14:51:14 crc kubenswrapper[4897]: I1121 14:51:14.233578 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3de8f392-aca8-4f49-bd5c-b789ec0c7a0a-utilities" (OuterVolumeSpecName: "utilities") pod "3de8f392-aca8-4f49-bd5c-b789ec0c7a0a" (UID: "3de8f392-aca8-4f49-bd5c-b789ec0c7a0a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:51:14 crc kubenswrapper[4897]: I1121 14:51:14.246776 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3de8f392-aca8-4f49-bd5c-b789ec0c7a0a-kube-api-access-ldnxr" (OuterVolumeSpecName: "kube-api-access-ldnxr") pod "3de8f392-aca8-4f49-bd5c-b789ec0c7a0a" (UID: "3de8f392-aca8-4f49-bd5c-b789ec0c7a0a"). InnerVolumeSpecName "kube-api-access-ldnxr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:51:14 crc kubenswrapper[4897]: I1121 14:51:14.249577 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3de8f392-aca8-4f49-bd5c-b789ec0c7a0a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3de8f392-aca8-4f49-bd5c-b789ec0c7a0a" (UID: "3de8f392-aca8-4f49-bd5c-b789ec0c7a0a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:51:14 crc kubenswrapper[4897]: I1121 14:51:14.335047 4897 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3de8f392-aca8-4f49-bd5c-b789ec0c7a0a-utilities\") on node \"crc\" DevicePath \"\"" Nov 21 14:51:14 crc kubenswrapper[4897]: I1121 14:51:14.335310 4897 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3de8f392-aca8-4f49-bd5c-b789ec0c7a0a-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 21 14:51:14 crc kubenswrapper[4897]: I1121 14:51:14.335323 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ldnxr\" (UniqueName: \"kubernetes.io/projected/3de8f392-aca8-4f49-bd5c-b789ec0c7a0a-kube-api-access-ldnxr\") on node \"crc\" DevicePath \"\"" Nov 21 14:51:14 crc kubenswrapper[4897]: I1121 14:51:14.477617 4897 generic.go:334] "Generic (PLEG): container finished" podID="3de8f392-aca8-4f49-bd5c-b789ec0c7a0a" containerID="6384a723f6fc4dff1af702909de34b8bf5ec504fb07aefa07ce73817caa71f00" exitCode=0 Nov 21 14:51:14 crc kubenswrapper[4897]: I1121 14:51:14.477700 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kstjm" event={"ID":"3de8f392-aca8-4f49-bd5c-b789ec0c7a0a","Type":"ContainerDied","Data":"6384a723f6fc4dff1af702909de34b8bf5ec504fb07aefa07ce73817caa71f00"} Nov 21 14:51:14 crc kubenswrapper[4897]: I1121 14:51:14.477778 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kstjm" event={"ID":"3de8f392-aca8-4f49-bd5c-b789ec0c7a0a","Type":"ContainerDied","Data":"ef19f10b72c6bc2d456c4400d0dca970bdcc7bc63d890aa13184675006c132b4"} Nov 21 14:51:14 crc kubenswrapper[4897]: I1121 14:51:14.477795 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kstjm" Nov 21 14:51:14 crc kubenswrapper[4897]: I1121 14:51:14.477801 4897 scope.go:117] "RemoveContainer" containerID="6384a723f6fc4dff1af702909de34b8bf5ec504fb07aefa07ce73817caa71f00" Nov 21 14:51:14 crc kubenswrapper[4897]: I1121 14:51:14.514019 4897 scope.go:117] "RemoveContainer" containerID="0350f76a174d96729d18a4de3bff7b243203b8351e149e83b86222fd17dca0ed" Nov 21 14:51:14 crc kubenswrapper[4897]: I1121 14:51:14.534566 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-kstjm"] Nov 21 14:51:14 crc kubenswrapper[4897]: I1121 14:51:14.545113 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-kstjm"] Nov 21 14:51:14 crc kubenswrapper[4897]: I1121 14:51:14.548036 4897 scope.go:117] "RemoveContainer" containerID="9470c8dbf3e937dc9db0d4f050aef89ff547e399c4de8cbb4a9997d7fa49d632" Nov 21 14:51:14 crc kubenswrapper[4897]: I1121 14:51:14.612950 4897 scope.go:117] "RemoveContainer" containerID="6384a723f6fc4dff1af702909de34b8bf5ec504fb07aefa07ce73817caa71f00" Nov 21 14:51:14 crc kubenswrapper[4897]: E1121 14:51:14.613549 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6384a723f6fc4dff1af702909de34b8bf5ec504fb07aefa07ce73817caa71f00\": container with ID starting with 6384a723f6fc4dff1af702909de34b8bf5ec504fb07aefa07ce73817caa71f00 not found: ID does not exist" containerID="6384a723f6fc4dff1af702909de34b8bf5ec504fb07aefa07ce73817caa71f00" Nov 21 14:51:14 crc kubenswrapper[4897]: I1121 14:51:14.613624 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6384a723f6fc4dff1af702909de34b8bf5ec504fb07aefa07ce73817caa71f00"} err="failed to get container status \"6384a723f6fc4dff1af702909de34b8bf5ec504fb07aefa07ce73817caa71f00\": rpc error: code = NotFound desc = could not find container \"6384a723f6fc4dff1af702909de34b8bf5ec504fb07aefa07ce73817caa71f00\": container with ID starting with 6384a723f6fc4dff1af702909de34b8bf5ec504fb07aefa07ce73817caa71f00 not found: ID does not exist" Nov 21 14:51:14 crc kubenswrapper[4897]: I1121 14:51:14.613669 4897 scope.go:117] "RemoveContainer" containerID="0350f76a174d96729d18a4de3bff7b243203b8351e149e83b86222fd17dca0ed" Nov 21 14:51:14 crc kubenswrapper[4897]: E1121 14:51:14.614063 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0350f76a174d96729d18a4de3bff7b243203b8351e149e83b86222fd17dca0ed\": container with ID starting with 0350f76a174d96729d18a4de3bff7b243203b8351e149e83b86222fd17dca0ed not found: ID does not exist" containerID="0350f76a174d96729d18a4de3bff7b243203b8351e149e83b86222fd17dca0ed" Nov 21 14:51:14 crc kubenswrapper[4897]: I1121 14:51:14.614103 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0350f76a174d96729d18a4de3bff7b243203b8351e149e83b86222fd17dca0ed"} err="failed to get container status \"0350f76a174d96729d18a4de3bff7b243203b8351e149e83b86222fd17dca0ed\": rpc error: code = NotFound desc = could not find container \"0350f76a174d96729d18a4de3bff7b243203b8351e149e83b86222fd17dca0ed\": container with ID starting with 0350f76a174d96729d18a4de3bff7b243203b8351e149e83b86222fd17dca0ed not found: ID does not exist" Nov 21 14:51:14 crc kubenswrapper[4897]: I1121 14:51:14.614132 4897 scope.go:117] "RemoveContainer" containerID="9470c8dbf3e937dc9db0d4f050aef89ff547e399c4de8cbb4a9997d7fa49d632" Nov 21 14:51:14 crc kubenswrapper[4897]: E1121 14:51:14.614547 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9470c8dbf3e937dc9db0d4f050aef89ff547e399c4de8cbb4a9997d7fa49d632\": container with ID starting with 9470c8dbf3e937dc9db0d4f050aef89ff547e399c4de8cbb4a9997d7fa49d632 not found: ID does not exist" containerID="9470c8dbf3e937dc9db0d4f050aef89ff547e399c4de8cbb4a9997d7fa49d632" Nov 21 14:51:14 crc kubenswrapper[4897]: I1121 14:51:14.614578 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9470c8dbf3e937dc9db0d4f050aef89ff547e399c4de8cbb4a9997d7fa49d632"} err="failed to get container status \"9470c8dbf3e937dc9db0d4f050aef89ff547e399c4de8cbb4a9997d7fa49d632\": rpc error: code = NotFound desc = could not find container \"9470c8dbf3e937dc9db0d4f050aef89ff547e399c4de8cbb4a9997d7fa49d632\": container with ID starting with 9470c8dbf3e937dc9db0d4f050aef89ff547e399c4de8cbb4a9997d7fa49d632 not found: ID does not exist" Nov 21 14:51:16 crc kubenswrapper[4897]: I1121 14:51:16.091766 4897 scope.go:117] "RemoveContainer" containerID="dc3c0a9877f70b7d5c55314a92df879b827a4103351566cc755cfbf916773aa2" Nov 21 14:51:16 crc kubenswrapper[4897]: E1121 14:51:16.092248 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 14:51:16 crc kubenswrapper[4897]: I1121 14:51:16.103928 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3de8f392-aca8-4f49-bd5c-b789ec0c7a0a" path="/var/lib/kubelet/pods/3de8f392-aca8-4f49-bd5c-b789ec0c7a0a/volumes" Nov 21 14:51:26 crc kubenswrapper[4897]: I1121 14:51:26.866419 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-kmlvv"] Nov 21 14:51:26 crc kubenswrapper[4897]: E1121 14:51:26.867845 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3de8f392-aca8-4f49-bd5c-b789ec0c7a0a" containerName="extract-utilities" Nov 21 14:51:26 crc kubenswrapper[4897]: I1121 14:51:26.867885 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="3de8f392-aca8-4f49-bd5c-b789ec0c7a0a" containerName="extract-utilities" Nov 21 14:51:26 crc kubenswrapper[4897]: E1121 14:51:26.867911 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3de8f392-aca8-4f49-bd5c-b789ec0c7a0a" containerName="registry-server" Nov 21 14:51:26 crc kubenswrapper[4897]: I1121 14:51:26.867917 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="3de8f392-aca8-4f49-bd5c-b789ec0c7a0a" containerName="registry-server" Nov 21 14:51:26 crc kubenswrapper[4897]: E1121 14:51:26.867945 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3de8f392-aca8-4f49-bd5c-b789ec0c7a0a" containerName="extract-content" Nov 21 14:51:26 crc kubenswrapper[4897]: I1121 14:51:26.867952 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="3de8f392-aca8-4f49-bd5c-b789ec0c7a0a" containerName="extract-content" Nov 21 14:51:26 crc kubenswrapper[4897]: I1121 14:51:26.868170 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="3de8f392-aca8-4f49-bd5c-b789ec0c7a0a" containerName="registry-server" Nov 21 14:51:26 crc kubenswrapper[4897]: I1121 14:51:26.870084 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kmlvv" Nov 21 14:51:26 crc kubenswrapper[4897]: I1121 14:51:26.890672 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kmlvv"] Nov 21 14:51:27 crc kubenswrapper[4897]: I1121 14:51:27.053748 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mptz9\" (UniqueName: \"kubernetes.io/projected/7b468737-2893-4083-a91e-6c7b7d8d5d5a-kube-api-access-mptz9\") pod \"community-operators-kmlvv\" (UID: \"7b468737-2893-4083-a91e-6c7b7d8d5d5a\") " pod="openshift-marketplace/community-operators-kmlvv" Nov 21 14:51:27 crc kubenswrapper[4897]: I1121 14:51:27.054268 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b468737-2893-4083-a91e-6c7b7d8d5d5a-catalog-content\") pod \"community-operators-kmlvv\" (UID: \"7b468737-2893-4083-a91e-6c7b7d8d5d5a\") " pod="openshift-marketplace/community-operators-kmlvv" Nov 21 14:51:27 crc kubenswrapper[4897]: I1121 14:51:27.054380 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b468737-2893-4083-a91e-6c7b7d8d5d5a-utilities\") pod \"community-operators-kmlvv\" (UID: \"7b468737-2893-4083-a91e-6c7b7d8d5d5a\") " pod="openshift-marketplace/community-operators-kmlvv" Nov 21 14:51:27 crc kubenswrapper[4897]: I1121 14:51:27.158375 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mptz9\" (UniqueName: \"kubernetes.io/projected/7b468737-2893-4083-a91e-6c7b7d8d5d5a-kube-api-access-mptz9\") pod \"community-operators-kmlvv\" (UID: \"7b468737-2893-4083-a91e-6c7b7d8d5d5a\") " pod="openshift-marketplace/community-operators-kmlvv" Nov 21 14:51:27 crc kubenswrapper[4897]: I1121 14:51:27.158612 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b468737-2893-4083-a91e-6c7b7d8d5d5a-catalog-content\") pod \"community-operators-kmlvv\" (UID: \"7b468737-2893-4083-a91e-6c7b7d8d5d5a\") " pod="openshift-marketplace/community-operators-kmlvv" Nov 21 14:51:27 crc kubenswrapper[4897]: I1121 14:51:27.158647 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b468737-2893-4083-a91e-6c7b7d8d5d5a-utilities\") pod \"community-operators-kmlvv\" (UID: \"7b468737-2893-4083-a91e-6c7b7d8d5d5a\") " pod="openshift-marketplace/community-operators-kmlvv" Nov 21 14:51:27 crc kubenswrapper[4897]: I1121 14:51:27.159346 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b468737-2893-4083-a91e-6c7b7d8d5d5a-catalog-content\") pod \"community-operators-kmlvv\" (UID: \"7b468737-2893-4083-a91e-6c7b7d8d5d5a\") " pod="openshift-marketplace/community-operators-kmlvv" Nov 21 14:51:27 crc kubenswrapper[4897]: I1121 14:51:27.159456 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b468737-2893-4083-a91e-6c7b7d8d5d5a-utilities\") pod \"community-operators-kmlvv\" (UID: \"7b468737-2893-4083-a91e-6c7b7d8d5d5a\") " pod="openshift-marketplace/community-operators-kmlvv" Nov 21 14:51:27 crc kubenswrapper[4897]: I1121 14:51:27.195936 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mptz9\" (UniqueName: \"kubernetes.io/projected/7b468737-2893-4083-a91e-6c7b7d8d5d5a-kube-api-access-mptz9\") pod \"community-operators-kmlvv\" (UID: \"7b468737-2893-4083-a91e-6c7b7d8d5d5a\") " pod="openshift-marketplace/community-operators-kmlvv" Nov 21 14:51:27 crc kubenswrapper[4897]: I1121 14:51:27.494289 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kmlvv" Nov 21 14:51:28 crc kubenswrapper[4897]: I1121 14:51:28.045628 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kmlvv"] Nov 21 14:51:28 crc kubenswrapper[4897]: I1121 14:51:28.631016 4897 generic.go:334] "Generic (PLEG): container finished" podID="7b468737-2893-4083-a91e-6c7b7d8d5d5a" containerID="e51c8fa96208ad4651eb0b355cab685ffe76d8151a936c74fb22731e2b2424b4" exitCode=0 Nov 21 14:51:28 crc kubenswrapper[4897]: I1121 14:51:28.631206 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kmlvv" event={"ID":"7b468737-2893-4083-a91e-6c7b7d8d5d5a","Type":"ContainerDied","Data":"e51c8fa96208ad4651eb0b355cab685ffe76d8151a936c74fb22731e2b2424b4"} Nov 21 14:51:28 crc kubenswrapper[4897]: I1121 14:51:28.631892 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kmlvv" event={"ID":"7b468737-2893-4083-a91e-6c7b7d8d5d5a","Type":"ContainerStarted","Data":"bca4a841486779c5286b06defaa8ea27c2c73a1d4be91ab8d447210741c0d075"} Nov 21 14:51:29 crc kubenswrapper[4897]: I1121 14:51:29.647225 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kmlvv" event={"ID":"7b468737-2893-4083-a91e-6c7b7d8d5d5a","Type":"ContainerStarted","Data":"0fde6d0a071919275a2d6af817625c0d11f3636008b39dbbe5c88c3742da41a2"} Nov 21 14:51:31 crc kubenswrapper[4897]: I1121 14:51:31.088900 4897 scope.go:117] "RemoveContainer" containerID="dc3c0a9877f70b7d5c55314a92df879b827a4103351566cc755cfbf916773aa2" Nov 21 14:51:31 crc kubenswrapper[4897]: E1121 14:51:31.089487 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 14:51:31 crc kubenswrapper[4897]: I1121 14:51:31.667409 4897 generic.go:334] "Generic (PLEG): container finished" podID="7b468737-2893-4083-a91e-6c7b7d8d5d5a" containerID="0fde6d0a071919275a2d6af817625c0d11f3636008b39dbbe5c88c3742da41a2" exitCode=0 Nov 21 14:51:31 crc kubenswrapper[4897]: I1121 14:51:31.667453 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kmlvv" event={"ID":"7b468737-2893-4083-a91e-6c7b7d8d5d5a","Type":"ContainerDied","Data":"0fde6d0a071919275a2d6af817625c0d11f3636008b39dbbe5c88c3742da41a2"} Nov 21 14:51:32 crc kubenswrapper[4897]: I1121 14:51:32.679774 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kmlvv" event={"ID":"7b468737-2893-4083-a91e-6c7b7d8d5d5a","Type":"ContainerStarted","Data":"4feff54089f1e135d7b4b952da8c002ef83367fc415b372d57e35f74d16df637"} Nov 21 14:51:32 crc kubenswrapper[4897]: I1121 14:51:32.710806 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-kmlvv" podStartSLOduration=3.058963607 podStartE2EDuration="6.710780908s" podCreationTimestamp="2025-11-21 14:51:26 +0000 UTC" firstStartedPulling="2025-11-21 14:51:28.634524562 +0000 UTC m=+2565.919118037" lastFinishedPulling="2025-11-21 14:51:32.286341843 +0000 UTC m=+2569.570935338" observedRunningTime="2025-11-21 14:51:32.701642591 +0000 UTC m=+2569.986236076" watchObservedRunningTime="2025-11-21 14:51:32.710780908 +0000 UTC m=+2569.995374383" Nov 21 14:51:37 crc kubenswrapper[4897]: I1121 14:51:37.495207 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-kmlvv" Nov 21 14:51:37 crc kubenswrapper[4897]: I1121 14:51:37.495912 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-kmlvv" Nov 21 14:51:37 crc kubenswrapper[4897]: I1121 14:51:37.588326 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-kmlvv" Nov 21 14:51:37 crc kubenswrapper[4897]: I1121 14:51:37.792432 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-kmlvv" Nov 21 14:51:37 crc kubenswrapper[4897]: I1121 14:51:37.871230 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-kmlvv"] Nov 21 14:51:39 crc kubenswrapper[4897]: I1121 14:51:39.744061 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-kmlvv" podUID="7b468737-2893-4083-a91e-6c7b7d8d5d5a" containerName="registry-server" containerID="cri-o://4feff54089f1e135d7b4b952da8c002ef83367fc415b372d57e35f74d16df637" gracePeriod=2 Nov 21 14:51:40 crc kubenswrapper[4897]: I1121 14:51:40.756364 4897 generic.go:334] "Generic (PLEG): container finished" podID="7b468737-2893-4083-a91e-6c7b7d8d5d5a" containerID="4feff54089f1e135d7b4b952da8c002ef83367fc415b372d57e35f74d16df637" exitCode=0 Nov 21 14:51:40 crc kubenswrapper[4897]: I1121 14:51:40.756432 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kmlvv" event={"ID":"7b468737-2893-4083-a91e-6c7b7d8d5d5a","Type":"ContainerDied","Data":"4feff54089f1e135d7b4b952da8c002ef83367fc415b372d57e35f74d16df637"} Nov 21 14:51:41 crc kubenswrapper[4897]: I1121 14:51:41.411194 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kmlvv" Nov 21 14:51:41 crc kubenswrapper[4897]: I1121 14:51:41.533823 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b468737-2893-4083-a91e-6c7b7d8d5d5a-utilities\") pod \"7b468737-2893-4083-a91e-6c7b7d8d5d5a\" (UID: \"7b468737-2893-4083-a91e-6c7b7d8d5d5a\") " Nov 21 14:51:41 crc kubenswrapper[4897]: I1121 14:51:41.534021 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mptz9\" (UniqueName: \"kubernetes.io/projected/7b468737-2893-4083-a91e-6c7b7d8d5d5a-kube-api-access-mptz9\") pod \"7b468737-2893-4083-a91e-6c7b7d8d5d5a\" (UID: \"7b468737-2893-4083-a91e-6c7b7d8d5d5a\") " Nov 21 14:51:41 crc kubenswrapper[4897]: I1121 14:51:41.534113 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b468737-2893-4083-a91e-6c7b7d8d5d5a-catalog-content\") pod \"7b468737-2893-4083-a91e-6c7b7d8d5d5a\" (UID: \"7b468737-2893-4083-a91e-6c7b7d8d5d5a\") " Nov 21 14:51:41 crc kubenswrapper[4897]: I1121 14:51:41.535008 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7b468737-2893-4083-a91e-6c7b7d8d5d5a-utilities" (OuterVolumeSpecName: "utilities") pod "7b468737-2893-4083-a91e-6c7b7d8d5d5a" (UID: "7b468737-2893-4083-a91e-6c7b7d8d5d5a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:51:41 crc kubenswrapper[4897]: I1121 14:51:41.541072 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b468737-2893-4083-a91e-6c7b7d8d5d5a-kube-api-access-mptz9" (OuterVolumeSpecName: "kube-api-access-mptz9") pod "7b468737-2893-4083-a91e-6c7b7d8d5d5a" (UID: "7b468737-2893-4083-a91e-6c7b7d8d5d5a"). InnerVolumeSpecName "kube-api-access-mptz9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:51:41 crc kubenswrapper[4897]: I1121 14:51:41.586529 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7b468737-2893-4083-a91e-6c7b7d8d5d5a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7b468737-2893-4083-a91e-6c7b7d8d5d5a" (UID: "7b468737-2893-4083-a91e-6c7b7d8d5d5a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:51:41 crc kubenswrapper[4897]: I1121 14:51:41.637297 4897 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b468737-2893-4083-a91e-6c7b7d8d5d5a-utilities\") on node \"crc\" DevicePath \"\"" Nov 21 14:51:41 crc kubenswrapper[4897]: I1121 14:51:41.637625 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mptz9\" (UniqueName: \"kubernetes.io/projected/7b468737-2893-4083-a91e-6c7b7d8d5d5a-kube-api-access-mptz9\") on node \"crc\" DevicePath \"\"" Nov 21 14:51:41 crc kubenswrapper[4897]: I1121 14:51:41.637719 4897 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b468737-2893-4083-a91e-6c7b7d8d5d5a-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 21 14:51:41 crc kubenswrapper[4897]: I1121 14:51:41.779910 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kmlvv" event={"ID":"7b468737-2893-4083-a91e-6c7b7d8d5d5a","Type":"ContainerDied","Data":"bca4a841486779c5286b06defaa8ea27c2c73a1d4be91ab8d447210741c0d075"} Nov 21 14:51:41 crc kubenswrapper[4897]: I1121 14:51:41.779968 4897 scope.go:117] "RemoveContainer" containerID="4feff54089f1e135d7b4b952da8c002ef83367fc415b372d57e35f74d16df637" Nov 21 14:51:41 crc kubenswrapper[4897]: I1121 14:51:41.779963 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kmlvv" Nov 21 14:51:41 crc kubenswrapper[4897]: I1121 14:51:41.813486 4897 scope.go:117] "RemoveContainer" containerID="0fde6d0a071919275a2d6af817625c0d11f3636008b39dbbe5c88c3742da41a2" Nov 21 14:51:41 crc kubenswrapper[4897]: I1121 14:51:41.814114 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-kmlvv"] Nov 21 14:51:41 crc kubenswrapper[4897]: I1121 14:51:41.825397 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-kmlvv"] Nov 21 14:51:41 crc kubenswrapper[4897]: I1121 14:51:41.839850 4897 scope.go:117] "RemoveContainer" containerID="e51c8fa96208ad4651eb0b355cab685ffe76d8151a936c74fb22731e2b2424b4" Nov 21 14:51:42 crc kubenswrapper[4897]: I1121 14:51:42.104746 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7b468737-2893-4083-a91e-6c7b7d8d5d5a" path="/var/lib/kubelet/pods/7b468737-2893-4083-a91e-6c7b7d8d5d5a/volumes" Nov 21 14:51:46 crc kubenswrapper[4897]: I1121 14:51:46.089744 4897 scope.go:117] "RemoveContainer" containerID="dc3c0a9877f70b7d5c55314a92df879b827a4103351566cc755cfbf916773aa2" Nov 21 14:51:46 crc kubenswrapper[4897]: I1121 14:51:46.840396 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" event={"ID":"e7670227-d280-4847-b882-754429f56b0f","Type":"ContainerStarted","Data":"93211f44a4c3f56e3048bcba270e9ae8a65d8435c6fcfa8de9ac51f949ca9ee4"} Nov 21 14:51:53 crc kubenswrapper[4897]: I1121 14:51:53.039661 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-sync-slsml"] Nov 21 14:51:53 crc kubenswrapper[4897]: I1121 14:51:53.050990 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-sync-slsml"] Nov 21 14:51:54 crc kubenswrapper[4897]: I1121 14:51:54.102319 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f9422887-78a4-47c5-99ed-89f4ed9a97e9" path="/var/lib/kubelet/pods/f9422887-78a4-47c5-99ed-89f4ed9a97e9/volumes" Nov 21 14:52:21 crc kubenswrapper[4897]: I1121 14:52:21.195015 4897 generic.go:334] "Generic (PLEG): container finished" podID="b3fe592e-d817-4d38-8082-fe800be92147" containerID="11c606772bd58f64703e6275d28a02bc9e6c6ff3a42f24aba7761fb9771a1e38" exitCode=0 Nov 21 14:52:21 crc kubenswrapper[4897]: I1121 14:52:21.195448 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2prl7" event={"ID":"b3fe592e-d817-4d38-8082-fe800be92147","Type":"ContainerDied","Data":"11c606772bd58f64703e6275d28a02bc9e6c6ff3a42f24aba7761fb9771a1e38"} Nov 21 14:52:21 crc kubenswrapper[4897]: I1121 14:52:21.994780 4897 scope.go:117] "RemoveContainer" containerID="f86960bf3fcbb0166df044017ccd3c080affb6d27e16d204771c87bffedbeb96" Nov 21 14:52:22 crc kubenswrapper[4897]: I1121 14:52:22.714679 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2prl7" Nov 21 14:52:22 crc kubenswrapper[4897]: I1121 14:52:22.883148 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b3fe592e-d817-4d38-8082-fe800be92147-inventory\") pod \"b3fe592e-d817-4d38-8082-fe800be92147\" (UID: \"b3fe592e-d817-4d38-8082-fe800be92147\") " Nov 21 14:52:22 crc kubenswrapper[4897]: I1121 14:52:22.883459 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b3fe592e-d817-4d38-8082-fe800be92147-ssh-key\") pod \"b3fe592e-d817-4d38-8082-fe800be92147\" (UID: \"b3fe592e-d817-4d38-8082-fe800be92147\") " Nov 21 14:52:22 crc kubenswrapper[4897]: I1121 14:52:22.883525 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sq9st\" (UniqueName: \"kubernetes.io/projected/b3fe592e-d817-4d38-8082-fe800be92147-kube-api-access-sq9st\") pod \"b3fe592e-d817-4d38-8082-fe800be92147\" (UID: \"b3fe592e-d817-4d38-8082-fe800be92147\") " Nov 21 14:52:22 crc kubenswrapper[4897]: I1121 14:52:22.889420 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3fe592e-d817-4d38-8082-fe800be92147-kube-api-access-sq9st" (OuterVolumeSpecName: "kube-api-access-sq9st") pod "b3fe592e-d817-4d38-8082-fe800be92147" (UID: "b3fe592e-d817-4d38-8082-fe800be92147"). InnerVolumeSpecName "kube-api-access-sq9st". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:52:22 crc kubenswrapper[4897]: I1121 14:52:22.919295 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3fe592e-d817-4d38-8082-fe800be92147-inventory" (OuterVolumeSpecName: "inventory") pod "b3fe592e-d817-4d38-8082-fe800be92147" (UID: "b3fe592e-d817-4d38-8082-fe800be92147"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:52:22 crc kubenswrapper[4897]: I1121 14:52:22.932135 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3fe592e-d817-4d38-8082-fe800be92147-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b3fe592e-d817-4d38-8082-fe800be92147" (UID: "b3fe592e-d817-4d38-8082-fe800be92147"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:52:22 crc kubenswrapper[4897]: I1121 14:52:22.986807 4897 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b3fe592e-d817-4d38-8082-fe800be92147-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 21 14:52:22 crc kubenswrapper[4897]: I1121 14:52:22.986843 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sq9st\" (UniqueName: \"kubernetes.io/projected/b3fe592e-d817-4d38-8082-fe800be92147-kube-api-access-sq9st\") on node \"crc\" DevicePath \"\"" Nov 21 14:52:22 crc kubenswrapper[4897]: I1121 14:52:22.986856 4897 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b3fe592e-d817-4d38-8082-fe800be92147-inventory\") on node \"crc\" DevicePath \"\"" Nov 21 14:52:23 crc kubenswrapper[4897]: I1121 14:52:23.221905 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2prl7" event={"ID":"b3fe592e-d817-4d38-8082-fe800be92147","Type":"ContainerDied","Data":"0ecd1b3dd69349a80ead581f50fc3d4cee8bd58a3091855c85ac0bf61a1ec830"} Nov 21 14:52:23 crc kubenswrapper[4897]: I1121 14:52:23.222635 4897 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0ecd1b3dd69349a80ead581f50fc3d4cee8bd58a3091855c85ac0bf61a1ec830" Nov 21 14:52:23 crc kubenswrapper[4897]: I1121 14:52:23.222829 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2prl7" Nov 21 14:52:23 crc kubenswrapper[4897]: I1121 14:52:23.341590 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-hctxf"] Nov 21 14:52:23 crc kubenswrapper[4897]: E1121 14:52:23.342596 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b468737-2893-4083-a91e-6c7b7d8d5d5a" containerName="extract-utilities" Nov 21 14:52:23 crc kubenswrapper[4897]: I1121 14:52:23.342768 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b468737-2893-4083-a91e-6c7b7d8d5d5a" containerName="extract-utilities" Nov 21 14:52:23 crc kubenswrapper[4897]: E1121 14:52:23.342890 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b468737-2893-4083-a91e-6c7b7d8d5d5a" containerName="registry-server" Nov 21 14:52:23 crc kubenswrapper[4897]: I1121 14:52:23.342975 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b468737-2893-4083-a91e-6c7b7d8d5d5a" containerName="registry-server" Nov 21 14:52:23 crc kubenswrapper[4897]: E1121 14:52:23.343064 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b468737-2893-4083-a91e-6c7b7d8d5d5a" containerName="extract-content" Nov 21 14:52:23 crc kubenswrapper[4897]: I1121 14:52:23.343161 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b468737-2893-4083-a91e-6c7b7d8d5d5a" containerName="extract-content" Nov 21 14:52:23 crc kubenswrapper[4897]: E1121 14:52:23.343259 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3fe592e-d817-4d38-8082-fe800be92147" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 21 14:52:23 crc kubenswrapper[4897]: I1121 14:52:23.343335 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3fe592e-d817-4d38-8082-fe800be92147" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 21 14:52:23 crc kubenswrapper[4897]: I1121 14:52:23.343758 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b468737-2893-4083-a91e-6c7b7d8d5d5a" containerName="registry-server" Nov 21 14:52:23 crc kubenswrapper[4897]: I1121 14:52:23.343881 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3fe592e-d817-4d38-8082-fe800be92147" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 21 14:52:23 crc kubenswrapper[4897]: I1121 14:52:23.345120 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-hctxf" Nov 21 14:52:23 crc kubenswrapper[4897]: I1121 14:52:23.352108 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 21 14:52:23 crc kubenswrapper[4897]: I1121 14:52:23.352434 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-dj88m" Nov 21 14:52:23 crc kubenswrapper[4897]: I1121 14:52:23.352724 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 21 14:52:23 crc kubenswrapper[4897]: I1121 14:52:23.363076 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-hctxf"] Nov 21 14:52:23 crc kubenswrapper[4897]: I1121 14:52:23.368045 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 21 14:52:23 crc kubenswrapper[4897]: I1121 14:52:23.505813 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9fb25923-3c49-41ee-bc5e-c88778b2f80f-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-hctxf\" (UID: \"9fb25923-3c49-41ee-bc5e-c88778b2f80f\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-hctxf" Nov 21 14:52:23 crc kubenswrapper[4897]: I1121 14:52:23.505855 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9fb25923-3c49-41ee-bc5e-c88778b2f80f-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-hctxf\" (UID: \"9fb25923-3c49-41ee-bc5e-c88778b2f80f\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-hctxf" Nov 21 14:52:23 crc kubenswrapper[4897]: I1121 14:52:23.507193 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gg95f\" (UniqueName: \"kubernetes.io/projected/9fb25923-3c49-41ee-bc5e-c88778b2f80f-kube-api-access-gg95f\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-hctxf\" (UID: \"9fb25923-3c49-41ee-bc5e-c88778b2f80f\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-hctxf" Nov 21 14:52:23 crc kubenswrapper[4897]: I1121 14:52:23.609075 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9fb25923-3c49-41ee-bc5e-c88778b2f80f-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-hctxf\" (UID: \"9fb25923-3c49-41ee-bc5e-c88778b2f80f\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-hctxf" Nov 21 14:52:23 crc kubenswrapper[4897]: I1121 14:52:23.609126 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9fb25923-3c49-41ee-bc5e-c88778b2f80f-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-hctxf\" (UID: \"9fb25923-3c49-41ee-bc5e-c88778b2f80f\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-hctxf" Nov 21 14:52:23 crc kubenswrapper[4897]: I1121 14:52:23.610145 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gg95f\" (UniqueName: \"kubernetes.io/projected/9fb25923-3c49-41ee-bc5e-c88778b2f80f-kube-api-access-gg95f\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-hctxf\" (UID: \"9fb25923-3c49-41ee-bc5e-c88778b2f80f\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-hctxf" Nov 21 14:52:23 crc kubenswrapper[4897]: I1121 14:52:23.614953 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9fb25923-3c49-41ee-bc5e-c88778b2f80f-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-hctxf\" (UID: \"9fb25923-3c49-41ee-bc5e-c88778b2f80f\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-hctxf" Nov 21 14:52:23 crc kubenswrapper[4897]: I1121 14:52:23.615597 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9fb25923-3c49-41ee-bc5e-c88778b2f80f-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-hctxf\" (UID: \"9fb25923-3c49-41ee-bc5e-c88778b2f80f\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-hctxf" Nov 21 14:52:23 crc kubenswrapper[4897]: I1121 14:52:23.627952 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gg95f\" (UniqueName: \"kubernetes.io/projected/9fb25923-3c49-41ee-bc5e-c88778b2f80f-kube-api-access-gg95f\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-hctxf\" (UID: \"9fb25923-3c49-41ee-bc5e-c88778b2f80f\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-hctxf" Nov 21 14:52:23 crc kubenswrapper[4897]: I1121 14:52:23.667319 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-hctxf" Nov 21 14:52:24 crc kubenswrapper[4897]: I1121 14:52:24.252080 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-hctxf"] Nov 21 14:52:25 crc kubenswrapper[4897]: I1121 14:52:25.245107 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-hctxf" event={"ID":"9fb25923-3c49-41ee-bc5e-c88778b2f80f","Type":"ContainerStarted","Data":"e1fa051c72f15103db8602d4de10722df8e17c6fa9da3b16f6ee23ce55f5170b"} Nov 21 14:52:26 crc kubenswrapper[4897]: I1121 14:52:26.259736 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-hctxf" event={"ID":"9fb25923-3c49-41ee-bc5e-c88778b2f80f","Type":"ContainerStarted","Data":"c42f9339ad0d0c80d061249bfa2cfbbe856f5dfbabef949d7669af56469e43f0"} Nov 21 14:52:26 crc kubenswrapper[4897]: I1121 14:52:26.281373 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-hctxf" podStartSLOduration=2.204213874 podStartE2EDuration="3.281337783s" podCreationTimestamp="2025-11-21 14:52:23 +0000 UTC" firstStartedPulling="2025-11-21 14:52:24.266908351 +0000 UTC m=+2621.551501826" lastFinishedPulling="2025-11-21 14:52:25.34403226 +0000 UTC m=+2622.628625735" observedRunningTime="2025-11-21 14:52:26.277601162 +0000 UTC m=+2623.562194637" watchObservedRunningTime="2025-11-21 14:52:26.281337783 +0000 UTC m=+2623.565931308" Nov 21 14:52:32 crc kubenswrapper[4897]: I1121 14:52:32.323484 4897 generic.go:334] "Generic (PLEG): container finished" podID="9fb25923-3c49-41ee-bc5e-c88778b2f80f" containerID="c42f9339ad0d0c80d061249bfa2cfbbe856f5dfbabef949d7669af56469e43f0" exitCode=0 Nov 21 14:52:32 crc kubenswrapper[4897]: I1121 14:52:32.323648 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-hctxf" event={"ID":"9fb25923-3c49-41ee-bc5e-c88778b2f80f","Type":"ContainerDied","Data":"c42f9339ad0d0c80d061249bfa2cfbbe856f5dfbabef949d7669af56469e43f0"} Nov 21 14:52:33 crc kubenswrapper[4897]: I1121 14:52:33.836354 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-hctxf" Nov 21 14:52:33 crc kubenswrapper[4897]: I1121 14:52:33.997376 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9fb25923-3c49-41ee-bc5e-c88778b2f80f-inventory\") pod \"9fb25923-3c49-41ee-bc5e-c88778b2f80f\" (UID: \"9fb25923-3c49-41ee-bc5e-c88778b2f80f\") " Nov 21 14:52:33 crc kubenswrapper[4897]: I1121 14:52:33.997426 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9fb25923-3c49-41ee-bc5e-c88778b2f80f-ssh-key\") pod \"9fb25923-3c49-41ee-bc5e-c88778b2f80f\" (UID: \"9fb25923-3c49-41ee-bc5e-c88778b2f80f\") " Nov 21 14:52:33 crc kubenswrapper[4897]: I1121 14:52:33.997827 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gg95f\" (UniqueName: \"kubernetes.io/projected/9fb25923-3c49-41ee-bc5e-c88778b2f80f-kube-api-access-gg95f\") pod \"9fb25923-3c49-41ee-bc5e-c88778b2f80f\" (UID: \"9fb25923-3c49-41ee-bc5e-c88778b2f80f\") " Nov 21 14:52:34 crc kubenswrapper[4897]: I1121 14:52:34.004824 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9fb25923-3c49-41ee-bc5e-c88778b2f80f-kube-api-access-gg95f" (OuterVolumeSpecName: "kube-api-access-gg95f") pod "9fb25923-3c49-41ee-bc5e-c88778b2f80f" (UID: "9fb25923-3c49-41ee-bc5e-c88778b2f80f"). InnerVolumeSpecName "kube-api-access-gg95f". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:52:34 crc kubenswrapper[4897]: I1121 14:52:34.033892 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9fb25923-3c49-41ee-bc5e-c88778b2f80f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "9fb25923-3c49-41ee-bc5e-c88778b2f80f" (UID: "9fb25923-3c49-41ee-bc5e-c88778b2f80f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:52:34 crc kubenswrapper[4897]: I1121 14:52:34.045255 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9fb25923-3c49-41ee-bc5e-c88778b2f80f-inventory" (OuterVolumeSpecName: "inventory") pod "9fb25923-3c49-41ee-bc5e-c88778b2f80f" (UID: "9fb25923-3c49-41ee-bc5e-c88778b2f80f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:52:34 crc kubenswrapper[4897]: I1121 14:52:34.100947 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gg95f\" (UniqueName: \"kubernetes.io/projected/9fb25923-3c49-41ee-bc5e-c88778b2f80f-kube-api-access-gg95f\") on node \"crc\" DevicePath \"\"" Nov 21 14:52:34 crc kubenswrapper[4897]: I1121 14:52:34.100983 4897 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9fb25923-3c49-41ee-bc5e-c88778b2f80f-inventory\") on node \"crc\" DevicePath \"\"" Nov 21 14:52:34 crc kubenswrapper[4897]: I1121 14:52:34.100992 4897 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9fb25923-3c49-41ee-bc5e-c88778b2f80f-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 21 14:52:34 crc kubenswrapper[4897]: I1121 14:52:34.353562 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-hctxf" event={"ID":"9fb25923-3c49-41ee-bc5e-c88778b2f80f","Type":"ContainerDied","Data":"e1fa051c72f15103db8602d4de10722df8e17c6fa9da3b16f6ee23ce55f5170b"} Nov 21 14:52:34 crc kubenswrapper[4897]: I1121 14:52:34.353615 4897 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e1fa051c72f15103db8602d4de10722df8e17c6fa9da3b16f6ee23ce55f5170b" Nov 21 14:52:34 crc kubenswrapper[4897]: I1121 14:52:34.353644 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-hctxf" Nov 21 14:52:34 crc kubenswrapper[4897]: I1121 14:52:34.444050 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-mhqzh"] Nov 21 14:52:34 crc kubenswrapper[4897]: E1121 14:52:34.447173 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9fb25923-3c49-41ee-bc5e-c88778b2f80f" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 21 14:52:34 crc kubenswrapper[4897]: I1121 14:52:34.447206 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="9fb25923-3c49-41ee-bc5e-c88778b2f80f" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 21 14:52:34 crc kubenswrapper[4897]: I1121 14:52:34.448120 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="9fb25923-3c49-41ee-bc5e-c88778b2f80f" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 21 14:52:34 crc kubenswrapper[4897]: I1121 14:52:34.449639 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mhqzh" Nov 21 14:52:34 crc kubenswrapper[4897]: I1121 14:52:34.452350 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 21 14:52:34 crc kubenswrapper[4897]: I1121 14:52:34.452898 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 21 14:52:34 crc kubenswrapper[4897]: I1121 14:52:34.453108 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 21 14:52:34 crc kubenswrapper[4897]: I1121 14:52:34.453246 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-dj88m" Nov 21 14:52:34 crc kubenswrapper[4897]: I1121 14:52:34.458102 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-mhqzh"] Nov 21 14:52:34 crc kubenswrapper[4897]: I1121 14:52:34.514678 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0056e4f5-937c-414d-a964-2d72c298ea6d-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-mhqzh\" (UID: \"0056e4f5-937c-414d-a964-2d72c298ea6d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mhqzh" Nov 21 14:52:34 crc kubenswrapper[4897]: I1121 14:52:34.514798 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qjndf\" (UniqueName: \"kubernetes.io/projected/0056e4f5-937c-414d-a964-2d72c298ea6d-kube-api-access-qjndf\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-mhqzh\" (UID: \"0056e4f5-937c-414d-a964-2d72c298ea6d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mhqzh" Nov 21 14:52:34 crc kubenswrapper[4897]: I1121 14:52:34.514988 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0056e4f5-937c-414d-a964-2d72c298ea6d-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-mhqzh\" (UID: \"0056e4f5-937c-414d-a964-2d72c298ea6d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mhqzh" Nov 21 14:52:34 crc kubenswrapper[4897]: I1121 14:52:34.617066 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0056e4f5-937c-414d-a964-2d72c298ea6d-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-mhqzh\" (UID: \"0056e4f5-937c-414d-a964-2d72c298ea6d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mhqzh" Nov 21 14:52:34 crc kubenswrapper[4897]: I1121 14:52:34.617360 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qjndf\" (UniqueName: \"kubernetes.io/projected/0056e4f5-937c-414d-a964-2d72c298ea6d-kube-api-access-qjndf\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-mhqzh\" (UID: \"0056e4f5-937c-414d-a964-2d72c298ea6d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mhqzh" Nov 21 14:52:34 crc kubenswrapper[4897]: I1121 14:52:34.617455 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0056e4f5-937c-414d-a964-2d72c298ea6d-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-mhqzh\" (UID: \"0056e4f5-937c-414d-a964-2d72c298ea6d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mhqzh" Nov 21 14:52:34 crc kubenswrapper[4897]: I1121 14:52:34.623412 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0056e4f5-937c-414d-a964-2d72c298ea6d-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-mhqzh\" (UID: \"0056e4f5-937c-414d-a964-2d72c298ea6d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mhqzh" Nov 21 14:52:34 crc kubenswrapper[4897]: I1121 14:52:34.623443 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0056e4f5-937c-414d-a964-2d72c298ea6d-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-mhqzh\" (UID: \"0056e4f5-937c-414d-a964-2d72c298ea6d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mhqzh" Nov 21 14:52:34 crc kubenswrapper[4897]: I1121 14:52:34.645075 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qjndf\" (UniqueName: \"kubernetes.io/projected/0056e4f5-937c-414d-a964-2d72c298ea6d-kube-api-access-qjndf\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-mhqzh\" (UID: \"0056e4f5-937c-414d-a964-2d72c298ea6d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mhqzh" Nov 21 14:52:34 crc kubenswrapper[4897]: I1121 14:52:34.769835 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mhqzh" Nov 21 14:52:35 crc kubenswrapper[4897]: I1121 14:52:35.306697 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-mhqzh"] Nov 21 14:52:35 crc kubenswrapper[4897]: I1121 14:52:35.365910 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mhqzh" event={"ID":"0056e4f5-937c-414d-a964-2d72c298ea6d","Type":"ContainerStarted","Data":"d2157c9e6225416a5c10c4366e1c67aad963c41b7383e3c8a3dba24feefb99f8"} Nov 21 14:52:36 crc kubenswrapper[4897]: I1121 14:52:36.379344 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mhqzh" event={"ID":"0056e4f5-937c-414d-a964-2d72c298ea6d","Type":"ContainerStarted","Data":"1e85738d486ac572d6990bb94b19538c8d9d39ccee3df07a671b2a53e8839e30"} Nov 21 14:52:36 crc kubenswrapper[4897]: I1121 14:52:36.411687 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mhqzh" podStartSLOduration=1.810462907 podStartE2EDuration="2.411670557s" podCreationTimestamp="2025-11-21 14:52:34 +0000 UTC" firstStartedPulling="2025-11-21 14:52:35.314091115 +0000 UTC m=+2632.598684590" lastFinishedPulling="2025-11-21 14:52:35.915298765 +0000 UTC m=+2633.199892240" observedRunningTime="2025-11-21 14:52:36.405030878 +0000 UTC m=+2633.689624353" watchObservedRunningTime="2025-11-21 14:52:36.411670557 +0000 UTC m=+2633.696264022" Nov 21 14:53:16 crc kubenswrapper[4897]: I1121 14:53:16.823338 4897 generic.go:334] "Generic (PLEG): container finished" podID="0056e4f5-937c-414d-a964-2d72c298ea6d" containerID="1e85738d486ac572d6990bb94b19538c8d9d39ccee3df07a671b2a53e8839e30" exitCode=0 Nov 21 14:53:16 crc kubenswrapper[4897]: I1121 14:53:16.823423 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mhqzh" event={"ID":"0056e4f5-937c-414d-a964-2d72c298ea6d","Type":"ContainerDied","Data":"1e85738d486ac572d6990bb94b19538c8d9d39ccee3df07a671b2a53e8839e30"} Nov 21 14:53:18 crc kubenswrapper[4897]: I1121 14:53:18.274214 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mhqzh" Nov 21 14:53:18 crc kubenswrapper[4897]: I1121 14:53:18.326544 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0056e4f5-937c-414d-a964-2d72c298ea6d-ssh-key\") pod \"0056e4f5-937c-414d-a964-2d72c298ea6d\" (UID: \"0056e4f5-937c-414d-a964-2d72c298ea6d\") " Nov 21 14:53:18 crc kubenswrapper[4897]: I1121 14:53:18.326637 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qjndf\" (UniqueName: \"kubernetes.io/projected/0056e4f5-937c-414d-a964-2d72c298ea6d-kube-api-access-qjndf\") pod \"0056e4f5-937c-414d-a964-2d72c298ea6d\" (UID: \"0056e4f5-937c-414d-a964-2d72c298ea6d\") " Nov 21 14:53:18 crc kubenswrapper[4897]: I1121 14:53:18.326741 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0056e4f5-937c-414d-a964-2d72c298ea6d-inventory\") pod \"0056e4f5-937c-414d-a964-2d72c298ea6d\" (UID: \"0056e4f5-937c-414d-a964-2d72c298ea6d\") " Nov 21 14:53:18 crc kubenswrapper[4897]: I1121 14:53:18.336282 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0056e4f5-937c-414d-a964-2d72c298ea6d-kube-api-access-qjndf" (OuterVolumeSpecName: "kube-api-access-qjndf") pod "0056e4f5-937c-414d-a964-2d72c298ea6d" (UID: "0056e4f5-937c-414d-a964-2d72c298ea6d"). InnerVolumeSpecName "kube-api-access-qjndf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:53:18 crc kubenswrapper[4897]: I1121 14:53:18.360577 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0056e4f5-937c-414d-a964-2d72c298ea6d-inventory" (OuterVolumeSpecName: "inventory") pod "0056e4f5-937c-414d-a964-2d72c298ea6d" (UID: "0056e4f5-937c-414d-a964-2d72c298ea6d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:53:18 crc kubenswrapper[4897]: I1121 14:53:18.366496 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0056e4f5-937c-414d-a964-2d72c298ea6d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "0056e4f5-937c-414d-a964-2d72c298ea6d" (UID: "0056e4f5-937c-414d-a964-2d72c298ea6d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:53:18 crc kubenswrapper[4897]: I1121 14:53:18.430379 4897 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0056e4f5-937c-414d-a964-2d72c298ea6d-inventory\") on node \"crc\" DevicePath \"\"" Nov 21 14:53:18 crc kubenswrapper[4897]: I1121 14:53:18.430420 4897 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0056e4f5-937c-414d-a964-2d72c298ea6d-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 21 14:53:18 crc kubenswrapper[4897]: I1121 14:53:18.430462 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qjndf\" (UniqueName: \"kubernetes.io/projected/0056e4f5-937c-414d-a964-2d72c298ea6d-kube-api-access-qjndf\") on node \"crc\" DevicePath \"\"" Nov 21 14:53:18 crc kubenswrapper[4897]: I1121 14:53:18.845023 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mhqzh" event={"ID":"0056e4f5-937c-414d-a964-2d72c298ea6d","Type":"ContainerDied","Data":"d2157c9e6225416a5c10c4366e1c67aad963c41b7383e3c8a3dba24feefb99f8"} Nov 21 14:53:18 crc kubenswrapper[4897]: I1121 14:53:18.845346 4897 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d2157c9e6225416a5c10c4366e1c67aad963c41b7383e3c8a3dba24feefb99f8" Nov 21 14:53:18 crc kubenswrapper[4897]: I1121 14:53:18.845065 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mhqzh" Nov 21 14:53:18 crc kubenswrapper[4897]: I1121 14:53:18.953621 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tzbxw"] Nov 21 14:53:18 crc kubenswrapper[4897]: E1121 14:53:18.954062 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0056e4f5-937c-414d-a964-2d72c298ea6d" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 21 14:53:18 crc kubenswrapper[4897]: I1121 14:53:18.954079 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="0056e4f5-937c-414d-a964-2d72c298ea6d" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 21 14:53:18 crc kubenswrapper[4897]: I1121 14:53:18.954323 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="0056e4f5-937c-414d-a964-2d72c298ea6d" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 21 14:53:18 crc kubenswrapper[4897]: I1121 14:53:18.956082 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tzbxw" Nov 21 14:53:18 crc kubenswrapper[4897]: I1121 14:53:18.963128 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 21 14:53:18 crc kubenswrapper[4897]: I1121 14:53:18.963474 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 21 14:53:18 crc kubenswrapper[4897]: I1121 14:53:18.963655 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 21 14:53:18 crc kubenswrapper[4897]: I1121 14:53:18.963924 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-dj88m" Nov 21 14:53:18 crc kubenswrapper[4897]: I1121 14:53:18.965802 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tzbxw"] Nov 21 14:53:19 crc kubenswrapper[4897]: I1121 14:53:19.041794 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9przn\" (UniqueName: \"kubernetes.io/projected/044196f5-debd-41e8-99a7-0c2bc61e3699-kube-api-access-9przn\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-tzbxw\" (UID: \"044196f5-debd-41e8-99a7-0c2bc61e3699\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tzbxw" Nov 21 14:53:19 crc kubenswrapper[4897]: I1121 14:53:19.041856 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/044196f5-debd-41e8-99a7-0c2bc61e3699-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-tzbxw\" (UID: \"044196f5-debd-41e8-99a7-0c2bc61e3699\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tzbxw" Nov 21 14:53:19 crc kubenswrapper[4897]: I1121 14:53:19.042315 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/044196f5-debd-41e8-99a7-0c2bc61e3699-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-tzbxw\" (UID: \"044196f5-debd-41e8-99a7-0c2bc61e3699\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tzbxw" Nov 21 14:53:19 crc kubenswrapper[4897]: I1121 14:53:19.144175 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9przn\" (UniqueName: \"kubernetes.io/projected/044196f5-debd-41e8-99a7-0c2bc61e3699-kube-api-access-9przn\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-tzbxw\" (UID: \"044196f5-debd-41e8-99a7-0c2bc61e3699\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tzbxw" Nov 21 14:53:19 crc kubenswrapper[4897]: I1121 14:53:19.144296 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/044196f5-debd-41e8-99a7-0c2bc61e3699-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-tzbxw\" (UID: \"044196f5-debd-41e8-99a7-0c2bc61e3699\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tzbxw" Nov 21 14:53:19 crc kubenswrapper[4897]: I1121 14:53:19.145834 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/044196f5-debd-41e8-99a7-0c2bc61e3699-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-tzbxw\" (UID: \"044196f5-debd-41e8-99a7-0c2bc61e3699\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tzbxw" Nov 21 14:53:19 crc kubenswrapper[4897]: I1121 14:53:19.155779 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/044196f5-debd-41e8-99a7-0c2bc61e3699-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-tzbxw\" (UID: \"044196f5-debd-41e8-99a7-0c2bc61e3699\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tzbxw" Nov 21 14:53:19 crc kubenswrapper[4897]: I1121 14:53:19.162670 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9przn\" (UniqueName: \"kubernetes.io/projected/044196f5-debd-41e8-99a7-0c2bc61e3699-kube-api-access-9przn\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-tzbxw\" (UID: \"044196f5-debd-41e8-99a7-0c2bc61e3699\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tzbxw" Nov 21 14:53:19 crc kubenswrapper[4897]: I1121 14:53:19.165570 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/044196f5-debd-41e8-99a7-0c2bc61e3699-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-tzbxw\" (UID: \"044196f5-debd-41e8-99a7-0c2bc61e3699\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tzbxw" Nov 21 14:53:19 crc kubenswrapper[4897]: I1121 14:53:19.273657 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tzbxw" Nov 21 14:53:19 crc kubenswrapper[4897]: I1121 14:53:19.806596 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tzbxw"] Nov 21 14:53:19 crc kubenswrapper[4897]: I1121 14:53:19.859341 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tzbxw" event={"ID":"044196f5-debd-41e8-99a7-0c2bc61e3699","Type":"ContainerStarted","Data":"0b52bb7eace289bc50acc3a5e7c52b1fa550d3b09fe12e0e1702ea0310b5243c"} Nov 21 14:53:21 crc kubenswrapper[4897]: I1121 14:53:21.880803 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tzbxw" event={"ID":"044196f5-debd-41e8-99a7-0c2bc61e3699","Type":"ContainerStarted","Data":"f9f94ccc4b71e47a97a96bf723ad240f502a214caaa34aa4de8b87fbec03514b"} Nov 21 14:53:21 crc kubenswrapper[4897]: I1121 14:53:21.904532 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tzbxw" podStartSLOduration=2.7895007659999997 podStartE2EDuration="3.904514379s" podCreationTimestamp="2025-11-21 14:53:18 +0000 UTC" firstStartedPulling="2025-11-21 14:53:19.830132089 +0000 UTC m=+2677.114725564" lastFinishedPulling="2025-11-21 14:53:20.945145692 +0000 UTC m=+2678.229739177" observedRunningTime="2025-11-21 14:53:21.896594314 +0000 UTC m=+2679.181187799" watchObservedRunningTime="2025-11-21 14:53:21.904514379 +0000 UTC m=+2679.189107854" Nov 21 14:54:04 crc kubenswrapper[4897]: I1121 14:54:04.371265 4897 patch_prober.go:28] interesting pod/machine-config-daemon-krv5b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 21 14:54:04 crc kubenswrapper[4897]: I1121 14:54:04.371951 4897 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 21 14:54:17 crc kubenswrapper[4897]: I1121 14:54:17.494438 4897 generic.go:334] "Generic (PLEG): container finished" podID="044196f5-debd-41e8-99a7-0c2bc61e3699" containerID="f9f94ccc4b71e47a97a96bf723ad240f502a214caaa34aa4de8b87fbec03514b" exitCode=0 Nov 21 14:54:17 crc kubenswrapper[4897]: I1121 14:54:17.494530 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tzbxw" event={"ID":"044196f5-debd-41e8-99a7-0c2bc61e3699","Type":"ContainerDied","Data":"f9f94ccc4b71e47a97a96bf723ad240f502a214caaa34aa4de8b87fbec03514b"} Nov 21 14:54:18 crc kubenswrapper[4897]: I1121 14:54:18.993991 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tzbxw" Nov 21 14:54:19 crc kubenswrapper[4897]: I1121 14:54:19.127012 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/044196f5-debd-41e8-99a7-0c2bc61e3699-ssh-key\") pod \"044196f5-debd-41e8-99a7-0c2bc61e3699\" (UID: \"044196f5-debd-41e8-99a7-0c2bc61e3699\") " Nov 21 14:54:19 crc kubenswrapper[4897]: I1121 14:54:19.127120 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/044196f5-debd-41e8-99a7-0c2bc61e3699-inventory\") pod \"044196f5-debd-41e8-99a7-0c2bc61e3699\" (UID: \"044196f5-debd-41e8-99a7-0c2bc61e3699\") " Nov 21 14:54:19 crc kubenswrapper[4897]: I1121 14:54:19.127160 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9przn\" (UniqueName: \"kubernetes.io/projected/044196f5-debd-41e8-99a7-0c2bc61e3699-kube-api-access-9przn\") pod \"044196f5-debd-41e8-99a7-0c2bc61e3699\" (UID: \"044196f5-debd-41e8-99a7-0c2bc61e3699\") " Nov 21 14:54:19 crc kubenswrapper[4897]: I1121 14:54:19.132844 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/044196f5-debd-41e8-99a7-0c2bc61e3699-kube-api-access-9przn" (OuterVolumeSpecName: "kube-api-access-9przn") pod "044196f5-debd-41e8-99a7-0c2bc61e3699" (UID: "044196f5-debd-41e8-99a7-0c2bc61e3699"). InnerVolumeSpecName "kube-api-access-9przn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:54:19 crc kubenswrapper[4897]: I1121 14:54:19.164638 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/044196f5-debd-41e8-99a7-0c2bc61e3699-inventory" (OuterVolumeSpecName: "inventory") pod "044196f5-debd-41e8-99a7-0c2bc61e3699" (UID: "044196f5-debd-41e8-99a7-0c2bc61e3699"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:54:19 crc kubenswrapper[4897]: I1121 14:54:19.177641 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/044196f5-debd-41e8-99a7-0c2bc61e3699-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "044196f5-debd-41e8-99a7-0c2bc61e3699" (UID: "044196f5-debd-41e8-99a7-0c2bc61e3699"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:54:19 crc kubenswrapper[4897]: I1121 14:54:19.230112 4897 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/044196f5-debd-41e8-99a7-0c2bc61e3699-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 21 14:54:19 crc kubenswrapper[4897]: I1121 14:54:19.230435 4897 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/044196f5-debd-41e8-99a7-0c2bc61e3699-inventory\") on node \"crc\" DevicePath \"\"" Nov 21 14:54:19 crc kubenswrapper[4897]: I1121 14:54:19.230445 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9przn\" (UniqueName: \"kubernetes.io/projected/044196f5-debd-41e8-99a7-0c2bc61e3699-kube-api-access-9przn\") on node \"crc\" DevicePath \"\"" Nov 21 14:54:19 crc kubenswrapper[4897]: I1121 14:54:19.526809 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tzbxw" event={"ID":"044196f5-debd-41e8-99a7-0c2bc61e3699","Type":"ContainerDied","Data":"0b52bb7eace289bc50acc3a5e7c52b1fa550d3b09fe12e0e1702ea0310b5243c"} Nov 21 14:54:19 crc kubenswrapper[4897]: I1121 14:54:19.526893 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tzbxw" Nov 21 14:54:19 crc kubenswrapper[4897]: I1121 14:54:19.526896 4897 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0b52bb7eace289bc50acc3a5e7c52b1fa550d3b09fe12e0e1702ea0310b5243c" Nov 21 14:54:19 crc kubenswrapper[4897]: I1121 14:54:19.599636 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-vj7t2"] Nov 21 14:54:19 crc kubenswrapper[4897]: E1121 14:54:19.600696 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="044196f5-debd-41e8-99a7-0c2bc61e3699" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 21 14:54:19 crc kubenswrapper[4897]: I1121 14:54:19.600796 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="044196f5-debd-41e8-99a7-0c2bc61e3699" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 21 14:54:19 crc kubenswrapper[4897]: I1121 14:54:19.601105 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="044196f5-debd-41e8-99a7-0c2bc61e3699" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 21 14:54:19 crc kubenswrapper[4897]: I1121 14:54:19.602308 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-vj7t2" Nov 21 14:54:19 crc kubenswrapper[4897]: I1121 14:54:19.605475 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 21 14:54:19 crc kubenswrapper[4897]: I1121 14:54:19.605667 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 21 14:54:19 crc kubenswrapper[4897]: I1121 14:54:19.605679 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 21 14:54:19 crc kubenswrapper[4897]: I1121 14:54:19.607818 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-dj88m" Nov 21 14:54:19 crc kubenswrapper[4897]: I1121 14:54:19.616285 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-vj7t2"] Nov 21 14:54:19 crc kubenswrapper[4897]: I1121 14:54:19.743861 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3d331ad5-87b2-433b-a53f-2be5b7f17eea-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-vj7t2\" (UID: \"3d331ad5-87b2-433b-a53f-2be5b7f17eea\") " pod="openstack/ssh-known-hosts-edpm-deployment-vj7t2" Nov 21 14:54:19 crc kubenswrapper[4897]: I1121 14:54:19.743925 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b4km9\" (UniqueName: \"kubernetes.io/projected/3d331ad5-87b2-433b-a53f-2be5b7f17eea-kube-api-access-b4km9\") pod \"ssh-known-hosts-edpm-deployment-vj7t2\" (UID: \"3d331ad5-87b2-433b-a53f-2be5b7f17eea\") " pod="openstack/ssh-known-hosts-edpm-deployment-vj7t2" Nov 21 14:54:19 crc kubenswrapper[4897]: I1121 14:54:19.744022 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/3d331ad5-87b2-433b-a53f-2be5b7f17eea-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-vj7t2\" (UID: \"3d331ad5-87b2-433b-a53f-2be5b7f17eea\") " pod="openstack/ssh-known-hosts-edpm-deployment-vj7t2" Nov 21 14:54:19 crc kubenswrapper[4897]: I1121 14:54:19.846294 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3d331ad5-87b2-433b-a53f-2be5b7f17eea-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-vj7t2\" (UID: \"3d331ad5-87b2-433b-a53f-2be5b7f17eea\") " pod="openstack/ssh-known-hosts-edpm-deployment-vj7t2" Nov 21 14:54:19 crc kubenswrapper[4897]: I1121 14:54:19.846375 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b4km9\" (UniqueName: \"kubernetes.io/projected/3d331ad5-87b2-433b-a53f-2be5b7f17eea-kube-api-access-b4km9\") pod \"ssh-known-hosts-edpm-deployment-vj7t2\" (UID: \"3d331ad5-87b2-433b-a53f-2be5b7f17eea\") " pod="openstack/ssh-known-hosts-edpm-deployment-vj7t2" Nov 21 14:54:19 crc kubenswrapper[4897]: I1121 14:54:19.846578 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/3d331ad5-87b2-433b-a53f-2be5b7f17eea-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-vj7t2\" (UID: \"3d331ad5-87b2-433b-a53f-2be5b7f17eea\") " pod="openstack/ssh-known-hosts-edpm-deployment-vj7t2" Nov 21 14:54:19 crc kubenswrapper[4897]: I1121 14:54:19.851008 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3d331ad5-87b2-433b-a53f-2be5b7f17eea-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-vj7t2\" (UID: \"3d331ad5-87b2-433b-a53f-2be5b7f17eea\") " pod="openstack/ssh-known-hosts-edpm-deployment-vj7t2" Nov 21 14:54:19 crc kubenswrapper[4897]: I1121 14:54:19.859268 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/3d331ad5-87b2-433b-a53f-2be5b7f17eea-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-vj7t2\" (UID: \"3d331ad5-87b2-433b-a53f-2be5b7f17eea\") " pod="openstack/ssh-known-hosts-edpm-deployment-vj7t2" Nov 21 14:54:19 crc kubenswrapper[4897]: I1121 14:54:19.872456 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b4km9\" (UniqueName: \"kubernetes.io/projected/3d331ad5-87b2-433b-a53f-2be5b7f17eea-kube-api-access-b4km9\") pod \"ssh-known-hosts-edpm-deployment-vj7t2\" (UID: \"3d331ad5-87b2-433b-a53f-2be5b7f17eea\") " pod="openstack/ssh-known-hosts-edpm-deployment-vj7t2" Nov 21 14:54:19 crc kubenswrapper[4897]: I1121 14:54:19.924064 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-vj7t2" Nov 21 14:54:20 crc kubenswrapper[4897]: I1121 14:54:20.531186 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-vj7t2"] Nov 21 14:54:21 crc kubenswrapper[4897]: I1121 14:54:21.550633 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-vj7t2" event={"ID":"3d331ad5-87b2-433b-a53f-2be5b7f17eea","Type":"ContainerStarted","Data":"b77f8f4359e487466023ed0826d6702a650373396fb5d60f8e0bacfaf47660ea"} Nov 21 14:54:25 crc kubenswrapper[4897]: I1121 14:54:25.603595 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-vj7t2" event={"ID":"3d331ad5-87b2-433b-a53f-2be5b7f17eea","Type":"ContainerStarted","Data":"e46555510754da7921c3c8fe5043a637696d04dac19dae8e72973359cb12513b"} Nov 21 14:54:25 crc kubenswrapper[4897]: I1121 14:54:25.621835 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-vj7t2" podStartSLOduration=2.642719569 podStartE2EDuration="6.621814864s" podCreationTimestamp="2025-11-21 14:54:19 +0000 UTC" firstStartedPulling="2025-11-21 14:54:20.536304809 +0000 UTC m=+2737.820898284" lastFinishedPulling="2025-11-21 14:54:24.515400084 +0000 UTC m=+2741.799993579" observedRunningTime="2025-11-21 14:54:25.61944128 +0000 UTC m=+2742.904034775" watchObservedRunningTime="2025-11-21 14:54:25.621814864 +0000 UTC m=+2742.906408349" Nov 21 14:54:34 crc kubenswrapper[4897]: I1121 14:54:34.371393 4897 patch_prober.go:28] interesting pod/machine-config-daemon-krv5b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 21 14:54:34 crc kubenswrapper[4897]: I1121 14:54:34.372306 4897 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 21 14:54:34 crc kubenswrapper[4897]: I1121 14:54:34.727286 4897 generic.go:334] "Generic (PLEG): container finished" podID="3d331ad5-87b2-433b-a53f-2be5b7f17eea" containerID="e46555510754da7921c3c8fe5043a637696d04dac19dae8e72973359cb12513b" exitCode=0 Nov 21 14:54:34 crc kubenswrapper[4897]: I1121 14:54:34.727341 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-vj7t2" event={"ID":"3d331ad5-87b2-433b-a53f-2be5b7f17eea","Type":"ContainerDied","Data":"e46555510754da7921c3c8fe5043a637696d04dac19dae8e72973359cb12513b"} Nov 21 14:54:35 crc kubenswrapper[4897]: I1121 14:54:35.641694 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-q6td4"] Nov 21 14:54:35 crc kubenswrapper[4897]: I1121 14:54:35.644112 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-q6td4" Nov 21 14:54:35 crc kubenswrapper[4897]: I1121 14:54:35.661413 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-q6td4"] Nov 21 14:54:35 crc kubenswrapper[4897]: I1121 14:54:35.767971 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c27aa401-13de-4f6e-beb1-d1683c852940-utilities\") pod \"certified-operators-q6td4\" (UID: \"c27aa401-13de-4f6e-beb1-d1683c852940\") " pod="openshift-marketplace/certified-operators-q6td4" Nov 21 14:54:35 crc kubenswrapper[4897]: I1121 14:54:35.768116 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c27aa401-13de-4f6e-beb1-d1683c852940-catalog-content\") pod \"certified-operators-q6td4\" (UID: \"c27aa401-13de-4f6e-beb1-d1683c852940\") " pod="openshift-marketplace/certified-operators-q6td4" Nov 21 14:54:35 crc kubenswrapper[4897]: I1121 14:54:35.768225 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9rnkb\" (UniqueName: \"kubernetes.io/projected/c27aa401-13de-4f6e-beb1-d1683c852940-kube-api-access-9rnkb\") pod \"certified-operators-q6td4\" (UID: \"c27aa401-13de-4f6e-beb1-d1683c852940\") " pod="openshift-marketplace/certified-operators-q6td4" Nov 21 14:54:35 crc kubenswrapper[4897]: I1121 14:54:35.869820 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9rnkb\" (UniqueName: \"kubernetes.io/projected/c27aa401-13de-4f6e-beb1-d1683c852940-kube-api-access-9rnkb\") pod \"certified-operators-q6td4\" (UID: \"c27aa401-13de-4f6e-beb1-d1683c852940\") " pod="openshift-marketplace/certified-operators-q6td4" Nov 21 14:54:35 crc kubenswrapper[4897]: I1121 14:54:35.869900 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c27aa401-13de-4f6e-beb1-d1683c852940-utilities\") pod \"certified-operators-q6td4\" (UID: \"c27aa401-13de-4f6e-beb1-d1683c852940\") " pod="openshift-marketplace/certified-operators-q6td4" Nov 21 14:54:35 crc kubenswrapper[4897]: I1121 14:54:35.870037 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c27aa401-13de-4f6e-beb1-d1683c852940-catalog-content\") pod \"certified-operators-q6td4\" (UID: \"c27aa401-13de-4f6e-beb1-d1683c852940\") " pod="openshift-marketplace/certified-operators-q6td4" Nov 21 14:54:35 crc kubenswrapper[4897]: I1121 14:54:35.870727 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c27aa401-13de-4f6e-beb1-d1683c852940-catalog-content\") pod \"certified-operators-q6td4\" (UID: \"c27aa401-13de-4f6e-beb1-d1683c852940\") " pod="openshift-marketplace/certified-operators-q6td4" Nov 21 14:54:35 crc kubenswrapper[4897]: I1121 14:54:35.871310 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c27aa401-13de-4f6e-beb1-d1683c852940-utilities\") pod \"certified-operators-q6td4\" (UID: \"c27aa401-13de-4f6e-beb1-d1683c852940\") " pod="openshift-marketplace/certified-operators-q6td4" Nov 21 14:54:35 crc kubenswrapper[4897]: I1121 14:54:35.889716 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9rnkb\" (UniqueName: \"kubernetes.io/projected/c27aa401-13de-4f6e-beb1-d1683c852940-kube-api-access-9rnkb\") pod \"certified-operators-q6td4\" (UID: \"c27aa401-13de-4f6e-beb1-d1683c852940\") " pod="openshift-marketplace/certified-operators-q6td4" Nov 21 14:54:35 crc kubenswrapper[4897]: I1121 14:54:35.965480 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-q6td4" Nov 21 14:54:36 crc kubenswrapper[4897]: I1121 14:54:36.466895 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-vj7t2" Nov 21 14:54:36 crc kubenswrapper[4897]: I1121 14:54:36.619737 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/3d331ad5-87b2-433b-a53f-2be5b7f17eea-inventory-0\") pod \"3d331ad5-87b2-433b-a53f-2be5b7f17eea\" (UID: \"3d331ad5-87b2-433b-a53f-2be5b7f17eea\") " Nov 21 14:54:36 crc kubenswrapper[4897]: I1121 14:54:36.619808 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b4km9\" (UniqueName: \"kubernetes.io/projected/3d331ad5-87b2-433b-a53f-2be5b7f17eea-kube-api-access-b4km9\") pod \"3d331ad5-87b2-433b-a53f-2be5b7f17eea\" (UID: \"3d331ad5-87b2-433b-a53f-2be5b7f17eea\") " Nov 21 14:54:36 crc kubenswrapper[4897]: I1121 14:54:36.620105 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3d331ad5-87b2-433b-a53f-2be5b7f17eea-ssh-key-openstack-edpm-ipam\") pod \"3d331ad5-87b2-433b-a53f-2be5b7f17eea\" (UID: \"3d331ad5-87b2-433b-a53f-2be5b7f17eea\") " Nov 21 14:54:36 crc kubenswrapper[4897]: I1121 14:54:36.629464 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d331ad5-87b2-433b-a53f-2be5b7f17eea-kube-api-access-b4km9" (OuterVolumeSpecName: "kube-api-access-b4km9") pod "3d331ad5-87b2-433b-a53f-2be5b7f17eea" (UID: "3d331ad5-87b2-433b-a53f-2be5b7f17eea"). InnerVolumeSpecName "kube-api-access-b4km9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:54:36 crc kubenswrapper[4897]: I1121 14:54:36.645236 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-q6td4"] Nov 21 14:54:36 crc kubenswrapper[4897]: I1121 14:54:36.682710 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d331ad5-87b2-433b-a53f-2be5b7f17eea-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "3d331ad5-87b2-433b-a53f-2be5b7f17eea" (UID: "3d331ad5-87b2-433b-a53f-2be5b7f17eea"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:54:36 crc kubenswrapper[4897]: I1121 14:54:36.704852 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d331ad5-87b2-433b-a53f-2be5b7f17eea-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "3d331ad5-87b2-433b-a53f-2be5b7f17eea" (UID: "3d331ad5-87b2-433b-a53f-2be5b7f17eea"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:54:36 crc kubenswrapper[4897]: I1121 14:54:36.723142 4897 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3d331ad5-87b2-433b-a53f-2be5b7f17eea-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Nov 21 14:54:36 crc kubenswrapper[4897]: I1121 14:54:36.723181 4897 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/3d331ad5-87b2-433b-a53f-2be5b7f17eea-inventory-0\") on node \"crc\" DevicePath \"\"" Nov 21 14:54:36 crc kubenswrapper[4897]: I1121 14:54:36.723193 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b4km9\" (UniqueName: \"kubernetes.io/projected/3d331ad5-87b2-433b-a53f-2be5b7f17eea-kube-api-access-b4km9\") on node \"crc\" DevicePath \"\"" Nov 21 14:54:36 crc kubenswrapper[4897]: I1121 14:54:36.763253 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q6td4" event={"ID":"c27aa401-13de-4f6e-beb1-d1683c852940","Type":"ContainerStarted","Data":"e49e53c317f4b31cc12a4ec3d30f70fb0937fd813576096fbef91aa7ce624b4a"} Nov 21 14:54:36 crc kubenswrapper[4897]: I1121 14:54:36.765233 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-vj7t2" event={"ID":"3d331ad5-87b2-433b-a53f-2be5b7f17eea","Type":"ContainerDied","Data":"b77f8f4359e487466023ed0826d6702a650373396fb5d60f8e0bacfaf47660ea"} Nov 21 14:54:36 crc kubenswrapper[4897]: I1121 14:54:36.765261 4897 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b77f8f4359e487466023ed0826d6702a650373396fb5d60f8e0bacfaf47660ea" Nov 21 14:54:36 crc kubenswrapper[4897]: I1121 14:54:36.765343 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-vj7t2" Nov 21 14:54:36 crc kubenswrapper[4897]: I1121 14:54:36.828706 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-2hspl"] Nov 21 14:54:36 crc kubenswrapper[4897]: E1121 14:54:36.829283 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d331ad5-87b2-433b-a53f-2be5b7f17eea" containerName="ssh-known-hosts-edpm-deployment" Nov 21 14:54:36 crc kubenswrapper[4897]: I1121 14:54:36.829311 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d331ad5-87b2-433b-a53f-2be5b7f17eea" containerName="ssh-known-hosts-edpm-deployment" Nov 21 14:54:36 crc kubenswrapper[4897]: I1121 14:54:36.829660 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d331ad5-87b2-433b-a53f-2be5b7f17eea" containerName="ssh-known-hosts-edpm-deployment" Nov 21 14:54:36 crc kubenswrapper[4897]: I1121 14:54:36.831463 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2hspl" Nov 21 14:54:36 crc kubenswrapper[4897]: I1121 14:54:36.833532 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 21 14:54:36 crc kubenswrapper[4897]: I1121 14:54:36.833630 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 21 14:54:36 crc kubenswrapper[4897]: I1121 14:54:36.833925 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 21 14:54:36 crc kubenswrapper[4897]: I1121 14:54:36.834035 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-dj88m" Nov 21 14:54:36 crc kubenswrapper[4897]: I1121 14:54:36.840162 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-2hspl"] Nov 21 14:54:36 crc kubenswrapper[4897]: I1121 14:54:36.926604 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w6kt9\" (UniqueName: \"kubernetes.io/projected/acb6440f-d6f8-4ef3-b05b-d1957afdd240-kube-api-access-w6kt9\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-2hspl\" (UID: \"acb6440f-d6f8-4ef3-b05b-d1957afdd240\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2hspl" Nov 21 14:54:36 crc kubenswrapper[4897]: I1121 14:54:36.926745 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/acb6440f-d6f8-4ef3-b05b-d1957afdd240-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-2hspl\" (UID: \"acb6440f-d6f8-4ef3-b05b-d1957afdd240\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2hspl" Nov 21 14:54:36 crc kubenswrapper[4897]: I1121 14:54:36.926820 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/acb6440f-d6f8-4ef3-b05b-d1957afdd240-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-2hspl\" (UID: \"acb6440f-d6f8-4ef3-b05b-d1957afdd240\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2hspl" Nov 21 14:54:37 crc kubenswrapper[4897]: I1121 14:54:37.029374 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/acb6440f-d6f8-4ef3-b05b-d1957afdd240-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-2hspl\" (UID: \"acb6440f-d6f8-4ef3-b05b-d1957afdd240\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2hspl" Nov 21 14:54:37 crc kubenswrapper[4897]: I1121 14:54:37.029525 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w6kt9\" (UniqueName: \"kubernetes.io/projected/acb6440f-d6f8-4ef3-b05b-d1957afdd240-kube-api-access-w6kt9\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-2hspl\" (UID: \"acb6440f-d6f8-4ef3-b05b-d1957afdd240\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2hspl" Nov 21 14:54:37 crc kubenswrapper[4897]: I1121 14:54:37.029620 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/acb6440f-d6f8-4ef3-b05b-d1957afdd240-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-2hspl\" (UID: \"acb6440f-d6f8-4ef3-b05b-d1957afdd240\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2hspl" Nov 21 14:54:37 crc kubenswrapper[4897]: I1121 14:54:37.033375 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/acb6440f-d6f8-4ef3-b05b-d1957afdd240-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-2hspl\" (UID: \"acb6440f-d6f8-4ef3-b05b-d1957afdd240\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2hspl" Nov 21 14:54:37 crc kubenswrapper[4897]: I1121 14:54:37.033726 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/acb6440f-d6f8-4ef3-b05b-d1957afdd240-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-2hspl\" (UID: \"acb6440f-d6f8-4ef3-b05b-d1957afdd240\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2hspl" Nov 21 14:54:37 crc kubenswrapper[4897]: I1121 14:54:37.049232 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w6kt9\" (UniqueName: \"kubernetes.io/projected/acb6440f-d6f8-4ef3-b05b-d1957afdd240-kube-api-access-w6kt9\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-2hspl\" (UID: \"acb6440f-d6f8-4ef3-b05b-d1957afdd240\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2hspl" Nov 21 14:54:37 crc kubenswrapper[4897]: I1121 14:54:37.158350 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2hspl" Nov 21 14:54:37 crc kubenswrapper[4897]: I1121 14:54:37.786180 4897 generic.go:334] "Generic (PLEG): container finished" podID="c27aa401-13de-4f6e-beb1-d1683c852940" containerID="1ef9e41a36284f61c96ae8e31011fb2e46e5b7a06be43b890d768d9d1c5a539c" exitCode=0 Nov 21 14:54:37 crc kubenswrapper[4897]: I1121 14:54:37.786746 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q6td4" event={"ID":"c27aa401-13de-4f6e-beb1-d1683c852940","Type":"ContainerDied","Data":"1ef9e41a36284f61c96ae8e31011fb2e46e5b7a06be43b890d768d9d1c5a539c"} Nov 21 14:54:37 crc kubenswrapper[4897]: I1121 14:54:37.815355 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-2hspl"] Nov 21 14:54:37 crc kubenswrapper[4897]: W1121 14:54:37.815658 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podacb6440f_d6f8_4ef3_b05b_d1957afdd240.slice/crio-344adf5734f7fb726c514912734219d69dae2215dfe8b8e77a38635b41ef9e04 WatchSource:0}: Error finding container 344adf5734f7fb726c514912734219d69dae2215dfe8b8e77a38635b41ef9e04: Status 404 returned error can't find the container with id 344adf5734f7fb726c514912734219d69dae2215dfe8b8e77a38635b41ef9e04 Nov 21 14:54:38 crc kubenswrapper[4897]: I1121 14:54:38.799981 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2hspl" event={"ID":"acb6440f-d6f8-4ef3-b05b-d1957afdd240","Type":"ContainerStarted","Data":"344adf5734f7fb726c514912734219d69dae2215dfe8b8e77a38635b41ef9e04"} Nov 21 14:54:41 crc kubenswrapper[4897]: I1121 14:54:41.831022 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2hspl" event={"ID":"acb6440f-d6f8-4ef3-b05b-d1957afdd240","Type":"ContainerStarted","Data":"41f3191fe594d23bbed5906047e0069db0b0c59d89908acb1bf3395d4a51df16"} Nov 21 14:54:41 crc kubenswrapper[4897]: I1121 14:54:41.836028 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q6td4" event={"ID":"c27aa401-13de-4f6e-beb1-d1683c852940","Type":"ContainerStarted","Data":"0bb79f630ec5fd12530a9a78048be4f84560f02746d5fbed01e2758c0a4d7dd9"} Nov 21 14:54:41 crc kubenswrapper[4897]: I1121 14:54:41.880274 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2hspl" podStartSLOduration=3.076447156 podStartE2EDuration="5.880255517s" podCreationTimestamp="2025-11-21 14:54:36 +0000 UTC" firstStartedPulling="2025-11-21 14:54:37.817801574 +0000 UTC m=+2755.102395049" lastFinishedPulling="2025-11-21 14:54:40.621609935 +0000 UTC m=+2757.906203410" observedRunningTime="2025-11-21 14:54:41.85010914 +0000 UTC m=+2759.134702635" watchObservedRunningTime="2025-11-21 14:54:41.880255517 +0000 UTC m=+2759.164848992" Nov 21 14:54:49 crc kubenswrapper[4897]: I1121 14:54:49.919386 4897 generic.go:334] "Generic (PLEG): container finished" podID="acb6440f-d6f8-4ef3-b05b-d1957afdd240" containerID="41f3191fe594d23bbed5906047e0069db0b0c59d89908acb1bf3395d4a51df16" exitCode=0 Nov 21 14:54:49 crc kubenswrapper[4897]: I1121 14:54:49.919635 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2hspl" event={"ID":"acb6440f-d6f8-4ef3-b05b-d1957afdd240","Type":"ContainerDied","Data":"41f3191fe594d23bbed5906047e0069db0b0c59d89908acb1bf3395d4a51df16"} Nov 21 14:54:50 crc kubenswrapper[4897]: I1121 14:54:50.930641 4897 generic.go:334] "Generic (PLEG): container finished" podID="c27aa401-13de-4f6e-beb1-d1683c852940" containerID="0bb79f630ec5fd12530a9a78048be4f84560f02746d5fbed01e2758c0a4d7dd9" exitCode=0 Nov 21 14:54:50 crc kubenswrapper[4897]: I1121 14:54:50.930719 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q6td4" event={"ID":"c27aa401-13de-4f6e-beb1-d1683c852940","Type":"ContainerDied","Data":"0bb79f630ec5fd12530a9a78048be4f84560f02746d5fbed01e2758c0a4d7dd9"} Nov 21 14:54:51 crc kubenswrapper[4897]: I1121 14:54:51.492423 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2hspl" Nov 21 14:54:51 crc kubenswrapper[4897]: I1121 14:54:51.652645 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w6kt9\" (UniqueName: \"kubernetes.io/projected/acb6440f-d6f8-4ef3-b05b-d1957afdd240-kube-api-access-w6kt9\") pod \"acb6440f-d6f8-4ef3-b05b-d1957afdd240\" (UID: \"acb6440f-d6f8-4ef3-b05b-d1957afdd240\") " Nov 21 14:54:51 crc kubenswrapper[4897]: I1121 14:54:51.652702 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/acb6440f-d6f8-4ef3-b05b-d1957afdd240-inventory\") pod \"acb6440f-d6f8-4ef3-b05b-d1957afdd240\" (UID: \"acb6440f-d6f8-4ef3-b05b-d1957afdd240\") " Nov 21 14:54:51 crc kubenswrapper[4897]: I1121 14:54:51.652858 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/acb6440f-d6f8-4ef3-b05b-d1957afdd240-ssh-key\") pod \"acb6440f-d6f8-4ef3-b05b-d1957afdd240\" (UID: \"acb6440f-d6f8-4ef3-b05b-d1957afdd240\") " Nov 21 14:54:51 crc kubenswrapper[4897]: I1121 14:54:51.672810 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/acb6440f-d6f8-4ef3-b05b-d1957afdd240-kube-api-access-w6kt9" (OuterVolumeSpecName: "kube-api-access-w6kt9") pod "acb6440f-d6f8-4ef3-b05b-d1957afdd240" (UID: "acb6440f-d6f8-4ef3-b05b-d1957afdd240"). InnerVolumeSpecName "kube-api-access-w6kt9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:54:51 crc kubenswrapper[4897]: I1121 14:54:51.755153 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w6kt9\" (UniqueName: \"kubernetes.io/projected/acb6440f-d6f8-4ef3-b05b-d1957afdd240-kube-api-access-w6kt9\") on node \"crc\" DevicePath \"\"" Nov 21 14:54:51 crc kubenswrapper[4897]: I1121 14:54:51.756635 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/acb6440f-d6f8-4ef3-b05b-d1957afdd240-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "acb6440f-d6f8-4ef3-b05b-d1957afdd240" (UID: "acb6440f-d6f8-4ef3-b05b-d1957afdd240"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:54:51 crc kubenswrapper[4897]: I1121 14:54:51.843103 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/acb6440f-d6f8-4ef3-b05b-d1957afdd240-inventory" (OuterVolumeSpecName: "inventory") pod "acb6440f-d6f8-4ef3-b05b-d1957afdd240" (UID: "acb6440f-d6f8-4ef3-b05b-d1957afdd240"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:54:51 crc kubenswrapper[4897]: I1121 14:54:51.859397 4897 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/acb6440f-d6f8-4ef3-b05b-d1957afdd240-inventory\") on node \"crc\" DevicePath \"\"" Nov 21 14:54:51 crc kubenswrapper[4897]: I1121 14:54:51.859440 4897 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/acb6440f-d6f8-4ef3-b05b-d1957afdd240-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 21 14:54:51 crc kubenswrapper[4897]: I1121 14:54:51.942846 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2hspl" event={"ID":"acb6440f-d6f8-4ef3-b05b-d1957afdd240","Type":"ContainerDied","Data":"344adf5734f7fb726c514912734219d69dae2215dfe8b8e77a38635b41ef9e04"} Nov 21 14:54:51 crc kubenswrapper[4897]: I1121 14:54:51.942890 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2hspl" Nov 21 14:54:51 crc kubenswrapper[4897]: I1121 14:54:51.942892 4897 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="344adf5734f7fb726c514912734219d69dae2215dfe8b8e77a38635b41ef9e04" Nov 21 14:54:52 crc kubenswrapper[4897]: I1121 14:54:52.072059 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qwzfr"] Nov 21 14:54:52 crc kubenswrapper[4897]: E1121 14:54:52.072888 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="acb6440f-d6f8-4ef3-b05b-d1957afdd240" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 21 14:54:52 crc kubenswrapper[4897]: I1121 14:54:52.072909 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="acb6440f-d6f8-4ef3-b05b-d1957afdd240" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 21 14:54:52 crc kubenswrapper[4897]: I1121 14:54:52.073156 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="acb6440f-d6f8-4ef3-b05b-d1957afdd240" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 21 14:54:52 crc kubenswrapper[4897]: I1121 14:54:52.074173 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qwzfr" Nov 21 14:54:52 crc kubenswrapper[4897]: I1121 14:54:52.086766 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 21 14:54:52 crc kubenswrapper[4897]: I1121 14:54:52.087037 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 21 14:54:52 crc kubenswrapper[4897]: I1121 14:54:52.087201 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-dj88m" Nov 21 14:54:52 crc kubenswrapper[4897]: I1121 14:54:52.087275 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 21 14:54:52 crc kubenswrapper[4897]: I1121 14:54:52.113734 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qwzfr"] Nov 21 14:54:52 crc kubenswrapper[4897]: I1121 14:54:52.168188 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7b06e5fb-282d-4372-b287-1cad5f621223-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-qwzfr\" (UID: \"7b06e5fb-282d-4372-b287-1cad5f621223\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qwzfr" Nov 21 14:54:52 crc kubenswrapper[4897]: I1121 14:54:52.168246 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6p2xb\" (UniqueName: \"kubernetes.io/projected/7b06e5fb-282d-4372-b287-1cad5f621223-kube-api-access-6p2xb\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-qwzfr\" (UID: \"7b06e5fb-282d-4372-b287-1cad5f621223\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qwzfr" Nov 21 14:54:52 crc kubenswrapper[4897]: I1121 14:54:52.168364 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7b06e5fb-282d-4372-b287-1cad5f621223-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-qwzfr\" (UID: \"7b06e5fb-282d-4372-b287-1cad5f621223\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qwzfr" Nov 21 14:54:52 crc kubenswrapper[4897]: I1121 14:54:52.270605 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7b06e5fb-282d-4372-b287-1cad5f621223-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-qwzfr\" (UID: \"7b06e5fb-282d-4372-b287-1cad5f621223\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qwzfr" Nov 21 14:54:52 crc kubenswrapper[4897]: I1121 14:54:52.270920 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6p2xb\" (UniqueName: \"kubernetes.io/projected/7b06e5fb-282d-4372-b287-1cad5f621223-kube-api-access-6p2xb\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-qwzfr\" (UID: \"7b06e5fb-282d-4372-b287-1cad5f621223\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qwzfr" Nov 21 14:54:52 crc kubenswrapper[4897]: I1121 14:54:52.271066 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7b06e5fb-282d-4372-b287-1cad5f621223-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-qwzfr\" (UID: \"7b06e5fb-282d-4372-b287-1cad5f621223\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qwzfr" Nov 21 14:54:52 crc kubenswrapper[4897]: I1121 14:54:52.274949 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7b06e5fb-282d-4372-b287-1cad5f621223-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-qwzfr\" (UID: \"7b06e5fb-282d-4372-b287-1cad5f621223\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qwzfr" Nov 21 14:54:52 crc kubenswrapper[4897]: I1121 14:54:52.276140 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7b06e5fb-282d-4372-b287-1cad5f621223-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-qwzfr\" (UID: \"7b06e5fb-282d-4372-b287-1cad5f621223\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qwzfr" Nov 21 14:54:52 crc kubenswrapper[4897]: I1121 14:54:52.288006 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6p2xb\" (UniqueName: \"kubernetes.io/projected/7b06e5fb-282d-4372-b287-1cad5f621223-kube-api-access-6p2xb\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-qwzfr\" (UID: \"7b06e5fb-282d-4372-b287-1cad5f621223\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qwzfr" Nov 21 14:54:52 crc kubenswrapper[4897]: I1121 14:54:52.411528 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qwzfr" Nov 21 14:54:54 crc kubenswrapper[4897]: W1121 14:54:53.309742 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7b06e5fb_282d_4372_b287_1cad5f621223.slice/crio-8e3efdafa82db64a9d9c4b08edd0a36d15d6ca28d7c75383805885a68300ef8d WatchSource:0}: Error finding container 8e3efdafa82db64a9d9c4b08edd0a36d15d6ca28d7c75383805885a68300ef8d: Status 404 returned error can't find the container with id 8e3efdafa82db64a9d9c4b08edd0a36d15d6ca28d7c75383805885a68300ef8d Nov 21 14:54:54 crc kubenswrapper[4897]: I1121 14:54:53.319687 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qwzfr"] Nov 21 14:54:54 crc kubenswrapper[4897]: I1121 14:54:53.973807 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qwzfr" event={"ID":"7b06e5fb-282d-4372-b287-1cad5f621223","Type":"ContainerStarted","Data":"8e3efdafa82db64a9d9c4b08edd0a36d15d6ca28d7c75383805885a68300ef8d"} Nov 21 14:54:54 crc kubenswrapper[4897]: I1121 14:54:53.977462 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q6td4" event={"ID":"c27aa401-13de-4f6e-beb1-d1683c852940","Type":"ContainerStarted","Data":"03445d84a2b7eb109f672a68001ec9b4b438f171cd7c0a448d8ae667289ae1b8"} Nov 21 14:54:54 crc kubenswrapper[4897]: I1121 14:54:54.007269 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-q6td4" podStartSLOduration=4.094297259 podStartE2EDuration="19.007251099s" podCreationTimestamp="2025-11-21 14:54:35 +0000 UTC" firstStartedPulling="2025-11-21 14:54:37.789701603 +0000 UTC m=+2755.074295078" lastFinishedPulling="2025-11-21 14:54:52.702655443 +0000 UTC m=+2769.987248918" observedRunningTime="2025-11-21 14:54:54.004568376 +0000 UTC m=+2771.289161861" watchObservedRunningTime="2025-11-21 14:54:54.007251099 +0000 UTC m=+2771.291844574" Nov 21 14:54:54 crc kubenswrapper[4897]: I1121 14:54:54.993860 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qwzfr" event={"ID":"7b06e5fb-282d-4372-b287-1cad5f621223","Type":"ContainerStarted","Data":"1436927a60122769a6edb652c4ad7ab79b1d23252d813ad379263ae067f2f650"} Nov 21 14:54:55 crc kubenswrapper[4897]: I1121 14:54:55.029372 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qwzfr" podStartSLOduration=2.437466908 podStartE2EDuration="3.029353895s" podCreationTimestamp="2025-11-21 14:54:52 +0000 UTC" firstStartedPulling="2025-11-21 14:54:53.314597953 +0000 UTC m=+2770.599191428" lastFinishedPulling="2025-11-21 14:54:53.90648494 +0000 UTC m=+2771.191078415" observedRunningTime="2025-11-21 14:54:55.025936912 +0000 UTC m=+2772.310530397" watchObservedRunningTime="2025-11-21 14:54:55.029353895 +0000 UTC m=+2772.313947370" Nov 21 14:54:55 crc kubenswrapper[4897]: I1121 14:54:55.966382 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-q6td4" Nov 21 14:54:55 crc kubenswrapper[4897]: I1121 14:54:55.966761 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-q6td4" Nov 21 14:54:56 crc kubenswrapper[4897]: I1121 14:54:56.023915 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-q6td4" Nov 21 14:55:04 crc kubenswrapper[4897]: I1121 14:55:04.093559 4897 generic.go:334] "Generic (PLEG): container finished" podID="7b06e5fb-282d-4372-b287-1cad5f621223" containerID="1436927a60122769a6edb652c4ad7ab79b1d23252d813ad379263ae067f2f650" exitCode=0 Nov 21 14:55:04 crc kubenswrapper[4897]: I1121 14:55:04.103755 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qwzfr" event={"ID":"7b06e5fb-282d-4372-b287-1cad5f621223","Type":"ContainerDied","Data":"1436927a60122769a6edb652c4ad7ab79b1d23252d813ad379263ae067f2f650"} Nov 21 14:55:04 crc kubenswrapper[4897]: I1121 14:55:04.371194 4897 patch_prober.go:28] interesting pod/machine-config-daemon-krv5b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 21 14:55:04 crc kubenswrapper[4897]: I1121 14:55:04.371288 4897 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 21 14:55:04 crc kubenswrapper[4897]: I1121 14:55:04.371358 4897 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" Nov 21 14:55:04 crc kubenswrapper[4897]: I1121 14:55:04.372566 4897 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"93211f44a4c3f56e3048bcba270e9ae8a65d8435c6fcfa8de9ac51f949ca9ee4"} pod="openshift-machine-config-operator/machine-config-daemon-krv5b" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 21 14:55:04 crc kubenswrapper[4897]: I1121 14:55:04.372640 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" containerName="machine-config-daemon" containerID="cri-o://93211f44a4c3f56e3048bcba270e9ae8a65d8435c6fcfa8de9ac51f949ca9ee4" gracePeriod=600 Nov 21 14:55:05 crc kubenswrapper[4897]: I1121 14:55:05.108459 4897 generic.go:334] "Generic (PLEG): container finished" podID="e7670227-d280-4847-b882-754429f56b0f" containerID="93211f44a4c3f56e3048bcba270e9ae8a65d8435c6fcfa8de9ac51f949ca9ee4" exitCode=0 Nov 21 14:55:05 crc kubenswrapper[4897]: I1121 14:55:05.108628 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" event={"ID":"e7670227-d280-4847-b882-754429f56b0f","Type":"ContainerDied","Data":"93211f44a4c3f56e3048bcba270e9ae8a65d8435c6fcfa8de9ac51f949ca9ee4"} Nov 21 14:55:05 crc kubenswrapper[4897]: I1121 14:55:05.108998 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" event={"ID":"e7670227-d280-4847-b882-754429f56b0f","Type":"ContainerStarted","Data":"a17b834dbb567d490c0f7ce8a22ab5aa6ecb645fc460c063af4ca43f58348563"} Nov 21 14:55:05 crc kubenswrapper[4897]: I1121 14:55:05.109020 4897 scope.go:117] "RemoveContainer" containerID="dc3c0a9877f70b7d5c55314a92df879b827a4103351566cc755cfbf916773aa2" Nov 21 14:55:05 crc kubenswrapper[4897]: I1121 14:55:05.561212 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qwzfr" Nov 21 14:55:05 crc kubenswrapper[4897]: I1121 14:55:05.595448 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6p2xb\" (UniqueName: \"kubernetes.io/projected/7b06e5fb-282d-4372-b287-1cad5f621223-kube-api-access-6p2xb\") pod \"7b06e5fb-282d-4372-b287-1cad5f621223\" (UID: \"7b06e5fb-282d-4372-b287-1cad5f621223\") " Nov 21 14:55:05 crc kubenswrapper[4897]: I1121 14:55:05.595636 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7b06e5fb-282d-4372-b287-1cad5f621223-inventory\") pod \"7b06e5fb-282d-4372-b287-1cad5f621223\" (UID: \"7b06e5fb-282d-4372-b287-1cad5f621223\") " Nov 21 14:55:05 crc kubenswrapper[4897]: I1121 14:55:05.595663 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7b06e5fb-282d-4372-b287-1cad5f621223-ssh-key\") pod \"7b06e5fb-282d-4372-b287-1cad5f621223\" (UID: \"7b06e5fb-282d-4372-b287-1cad5f621223\") " Nov 21 14:55:05 crc kubenswrapper[4897]: I1121 14:55:05.604316 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b06e5fb-282d-4372-b287-1cad5f621223-kube-api-access-6p2xb" (OuterVolumeSpecName: "kube-api-access-6p2xb") pod "7b06e5fb-282d-4372-b287-1cad5f621223" (UID: "7b06e5fb-282d-4372-b287-1cad5f621223"). InnerVolumeSpecName "kube-api-access-6p2xb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:55:05 crc kubenswrapper[4897]: I1121 14:55:05.633406 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b06e5fb-282d-4372-b287-1cad5f621223-inventory" (OuterVolumeSpecName: "inventory") pod "7b06e5fb-282d-4372-b287-1cad5f621223" (UID: "7b06e5fb-282d-4372-b287-1cad5f621223"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:55:05 crc kubenswrapper[4897]: I1121 14:55:05.657781 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b06e5fb-282d-4372-b287-1cad5f621223-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "7b06e5fb-282d-4372-b287-1cad5f621223" (UID: "7b06e5fb-282d-4372-b287-1cad5f621223"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:55:05 crc kubenswrapper[4897]: I1121 14:55:05.698073 4897 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7b06e5fb-282d-4372-b287-1cad5f621223-inventory\") on node \"crc\" DevicePath \"\"" Nov 21 14:55:05 crc kubenswrapper[4897]: I1121 14:55:05.698169 4897 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7b06e5fb-282d-4372-b287-1cad5f621223-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 21 14:55:05 crc kubenswrapper[4897]: I1121 14:55:05.698195 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6p2xb\" (UniqueName: \"kubernetes.io/projected/7b06e5fb-282d-4372-b287-1cad5f621223-kube-api-access-6p2xb\") on node \"crc\" DevicePath \"\"" Nov 21 14:55:06 crc kubenswrapper[4897]: I1121 14:55:06.023320 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-q6td4" Nov 21 14:55:06 crc kubenswrapper[4897]: I1121 14:55:06.080997 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-q6td4"] Nov 21 14:55:06 crc kubenswrapper[4897]: I1121 14:55:06.119732 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qwzfr" Nov 21 14:55:06 crc kubenswrapper[4897]: I1121 14:55:06.119711 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qwzfr" event={"ID":"7b06e5fb-282d-4372-b287-1cad5f621223","Type":"ContainerDied","Data":"8e3efdafa82db64a9d9c4b08edd0a36d15d6ca28d7c75383805885a68300ef8d"} Nov 21 14:55:06 crc kubenswrapper[4897]: I1121 14:55:06.119880 4897 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8e3efdafa82db64a9d9c4b08edd0a36d15d6ca28d7c75383805885a68300ef8d" Nov 21 14:55:06 crc kubenswrapper[4897]: I1121 14:55:06.122947 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-q6td4" podUID="c27aa401-13de-4f6e-beb1-d1683c852940" containerName="registry-server" containerID="cri-o://03445d84a2b7eb109f672a68001ec9b4b438f171cd7c0a448d8ae667289ae1b8" gracePeriod=2 Nov 21 14:55:06 crc kubenswrapper[4897]: I1121 14:55:06.247568 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr"] Nov 21 14:55:06 crc kubenswrapper[4897]: E1121 14:55:06.250393 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b06e5fb-282d-4372-b287-1cad5f621223" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 21 14:55:06 crc kubenswrapper[4897]: I1121 14:55:06.250425 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b06e5fb-282d-4372-b287-1cad5f621223" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 21 14:55:06 crc kubenswrapper[4897]: I1121 14:55:06.250737 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b06e5fb-282d-4372-b287-1cad5f621223" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 21 14:55:06 crc kubenswrapper[4897]: I1121 14:55:06.251928 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr" Nov 21 14:55:06 crc kubenswrapper[4897]: I1121 14:55:06.258624 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 21 14:55:06 crc kubenswrapper[4897]: I1121 14:55:06.258781 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Nov 21 14:55:06 crc kubenswrapper[4897]: I1121 14:55:06.258997 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Nov 21 14:55:06 crc kubenswrapper[4897]: I1121 14:55:06.259038 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 21 14:55:06 crc kubenswrapper[4897]: I1121 14:55:06.259579 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Nov 21 14:55:06 crc kubenswrapper[4897]: I1121 14:55:06.260353 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0" Nov 21 14:55:06 crc kubenswrapper[4897]: I1121 14:55:06.261000 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-dj88m" Nov 21 14:55:06 crc kubenswrapper[4897]: I1121 14:55:06.262026 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 21 14:55:06 crc kubenswrapper[4897]: I1121 14:55:06.262260 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Nov 21 14:55:06 crc kubenswrapper[4897]: I1121 14:55:06.289588 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr"] Nov 21 14:55:06 crc kubenswrapper[4897]: I1121 14:55:06.315942 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19d6619a-d692-43fd-b62b-19ca1c3c5ae3-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr\" (UID: \"19d6619a-d692-43fd-b62b-19ca1c3c5ae3\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr" Nov 21 14:55:06 crc kubenswrapper[4897]: I1121 14:55:06.316012 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/19d6619a-d692-43fd-b62b-19ca1c3c5ae3-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr\" (UID: \"19d6619a-d692-43fd-b62b-19ca1c3c5ae3\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr" Nov 21 14:55:06 crc kubenswrapper[4897]: I1121 14:55:06.316088 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/19d6619a-d692-43fd-b62b-19ca1c3c5ae3-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr\" (UID: \"19d6619a-d692-43fd-b62b-19ca1c3c5ae3\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr" Nov 21 14:55:06 crc kubenswrapper[4897]: I1121 14:55:06.316194 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/19d6619a-d692-43fd-b62b-19ca1c3c5ae3-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr\" (UID: \"19d6619a-d692-43fd-b62b-19ca1c3c5ae3\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr" Nov 21 14:55:06 crc kubenswrapper[4897]: I1121 14:55:06.316234 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/19d6619a-d692-43fd-b62b-19ca1c3c5ae3-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr\" (UID: \"19d6619a-d692-43fd-b62b-19ca1c3c5ae3\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr" Nov 21 14:55:06 crc kubenswrapper[4897]: I1121 14:55:06.316287 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19d6619a-d692-43fd-b62b-19ca1c3c5ae3-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr\" (UID: \"19d6619a-d692-43fd-b62b-19ca1c3c5ae3\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr" Nov 21 14:55:06 crc kubenswrapper[4897]: I1121 14:55:06.316314 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19d6619a-d692-43fd-b62b-19ca1c3c5ae3-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr\" (UID: \"19d6619a-d692-43fd-b62b-19ca1c3c5ae3\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr" Nov 21 14:55:06 crc kubenswrapper[4897]: I1121 14:55:06.316356 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19d6619a-d692-43fd-b62b-19ca1c3c5ae3-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr\" (UID: \"19d6619a-d692-43fd-b62b-19ca1c3c5ae3\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr" Nov 21 14:55:06 crc kubenswrapper[4897]: I1121 14:55:06.316439 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5hnc6\" (UniqueName: \"kubernetes.io/projected/19d6619a-d692-43fd-b62b-19ca1c3c5ae3-kube-api-access-5hnc6\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr\" (UID: \"19d6619a-d692-43fd-b62b-19ca1c3c5ae3\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr" Nov 21 14:55:06 crc kubenswrapper[4897]: I1121 14:55:06.316465 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/19d6619a-d692-43fd-b62b-19ca1c3c5ae3-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr\" (UID: \"19d6619a-d692-43fd-b62b-19ca1c3c5ae3\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr" Nov 21 14:55:06 crc kubenswrapper[4897]: I1121 14:55:06.316667 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/19d6619a-d692-43fd-b62b-19ca1c3c5ae3-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr\" (UID: \"19d6619a-d692-43fd-b62b-19ca1c3c5ae3\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr" Nov 21 14:55:06 crc kubenswrapper[4897]: I1121 14:55:06.316776 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19d6619a-d692-43fd-b62b-19ca1c3c5ae3-telemetry-power-monitoring-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr\" (UID: \"19d6619a-d692-43fd-b62b-19ca1c3c5ae3\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr" Nov 21 14:55:06 crc kubenswrapper[4897]: I1121 14:55:06.316873 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/19d6619a-d692-43fd-b62b-19ca1c3c5ae3-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr\" (UID: \"19d6619a-d692-43fd-b62b-19ca1c3c5ae3\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr" Nov 21 14:55:06 crc kubenswrapper[4897]: I1121 14:55:06.316947 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19d6619a-d692-43fd-b62b-19ca1c3c5ae3-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr\" (UID: \"19d6619a-d692-43fd-b62b-19ca1c3c5ae3\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr" Nov 21 14:55:06 crc kubenswrapper[4897]: I1121 14:55:06.317048 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19d6619a-d692-43fd-b62b-19ca1c3c5ae3-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr\" (UID: \"19d6619a-d692-43fd-b62b-19ca1c3c5ae3\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr" Nov 21 14:55:06 crc kubenswrapper[4897]: I1121 14:55:06.317211 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19d6619a-d692-43fd-b62b-19ca1c3c5ae3-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr\" (UID: \"19d6619a-d692-43fd-b62b-19ca1c3c5ae3\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr" Nov 21 14:55:06 crc kubenswrapper[4897]: I1121 14:55:06.421058 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19d6619a-d692-43fd-b62b-19ca1c3c5ae3-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr\" (UID: \"19d6619a-d692-43fd-b62b-19ca1c3c5ae3\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr" Nov 21 14:55:06 crc kubenswrapper[4897]: I1121 14:55:06.421230 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19d6619a-d692-43fd-b62b-19ca1c3c5ae3-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr\" (UID: \"19d6619a-d692-43fd-b62b-19ca1c3c5ae3\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr" Nov 21 14:55:06 crc kubenswrapper[4897]: I1121 14:55:06.421269 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/19d6619a-d692-43fd-b62b-19ca1c3c5ae3-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr\" (UID: \"19d6619a-d692-43fd-b62b-19ca1c3c5ae3\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr" Nov 21 14:55:06 crc kubenswrapper[4897]: I1121 14:55:06.421327 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/19d6619a-d692-43fd-b62b-19ca1c3c5ae3-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr\" (UID: \"19d6619a-d692-43fd-b62b-19ca1c3c5ae3\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr" Nov 21 14:55:06 crc kubenswrapper[4897]: I1121 14:55:06.421361 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/19d6619a-d692-43fd-b62b-19ca1c3c5ae3-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr\" (UID: \"19d6619a-d692-43fd-b62b-19ca1c3c5ae3\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr" Nov 21 14:55:06 crc kubenswrapper[4897]: I1121 14:55:06.421382 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/19d6619a-d692-43fd-b62b-19ca1c3c5ae3-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr\" (UID: \"19d6619a-d692-43fd-b62b-19ca1c3c5ae3\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr" Nov 21 14:55:06 crc kubenswrapper[4897]: I1121 14:55:06.421404 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19d6619a-d692-43fd-b62b-19ca1c3c5ae3-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr\" (UID: \"19d6619a-d692-43fd-b62b-19ca1c3c5ae3\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr" Nov 21 14:55:06 crc kubenswrapper[4897]: I1121 14:55:06.421426 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19d6619a-d692-43fd-b62b-19ca1c3c5ae3-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr\" (UID: \"19d6619a-d692-43fd-b62b-19ca1c3c5ae3\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr" Nov 21 14:55:06 crc kubenswrapper[4897]: I1121 14:55:06.421445 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19d6619a-d692-43fd-b62b-19ca1c3c5ae3-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr\" (UID: \"19d6619a-d692-43fd-b62b-19ca1c3c5ae3\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr" Nov 21 14:55:06 crc kubenswrapper[4897]: I1121 14:55:06.421500 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5hnc6\" (UniqueName: \"kubernetes.io/projected/19d6619a-d692-43fd-b62b-19ca1c3c5ae3-kube-api-access-5hnc6\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr\" (UID: \"19d6619a-d692-43fd-b62b-19ca1c3c5ae3\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr" Nov 21 14:55:06 crc kubenswrapper[4897]: I1121 14:55:06.421535 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/19d6619a-d692-43fd-b62b-19ca1c3c5ae3-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr\" (UID: \"19d6619a-d692-43fd-b62b-19ca1c3c5ae3\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr" Nov 21 14:55:06 crc kubenswrapper[4897]: I1121 14:55:06.421558 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/19d6619a-d692-43fd-b62b-19ca1c3c5ae3-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr\" (UID: \"19d6619a-d692-43fd-b62b-19ca1c3c5ae3\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr" Nov 21 14:55:06 crc kubenswrapper[4897]: I1121 14:55:06.421589 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19d6619a-d692-43fd-b62b-19ca1c3c5ae3-telemetry-power-monitoring-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr\" (UID: \"19d6619a-d692-43fd-b62b-19ca1c3c5ae3\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr" Nov 21 14:55:06 crc kubenswrapper[4897]: I1121 14:55:06.421619 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/19d6619a-d692-43fd-b62b-19ca1c3c5ae3-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr\" (UID: \"19d6619a-d692-43fd-b62b-19ca1c3c5ae3\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr" Nov 21 14:55:06 crc kubenswrapper[4897]: I1121 14:55:06.421650 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19d6619a-d692-43fd-b62b-19ca1c3c5ae3-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr\" (UID: \"19d6619a-d692-43fd-b62b-19ca1c3c5ae3\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr" Nov 21 14:55:06 crc kubenswrapper[4897]: I1121 14:55:06.421686 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19d6619a-d692-43fd-b62b-19ca1c3c5ae3-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr\" (UID: \"19d6619a-d692-43fd-b62b-19ca1c3c5ae3\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr" Nov 21 14:55:06 crc kubenswrapper[4897]: I1121 14:55:06.429268 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/19d6619a-d692-43fd-b62b-19ca1c3c5ae3-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr\" (UID: \"19d6619a-d692-43fd-b62b-19ca1c3c5ae3\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr" Nov 21 14:55:06 crc kubenswrapper[4897]: I1121 14:55:06.440312 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/19d6619a-d692-43fd-b62b-19ca1c3c5ae3-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr\" (UID: \"19d6619a-d692-43fd-b62b-19ca1c3c5ae3\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr" Nov 21 14:55:06 crc kubenswrapper[4897]: I1121 14:55:06.444737 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19d6619a-d692-43fd-b62b-19ca1c3c5ae3-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr\" (UID: \"19d6619a-d692-43fd-b62b-19ca1c3c5ae3\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr" Nov 21 14:55:06 crc kubenswrapper[4897]: I1121 14:55:06.445324 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/19d6619a-d692-43fd-b62b-19ca1c3c5ae3-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr\" (UID: \"19d6619a-d692-43fd-b62b-19ca1c3c5ae3\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr" Nov 21 14:55:06 crc kubenswrapper[4897]: I1121 14:55:06.458480 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/19d6619a-d692-43fd-b62b-19ca1c3c5ae3-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr\" (UID: \"19d6619a-d692-43fd-b62b-19ca1c3c5ae3\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr" Nov 21 14:55:06 crc kubenswrapper[4897]: I1121 14:55:06.458540 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19d6619a-d692-43fd-b62b-19ca1c3c5ae3-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr\" (UID: \"19d6619a-d692-43fd-b62b-19ca1c3c5ae3\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr" Nov 21 14:55:06 crc kubenswrapper[4897]: I1121 14:55:06.469899 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/19d6619a-d692-43fd-b62b-19ca1c3c5ae3-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr\" (UID: \"19d6619a-d692-43fd-b62b-19ca1c3c5ae3\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr" Nov 21 14:55:06 crc kubenswrapper[4897]: I1121 14:55:06.469946 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5hnc6\" (UniqueName: \"kubernetes.io/projected/19d6619a-d692-43fd-b62b-19ca1c3c5ae3-kube-api-access-5hnc6\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr\" (UID: \"19d6619a-d692-43fd-b62b-19ca1c3c5ae3\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr" Nov 21 14:55:06 crc kubenswrapper[4897]: I1121 14:55:06.470024 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/19d6619a-d692-43fd-b62b-19ca1c3c5ae3-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr\" (UID: \"19d6619a-d692-43fd-b62b-19ca1c3c5ae3\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr" Nov 21 14:55:06 crc kubenswrapper[4897]: I1121 14:55:06.470470 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19d6619a-d692-43fd-b62b-19ca1c3c5ae3-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr\" (UID: \"19d6619a-d692-43fd-b62b-19ca1c3c5ae3\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr" Nov 21 14:55:06 crc kubenswrapper[4897]: I1121 14:55:06.470785 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19d6619a-d692-43fd-b62b-19ca1c3c5ae3-telemetry-power-monitoring-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr\" (UID: \"19d6619a-d692-43fd-b62b-19ca1c3c5ae3\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr" Nov 21 14:55:06 crc kubenswrapper[4897]: I1121 14:55:06.471048 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19d6619a-d692-43fd-b62b-19ca1c3c5ae3-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr\" (UID: \"19d6619a-d692-43fd-b62b-19ca1c3c5ae3\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr" Nov 21 14:55:06 crc kubenswrapper[4897]: I1121 14:55:06.471180 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/19d6619a-d692-43fd-b62b-19ca1c3c5ae3-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr\" (UID: \"19d6619a-d692-43fd-b62b-19ca1c3c5ae3\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr" Nov 21 14:55:06 crc kubenswrapper[4897]: I1121 14:55:06.481637 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19d6619a-d692-43fd-b62b-19ca1c3c5ae3-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr\" (UID: \"19d6619a-d692-43fd-b62b-19ca1c3c5ae3\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr" Nov 21 14:55:06 crc kubenswrapper[4897]: I1121 14:55:06.485939 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19d6619a-d692-43fd-b62b-19ca1c3c5ae3-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr\" (UID: \"19d6619a-d692-43fd-b62b-19ca1c3c5ae3\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr" Nov 21 14:55:06 crc kubenswrapper[4897]: I1121 14:55:06.508366 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19d6619a-d692-43fd-b62b-19ca1c3c5ae3-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr\" (UID: \"19d6619a-d692-43fd-b62b-19ca1c3c5ae3\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr" Nov 21 14:55:06 crc kubenswrapper[4897]: I1121 14:55:06.669145 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr" Nov 21 14:55:06 crc kubenswrapper[4897]: I1121 14:55:06.847632 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-q6td4" Nov 21 14:55:06 crc kubenswrapper[4897]: I1121 14:55:06.938065 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c27aa401-13de-4f6e-beb1-d1683c852940-utilities\") pod \"c27aa401-13de-4f6e-beb1-d1683c852940\" (UID: \"c27aa401-13de-4f6e-beb1-d1683c852940\") " Nov 21 14:55:06 crc kubenswrapper[4897]: I1121 14:55:06.938398 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9rnkb\" (UniqueName: \"kubernetes.io/projected/c27aa401-13de-4f6e-beb1-d1683c852940-kube-api-access-9rnkb\") pod \"c27aa401-13de-4f6e-beb1-d1683c852940\" (UID: \"c27aa401-13de-4f6e-beb1-d1683c852940\") " Nov 21 14:55:06 crc kubenswrapper[4897]: I1121 14:55:06.938467 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c27aa401-13de-4f6e-beb1-d1683c852940-catalog-content\") pod \"c27aa401-13de-4f6e-beb1-d1683c852940\" (UID: \"c27aa401-13de-4f6e-beb1-d1683c852940\") " Nov 21 14:55:06 crc kubenswrapper[4897]: I1121 14:55:06.939257 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c27aa401-13de-4f6e-beb1-d1683c852940-utilities" (OuterVolumeSpecName: "utilities") pod "c27aa401-13de-4f6e-beb1-d1683c852940" (UID: "c27aa401-13de-4f6e-beb1-d1683c852940"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:55:06 crc kubenswrapper[4897]: I1121 14:55:06.945247 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c27aa401-13de-4f6e-beb1-d1683c852940-kube-api-access-9rnkb" (OuterVolumeSpecName: "kube-api-access-9rnkb") pod "c27aa401-13de-4f6e-beb1-d1683c852940" (UID: "c27aa401-13de-4f6e-beb1-d1683c852940"). InnerVolumeSpecName "kube-api-access-9rnkb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:55:06 crc kubenswrapper[4897]: I1121 14:55:06.995479 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c27aa401-13de-4f6e-beb1-d1683c852940-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c27aa401-13de-4f6e-beb1-d1683c852940" (UID: "c27aa401-13de-4f6e-beb1-d1683c852940"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:55:07 crc kubenswrapper[4897]: I1121 14:55:07.041225 4897 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c27aa401-13de-4f6e-beb1-d1683c852940-utilities\") on node \"crc\" DevicePath \"\"" Nov 21 14:55:07 crc kubenswrapper[4897]: I1121 14:55:07.041605 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9rnkb\" (UniqueName: \"kubernetes.io/projected/c27aa401-13de-4f6e-beb1-d1683c852940-kube-api-access-9rnkb\") on node \"crc\" DevicePath \"\"" Nov 21 14:55:07 crc kubenswrapper[4897]: I1121 14:55:07.041615 4897 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c27aa401-13de-4f6e-beb1-d1683c852940-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 21 14:55:07 crc kubenswrapper[4897]: I1121 14:55:07.142323 4897 generic.go:334] "Generic (PLEG): container finished" podID="c27aa401-13de-4f6e-beb1-d1683c852940" containerID="03445d84a2b7eb109f672a68001ec9b4b438f171cd7c0a448d8ae667289ae1b8" exitCode=0 Nov 21 14:55:07 crc kubenswrapper[4897]: I1121 14:55:07.142378 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q6td4" event={"ID":"c27aa401-13de-4f6e-beb1-d1683c852940","Type":"ContainerDied","Data":"03445d84a2b7eb109f672a68001ec9b4b438f171cd7c0a448d8ae667289ae1b8"} Nov 21 14:55:07 crc kubenswrapper[4897]: I1121 14:55:07.142455 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q6td4" event={"ID":"c27aa401-13de-4f6e-beb1-d1683c852940","Type":"ContainerDied","Data":"e49e53c317f4b31cc12a4ec3d30f70fb0937fd813576096fbef91aa7ce624b4a"} Nov 21 14:55:07 crc kubenswrapper[4897]: I1121 14:55:07.142479 4897 scope.go:117] "RemoveContainer" containerID="03445d84a2b7eb109f672a68001ec9b4b438f171cd7c0a448d8ae667289ae1b8" Nov 21 14:55:07 crc kubenswrapper[4897]: I1121 14:55:07.142414 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-q6td4" Nov 21 14:55:07 crc kubenswrapper[4897]: I1121 14:55:07.183141 4897 scope.go:117] "RemoveContainer" containerID="0bb79f630ec5fd12530a9a78048be4f84560f02746d5fbed01e2758c0a4d7dd9" Nov 21 14:55:07 crc kubenswrapper[4897]: I1121 14:55:07.194039 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-q6td4"] Nov 21 14:55:07 crc kubenswrapper[4897]: I1121 14:55:07.217881 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-q6td4"] Nov 21 14:55:07 crc kubenswrapper[4897]: I1121 14:55:07.228642 4897 scope.go:117] "RemoveContainer" containerID="1ef9e41a36284f61c96ae8e31011fb2e46e5b7a06be43b890d768d9d1c5a539c" Nov 21 14:55:07 crc kubenswrapper[4897]: I1121 14:55:07.276750 4897 scope.go:117] "RemoveContainer" containerID="03445d84a2b7eb109f672a68001ec9b4b438f171cd7c0a448d8ae667289ae1b8" Nov 21 14:55:07 crc kubenswrapper[4897]: E1121 14:55:07.279162 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"03445d84a2b7eb109f672a68001ec9b4b438f171cd7c0a448d8ae667289ae1b8\": container with ID starting with 03445d84a2b7eb109f672a68001ec9b4b438f171cd7c0a448d8ae667289ae1b8 not found: ID does not exist" containerID="03445d84a2b7eb109f672a68001ec9b4b438f171cd7c0a448d8ae667289ae1b8" Nov 21 14:55:07 crc kubenswrapper[4897]: I1121 14:55:07.279231 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"03445d84a2b7eb109f672a68001ec9b4b438f171cd7c0a448d8ae667289ae1b8"} err="failed to get container status \"03445d84a2b7eb109f672a68001ec9b4b438f171cd7c0a448d8ae667289ae1b8\": rpc error: code = NotFound desc = could not find container \"03445d84a2b7eb109f672a68001ec9b4b438f171cd7c0a448d8ae667289ae1b8\": container with ID starting with 03445d84a2b7eb109f672a68001ec9b4b438f171cd7c0a448d8ae667289ae1b8 not found: ID does not exist" Nov 21 14:55:07 crc kubenswrapper[4897]: I1121 14:55:07.279265 4897 scope.go:117] "RemoveContainer" containerID="0bb79f630ec5fd12530a9a78048be4f84560f02746d5fbed01e2758c0a4d7dd9" Nov 21 14:55:07 crc kubenswrapper[4897]: E1121 14:55:07.283076 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0bb79f630ec5fd12530a9a78048be4f84560f02746d5fbed01e2758c0a4d7dd9\": container with ID starting with 0bb79f630ec5fd12530a9a78048be4f84560f02746d5fbed01e2758c0a4d7dd9 not found: ID does not exist" containerID="0bb79f630ec5fd12530a9a78048be4f84560f02746d5fbed01e2758c0a4d7dd9" Nov 21 14:55:07 crc kubenswrapper[4897]: I1121 14:55:07.283123 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0bb79f630ec5fd12530a9a78048be4f84560f02746d5fbed01e2758c0a4d7dd9"} err="failed to get container status \"0bb79f630ec5fd12530a9a78048be4f84560f02746d5fbed01e2758c0a4d7dd9\": rpc error: code = NotFound desc = could not find container \"0bb79f630ec5fd12530a9a78048be4f84560f02746d5fbed01e2758c0a4d7dd9\": container with ID starting with 0bb79f630ec5fd12530a9a78048be4f84560f02746d5fbed01e2758c0a4d7dd9 not found: ID does not exist" Nov 21 14:55:07 crc kubenswrapper[4897]: I1121 14:55:07.283149 4897 scope.go:117] "RemoveContainer" containerID="1ef9e41a36284f61c96ae8e31011fb2e46e5b7a06be43b890d768d9d1c5a539c" Nov 21 14:55:07 crc kubenswrapper[4897]: E1121 14:55:07.283599 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ef9e41a36284f61c96ae8e31011fb2e46e5b7a06be43b890d768d9d1c5a539c\": container with ID starting with 1ef9e41a36284f61c96ae8e31011fb2e46e5b7a06be43b890d768d9d1c5a539c not found: ID does not exist" containerID="1ef9e41a36284f61c96ae8e31011fb2e46e5b7a06be43b890d768d9d1c5a539c" Nov 21 14:55:07 crc kubenswrapper[4897]: I1121 14:55:07.283641 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ef9e41a36284f61c96ae8e31011fb2e46e5b7a06be43b890d768d9d1c5a539c"} err="failed to get container status \"1ef9e41a36284f61c96ae8e31011fb2e46e5b7a06be43b890d768d9d1c5a539c\": rpc error: code = NotFound desc = could not find container \"1ef9e41a36284f61c96ae8e31011fb2e46e5b7a06be43b890d768d9d1c5a539c\": container with ID starting with 1ef9e41a36284f61c96ae8e31011fb2e46e5b7a06be43b890d768d9d1c5a539c not found: ID does not exist" Nov 21 14:55:07 crc kubenswrapper[4897]: I1121 14:55:07.292339 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr"] Nov 21 14:55:07 crc kubenswrapper[4897]: W1121 14:55:07.295706 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod19d6619a_d692_43fd_b62b_19ca1c3c5ae3.slice/crio-1fbd3ea3264eebf4fca0cf12d8374e6d29673ab4e18b4b6ab4f017f03056a381 WatchSource:0}: Error finding container 1fbd3ea3264eebf4fca0cf12d8374e6d29673ab4e18b4b6ab4f017f03056a381: Status 404 returned error can't find the container with id 1fbd3ea3264eebf4fca0cf12d8374e6d29673ab4e18b4b6ab4f017f03056a381 Nov 21 14:55:08 crc kubenswrapper[4897]: I1121 14:55:08.111538 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c27aa401-13de-4f6e-beb1-d1683c852940" path="/var/lib/kubelet/pods/c27aa401-13de-4f6e-beb1-d1683c852940/volumes" Nov 21 14:55:08 crc kubenswrapper[4897]: I1121 14:55:08.155217 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr" event={"ID":"19d6619a-d692-43fd-b62b-19ca1c3c5ae3","Type":"ContainerStarted","Data":"1fbd3ea3264eebf4fca0cf12d8374e6d29673ab4e18b4b6ab4f017f03056a381"} Nov 21 14:55:10 crc kubenswrapper[4897]: I1121 14:55:10.178846 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr" event={"ID":"19d6619a-d692-43fd-b62b-19ca1c3c5ae3","Type":"ContainerStarted","Data":"474d494a290ee1b0ba1adb212b3abad486108e453e4c37ba5427dba2b1c93473"} Nov 21 14:55:10 crc kubenswrapper[4897]: I1121 14:55:10.209119 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr" podStartSLOduration=1.7253273930000002 podStartE2EDuration="4.209096098s" podCreationTimestamp="2025-11-21 14:55:06 +0000 UTC" firstStartedPulling="2025-11-21 14:55:07.299265227 +0000 UTC m=+2784.583858702" lastFinishedPulling="2025-11-21 14:55:09.783033932 +0000 UTC m=+2787.067627407" observedRunningTime="2025-11-21 14:55:10.198031598 +0000 UTC m=+2787.482625103" watchObservedRunningTime="2025-11-21 14:55:10.209096098 +0000 UTC m=+2787.493689573" Nov 21 14:55:58 crc kubenswrapper[4897]: I1121 14:55:58.739065 4897 generic.go:334] "Generic (PLEG): container finished" podID="19d6619a-d692-43fd-b62b-19ca1c3c5ae3" containerID="474d494a290ee1b0ba1adb212b3abad486108e453e4c37ba5427dba2b1c93473" exitCode=0 Nov 21 14:55:58 crc kubenswrapper[4897]: I1121 14:55:58.739176 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr" event={"ID":"19d6619a-d692-43fd-b62b-19ca1c3c5ae3","Type":"ContainerDied","Data":"474d494a290ee1b0ba1adb212b3abad486108e453e4c37ba5427dba2b1c93473"} Nov 21 14:56:00 crc kubenswrapper[4897]: I1121 14:56:00.170849 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr" Nov 21 14:56:00 crc kubenswrapper[4897]: I1121 14:56:00.307603 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19d6619a-d692-43fd-b62b-19ca1c3c5ae3-neutron-metadata-combined-ca-bundle\") pod \"19d6619a-d692-43fd-b62b-19ca1c3c5ae3\" (UID: \"19d6619a-d692-43fd-b62b-19ca1c3c5ae3\") " Nov 21 14:56:00 crc kubenswrapper[4897]: I1121 14:56:00.307656 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19d6619a-d692-43fd-b62b-19ca1c3c5ae3-bootstrap-combined-ca-bundle\") pod \"19d6619a-d692-43fd-b62b-19ca1c3c5ae3\" (UID: \"19d6619a-d692-43fd-b62b-19ca1c3c5ae3\") " Nov 21 14:56:00 crc kubenswrapper[4897]: I1121 14:56:00.307719 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19d6619a-d692-43fd-b62b-19ca1c3c5ae3-ovn-combined-ca-bundle\") pod \"19d6619a-d692-43fd-b62b-19ca1c3c5ae3\" (UID: \"19d6619a-d692-43fd-b62b-19ca1c3c5ae3\") " Nov 21 14:56:00 crc kubenswrapper[4897]: I1121 14:56:00.307749 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19d6619a-d692-43fd-b62b-19ca1c3c5ae3-repo-setup-combined-ca-bundle\") pod \"19d6619a-d692-43fd-b62b-19ca1c3c5ae3\" (UID: \"19d6619a-d692-43fd-b62b-19ca1c3c5ae3\") " Nov 21 14:56:00 crc kubenswrapper[4897]: I1121 14:56:00.307798 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/19d6619a-d692-43fd-b62b-19ca1c3c5ae3-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"19d6619a-d692-43fd-b62b-19ca1c3c5ae3\" (UID: \"19d6619a-d692-43fd-b62b-19ca1c3c5ae3\") " Nov 21 14:56:00 crc kubenswrapper[4897]: I1121 14:56:00.307836 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/19d6619a-d692-43fd-b62b-19ca1c3c5ae3-openstack-edpm-ipam-ovn-default-certs-0\") pod \"19d6619a-d692-43fd-b62b-19ca1c3c5ae3\" (UID: \"19d6619a-d692-43fd-b62b-19ca1c3c5ae3\") " Nov 21 14:56:00 crc kubenswrapper[4897]: I1121 14:56:00.307882 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/19d6619a-d692-43fd-b62b-19ca1c3c5ae3-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"19d6619a-d692-43fd-b62b-19ca1c3c5ae3\" (UID: \"19d6619a-d692-43fd-b62b-19ca1c3c5ae3\") " Nov 21 14:56:00 crc kubenswrapper[4897]: I1121 14:56:00.307901 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/19d6619a-d692-43fd-b62b-19ca1c3c5ae3-ssh-key\") pod \"19d6619a-d692-43fd-b62b-19ca1c3c5ae3\" (UID: \"19d6619a-d692-43fd-b62b-19ca1c3c5ae3\") " Nov 21 14:56:00 crc kubenswrapper[4897]: I1121 14:56:00.307917 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19d6619a-d692-43fd-b62b-19ca1c3c5ae3-libvirt-combined-ca-bundle\") pod \"19d6619a-d692-43fd-b62b-19ca1c3c5ae3\" (UID: \"19d6619a-d692-43fd-b62b-19ca1c3c5ae3\") " Nov 21 14:56:00 crc kubenswrapper[4897]: I1121 14:56:00.307955 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/19d6619a-d692-43fd-b62b-19ca1c3c5ae3-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") pod \"19d6619a-d692-43fd-b62b-19ca1c3c5ae3\" (UID: \"19d6619a-d692-43fd-b62b-19ca1c3c5ae3\") " Nov 21 14:56:00 crc kubenswrapper[4897]: I1121 14:56:00.308052 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/19d6619a-d692-43fd-b62b-19ca1c3c5ae3-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"19d6619a-d692-43fd-b62b-19ca1c3c5ae3\" (UID: \"19d6619a-d692-43fd-b62b-19ca1c3c5ae3\") " Nov 21 14:56:00 crc kubenswrapper[4897]: I1121 14:56:00.308068 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19d6619a-d692-43fd-b62b-19ca1c3c5ae3-nova-combined-ca-bundle\") pod \"19d6619a-d692-43fd-b62b-19ca1c3c5ae3\" (UID: \"19d6619a-d692-43fd-b62b-19ca1c3c5ae3\") " Nov 21 14:56:00 crc kubenswrapper[4897]: I1121 14:56:00.308153 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19d6619a-d692-43fd-b62b-19ca1c3c5ae3-telemetry-power-monitoring-combined-ca-bundle\") pod \"19d6619a-d692-43fd-b62b-19ca1c3c5ae3\" (UID: \"19d6619a-d692-43fd-b62b-19ca1c3c5ae3\") " Nov 21 14:56:00 crc kubenswrapper[4897]: I1121 14:56:00.308180 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/19d6619a-d692-43fd-b62b-19ca1c3c5ae3-inventory\") pod \"19d6619a-d692-43fd-b62b-19ca1c3c5ae3\" (UID: \"19d6619a-d692-43fd-b62b-19ca1c3c5ae3\") " Nov 21 14:56:00 crc kubenswrapper[4897]: I1121 14:56:00.308203 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19d6619a-d692-43fd-b62b-19ca1c3c5ae3-telemetry-combined-ca-bundle\") pod \"19d6619a-d692-43fd-b62b-19ca1c3c5ae3\" (UID: \"19d6619a-d692-43fd-b62b-19ca1c3c5ae3\") " Nov 21 14:56:00 crc kubenswrapper[4897]: I1121 14:56:00.308234 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5hnc6\" (UniqueName: \"kubernetes.io/projected/19d6619a-d692-43fd-b62b-19ca1c3c5ae3-kube-api-access-5hnc6\") pod \"19d6619a-d692-43fd-b62b-19ca1c3c5ae3\" (UID: \"19d6619a-d692-43fd-b62b-19ca1c3c5ae3\") " Nov 21 14:56:00 crc kubenswrapper[4897]: I1121 14:56:00.315325 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19d6619a-d692-43fd-b62b-19ca1c3c5ae3-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "19d6619a-d692-43fd-b62b-19ca1c3c5ae3" (UID: "19d6619a-d692-43fd-b62b-19ca1c3c5ae3"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:56:00 crc kubenswrapper[4897]: I1121 14:56:00.316169 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19d6619a-d692-43fd-b62b-19ca1c3c5ae3-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "19d6619a-d692-43fd-b62b-19ca1c3c5ae3" (UID: "19d6619a-d692-43fd-b62b-19ca1c3c5ae3"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:56:00 crc kubenswrapper[4897]: I1121 14:56:00.316999 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19d6619a-d692-43fd-b62b-19ca1c3c5ae3-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "19d6619a-d692-43fd-b62b-19ca1c3c5ae3" (UID: "19d6619a-d692-43fd-b62b-19ca1c3c5ae3"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:56:00 crc kubenswrapper[4897]: I1121 14:56:00.317523 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19d6619a-d692-43fd-b62b-19ca1c3c5ae3-kube-api-access-5hnc6" (OuterVolumeSpecName: "kube-api-access-5hnc6") pod "19d6619a-d692-43fd-b62b-19ca1c3c5ae3" (UID: "19d6619a-d692-43fd-b62b-19ca1c3c5ae3"). InnerVolumeSpecName "kube-api-access-5hnc6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:56:00 crc kubenswrapper[4897]: I1121 14:56:00.317697 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19d6619a-d692-43fd-b62b-19ca1c3c5ae3-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "19d6619a-d692-43fd-b62b-19ca1c3c5ae3" (UID: "19d6619a-d692-43fd-b62b-19ca1c3c5ae3"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:56:00 crc kubenswrapper[4897]: I1121 14:56:00.317845 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19d6619a-d692-43fd-b62b-19ca1c3c5ae3-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0") pod "19d6619a-d692-43fd-b62b-19ca1c3c5ae3" (UID: "19d6619a-d692-43fd-b62b-19ca1c3c5ae3"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:56:00 crc kubenswrapper[4897]: I1121 14:56:00.318284 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19d6619a-d692-43fd-b62b-19ca1c3c5ae3-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "19d6619a-d692-43fd-b62b-19ca1c3c5ae3" (UID: "19d6619a-d692-43fd-b62b-19ca1c3c5ae3"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:56:00 crc kubenswrapper[4897]: I1121 14:56:00.319203 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19d6619a-d692-43fd-b62b-19ca1c3c5ae3-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "19d6619a-d692-43fd-b62b-19ca1c3c5ae3" (UID: "19d6619a-d692-43fd-b62b-19ca1c3c5ae3"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:56:00 crc kubenswrapper[4897]: I1121 14:56:00.319813 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19d6619a-d692-43fd-b62b-19ca1c3c5ae3-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "19d6619a-d692-43fd-b62b-19ca1c3c5ae3" (UID: "19d6619a-d692-43fd-b62b-19ca1c3c5ae3"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:56:00 crc kubenswrapper[4897]: I1121 14:56:00.320770 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19d6619a-d692-43fd-b62b-19ca1c3c5ae3-telemetry-power-monitoring-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-power-monitoring-combined-ca-bundle") pod "19d6619a-d692-43fd-b62b-19ca1c3c5ae3" (UID: "19d6619a-d692-43fd-b62b-19ca1c3c5ae3"). InnerVolumeSpecName "telemetry-power-monitoring-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:56:00 crc kubenswrapper[4897]: I1121 14:56:00.321640 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19d6619a-d692-43fd-b62b-19ca1c3c5ae3-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "19d6619a-d692-43fd-b62b-19ca1c3c5ae3" (UID: "19d6619a-d692-43fd-b62b-19ca1c3c5ae3"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:56:00 crc kubenswrapper[4897]: I1121 14:56:00.322286 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19d6619a-d692-43fd-b62b-19ca1c3c5ae3-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "19d6619a-d692-43fd-b62b-19ca1c3c5ae3" (UID: "19d6619a-d692-43fd-b62b-19ca1c3c5ae3"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:56:00 crc kubenswrapper[4897]: I1121 14:56:00.325025 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19d6619a-d692-43fd-b62b-19ca1c3c5ae3-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "19d6619a-d692-43fd-b62b-19ca1c3c5ae3" (UID: "19d6619a-d692-43fd-b62b-19ca1c3c5ae3"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:56:00 crc kubenswrapper[4897]: I1121 14:56:00.325707 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19d6619a-d692-43fd-b62b-19ca1c3c5ae3-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "19d6619a-d692-43fd-b62b-19ca1c3c5ae3" (UID: "19d6619a-d692-43fd-b62b-19ca1c3c5ae3"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:56:00 crc kubenswrapper[4897]: I1121 14:56:00.357059 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19d6619a-d692-43fd-b62b-19ca1c3c5ae3-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "19d6619a-d692-43fd-b62b-19ca1c3c5ae3" (UID: "19d6619a-d692-43fd-b62b-19ca1c3c5ae3"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:56:00 crc kubenswrapper[4897]: I1121 14:56:00.359089 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19d6619a-d692-43fd-b62b-19ca1c3c5ae3-inventory" (OuterVolumeSpecName: "inventory") pod "19d6619a-d692-43fd-b62b-19ca1c3c5ae3" (UID: "19d6619a-d692-43fd-b62b-19ca1c3c5ae3"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:56:00 crc kubenswrapper[4897]: I1121 14:56:00.411557 4897 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/19d6619a-d692-43fd-b62b-19ca1c3c5ae3-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 21 14:56:00 crc kubenswrapper[4897]: I1121 14:56:00.411597 4897 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19d6619a-d692-43fd-b62b-19ca1c3c5ae3-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 21 14:56:00 crc kubenswrapper[4897]: I1121 14:56:00.411612 4897 reconciler_common.go:293] "Volume detached for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19d6619a-d692-43fd-b62b-19ca1c3c5ae3-telemetry-power-monitoring-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 21 14:56:00 crc kubenswrapper[4897]: I1121 14:56:00.411625 4897 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/19d6619a-d692-43fd-b62b-19ca1c3c5ae3-inventory\") on node \"crc\" DevicePath \"\"" Nov 21 14:56:00 crc kubenswrapper[4897]: I1121 14:56:00.411638 4897 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19d6619a-d692-43fd-b62b-19ca1c3c5ae3-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 21 14:56:00 crc kubenswrapper[4897]: I1121 14:56:00.411653 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5hnc6\" (UniqueName: \"kubernetes.io/projected/19d6619a-d692-43fd-b62b-19ca1c3c5ae3-kube-api-access-5hnc6\") on node \"crc\" DevicePath \"\"" Nov 21 14:56:00 crc kubenswrapper[4897]: I1121 14:56:00.411666 4897 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19d6619a-d692-43fd-b62b-19ca1c3c5ae3-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 21 14:56:00 crc kubenswrapper[4897]: I1121 14:56:00.411678 4897 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19d6619a-d692-43fd-b62b-19ca1c3c5ae3-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 21 14:56:00 crc kubenswrapper[4897]: I1121 14:56:00.411692 4897 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19d6619a-d692-43fd-b62b-19ca1c3c5ae3-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 21 14:56:00 crc kubenswrapper[4897]: I1121 14:56:00.411704 4897 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19d6619a-d692-43fd-b62b-19ca1c3c5ae3-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 21 14:56:00 crc kubenswrapper[4897]: I1121 14:56:00.411717 4897 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/19d6619a-d692-43fd-b62b-19ca1c3c5ae3-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 21 14:56:00 crc kubenswrapper[4897]: I1121 14:56:00.411730 4897 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/19d6619a-d692-43fd-b62b-19ca1c3c5ae3-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 21 14:56:00 crc kubenswrapper[4897]: I1121 14:56:00.411743 4897 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/19d6619a-d692-43fd-b62b-19ca1c3c5ae3-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 21 14:56:00 crc kubenswrapper[4897]: I1121 14:56:00.411755 4897 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/19d6619a-d692-43fd-b62b-19ca1c3c5ae3-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 21 14:56:00 crc kubenswrapper[4897]: I1121 14:56:00.411769 4897 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19d6619a-d692-43fd-b62b-19ca1c3c5ae3-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 21 14:56:00 crc kubenswrapper[4897]: I1121 14:56:00.411782 4897 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/19d6619a-d692-43fd-b62b-19ca1c3c5ae3-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 21 14:56:00 crc kubenswrapper[4897]: I1121 14:56:00.761574 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr" event={"ID":"19d6619a-d692-43fd-b62b-19ca1c3c5ae3","Type":"ContainerDied","Data":"1fbd3ea3264eebf4fca0cf12d8374e6d29673ab4e18b4b6ab4f017f03056a381"} Nov 21 14:56:00 crc kubenswrapper[4897]: I1121 14:56:00.761873 4897 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1fbd3ea3264eebf4fca0cf12d8374e6d29673ab4e18b4b6ab4f017f03056a381" Nov 21 14:56:00 crc kubenswrapper[4897]: I1121 14:56:00.761640 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-gjsjr" Nov 21 14:56:00 crc kubenswrapper[4897]: I1121 14:56:00.866030 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-x8zrj"] Nov 21 14:56:00 crc kubenswrapper[4897]: E1121 14:56:00.866603 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c27aa401-13de-4f6e-beb1-d1683c852940" containerName="extract-content" Nov 21 14:56:00 crc kubenswrapper[4897]: I1121 14:56:00.866630 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="c27aa401-13de-4f6e-beb1-d1683c852940" containerName="extract-content" Nov 21 14:56:00 crc kubenswrapper[4897]: E1121 14:56:00.866673 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19d6619a-d692-43fd-b62b-19ca1c3c5ae3" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Nov 21 14:56:00 crc kubenswrapper[4897]: I1121 14:56:00.866684 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="19d6619a-d692-43fd-b62b-19ca1c3c5ae3" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Nov 21 14:56:00 crc kubenswrapper[4897]: E1121 14:56:00.866702 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c27aa401-13de-4f6e-beb1-d1683c852940" containerName="registry-server" Nov 21 14:56:00 crc kubenswrapper[4897]: I1121 14:56:00.866711 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="c27aa401-13de-4f6e-beb1-d1683c852940" containerName="registry-server" Nov 21 14:56:00 crc kubenswrapper[4897]: E1121 14:56:00.866742 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c27aa401-13de-4f6e-beb1-d1683c852940" containerName="extract-utilities" Nov 21 14:56:00 crc kubenswrapper[4897]: I1121 14:56:00.866750 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="c27aa401-13de-4f6e-beb1-d1683c852940" containerName="extract-utilities" Nov 21 14:56:00 crc kubenswrapper[4897]: I1121 14:56:00.867021 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="19d6619a-d692-43fd-b62b-19ca1c3c5ae3" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Nov 21 14:56:00 crc kubenswrapper[4897]: I1121 14:56:00.867045 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="c27aa401-13de-4f6e-beb1-d1683c852940" containerName="registry-server" Nov 21 14:56:00 crc kubenswrapper[4897]: I1121 14:56:00.868082 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-x8zrj" Nov 21 14:56:00 crc kubenswrapper[4897]: I1121 14:56:00.870096 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 21 14:56:00 crc kubenswrapper[4897]: I1121 14:56:00.870683 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 21 14:56:00 crc kubenswrapper[4897]: I1121 14:56:00.875420 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Nov 21 14:56:00 crc kubenswrapper[4897]: I1121 14:56:00.875715 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 21 14:56:00 crc kubenswrapper[4897]: I1121 14:56:00.875868 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-dj88m" Nov 21 14:56:00 crc kubenswrapper[4897]: I1121 14:56:00.885216 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-x8zrj"] Nov 21 14:56:01 crc kubenswrapper[4897]: I1121 14:56:01.027044 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4855262f-81f6-44ee-b472-7a99a94b1b7f-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-x8zrj\" (UID: \"4855262f-81f6-44ee-b472-7a99a94b1b7f\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-x8zrj" Nov 21 14:56:01 crc kubenswrapper[4897]: I1121 14:56:01.027606 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4855262f-81f6-44ee-b472-7a99a94b1b7f-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-x8zrj\" (UID: \"4855262f-81f6-44ee-b472-7a99a94b1b7f\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-x8zrj" Nov 21 14:56:01 crc kubenswrapper[4897]: I1121 14:56:01.027691 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8gv9v\" (UniqueName: \"kubernetes.io/projected/4855262f-81f6-44ee-b472-7a99a94b1b7f-kube-api-access-8gv9v\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-x8zrj\" (UID: \"4855262f-81f6-44ee-b472-7a99a94b1b7f\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-x8zrj" Nov 21 14:56:01 crc kubenswrapper[4897]: I1121 14:56:01.027750 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/4855262f-81f6-44ee-b472-7a99a94b1b7f-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-x8zrj\" (UID: \"4855262f-81f6-44ee-b472-7a99a94b1b7f\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-x8zrj" Nov 21 14:56:01 crc kubenswrapper[4897]: I1121 14:56:01.028234 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4855262f-81f6-44ee-b472-7a99a94b1b7f-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-x8zrj\" (UID: \"4855262f-81f6-44ee-b472-7a99a94b1b7f\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-x8zrj" Nov 21 14:56:01 crc kubenswrapper[4897]: I1121 14:56:01.130633 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8gv9v\" (UniqueName: \"kubernetes.io/projected/4855262f-81f6-44ee-b472-7a99a94b1b7f-kube-api-access-8gv9v\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-x8zrj\" (UID: \"4855262f-81f6-44ee-b472-7a99a94b1b7f\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-x8zrj" Nov 21 14:56:01 crc kubenswrapper[4897]: I1121 14:56:01.130716 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/4855262f-81f6-44ee-b472-7a99a94b1b7f-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-x8zrj\" (UID: \"4855262f-81f6-44ee-b472-7a99a94b1b7f\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-x8zrj" Nov 21 14:56:01 crc kubenswrapper[4897]: I1121 14:56:01.130852 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4855262f-81f6-44ee-b472-7a99a94b1b7f-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-x8zrj\" (UID: \"4855262f-81f6-44ee-b472-7a99a94b1b7f\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-x8zrj" Nov 21 14:56:01 crc kubenswrapper[4897]: I1121 14:56:01.130892 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4855262f-81f6-44ee-b472-7a99a94b1b7f-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-x8zrj\" (UID: \"4855262f-81f6-44ee-b472-7a99a94b1b7f\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-x8zrj" Nov 21 14:56:01 crc kubenswrapper[4897]: I1121 14:56:01.130959 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4855262f-81f6-44ee-b472-7a99a94b1b7f-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-x8zrj\" (UID: \"4855262f-81f6-44ee-b472-7a99a94b1b7f\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-x8zrj" Nov 21 14:56:01 crc kubenswrapper[4897]: I1121 14:56:01.132143 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/4855262f-81f6-44ee-b472-7a99a94b1b7f-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-x8zrj\" (UID: \"4855262f-81f6-44ee-b472-7a99a94b1b7f\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-x8zrj" Nov 21 14:56:01 crc kubenswrapper[4897]: I1121 14:56:01.136525 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4855262f-81f6-44ee-b472-7a99a94b1b7f-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-x8zrj\" (UID: \"4855262f-81f6-44ee-b472-7a99a94b1b7f\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-x8zrj" Nov 21 14:56:01 crc kubenswrapper[4897]: I1121 14:56:01.136571 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4855262f-81f6-44ee-b472-7a99a94b1b7f-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-x8zrj\" (UID: \"4855262f-81f6-44ee-b472-7a99a94b1b7f\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-x8zrj" Nov 21 14:56:01 crc kubenswrapper[4897]: I1121 14:56:01.137947 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4855262f-81f6-44ee-b472-7a99a94b1b7f-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-x8zrj\" (UID: \"4855262f-81f6-44ee-b472-7a99a94b1b7f\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-x8zrj" Nov 21 14:56:01 crc kubenswrapper[4897]: I1121 14:56:01.154416 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8gv9v\" (UniqueName: \"kubernetes.io/projected/4855262f-81f6-44ee-b472-7a99a94b1b7f-kube-api-access-8gv9v\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-x8zrj\" (UID: \"4855262f-81f6-44ee-b472-7a99a94b1b7f\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-x8zrj" Nov 21 14:56:01 crc kubenswrapper[4897]: I1121 14:56:01.188649 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-x8zrj" Nov 21 14:56:01 crc kubenswrapper[4897]: I1121 14:56:01.740023 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-x8zrj"] Nov 21 14:56:01 crc kubenswrapper[4897]: I1121 14:56:01.742386 4897 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 21 14:56:01 crc kubenswrapper[4897]: I1121 14:56:01.778029 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-x8zrj" event={"ID":"4855262f-81f6-44ee-b472-7a99a94b1b7f","Type":"ContainerStarted","Data":"5a5abac618128685ac3d58e7dd63ff2fd962faf662d1018699d86d1dc194c07c"} Nov 21 14:56:04 crc kubenswrapper[4897]: I1121 14:56:04.816020 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-x8zrj" event={"ID":"4855262f-81f6-44ee-b472-7a99a94b1b7f","Type":"ContainerStarted","Data":"c27a120fd7dbeb5fa1b56c51fd92477c83a48c2a7622836d74d76739ab64e42f"} Nov 21 14:56:05 crc kubenswrapper[4897]: I1121 14:56:05.862905 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-x8zrj" podStartSLOduration=3.5884030510000002 podStartE2EDuration="5.862880549s" podCreationTimestamp="2025-11-21 14:56:00 +0000 UTC" firstStartedPulling="2025-11-21 14:56:01.742110718 +0000 UTC m=+2839.026704193" lastFinishedPulling="2025-11-21 14:56:04.016588216 +0000 UTC m=+2841.301181691" observedRunningTime="2025-11-21 14:56:05.848166591 +0000 UTC m=+2843.132760066" watchObservedRunningTime="2025-11-21 14:56:05.862880549 +0000 UTC m=+2843.147474034" Nov 21 14:57:04 crc kubenswrapper[4897]: I1121 14:57:04.371092 4897 patch_prober.go:28] interesting pod/machine-config-daemon-krv5b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 21 14:57:04 crc kubenswrapper[4897]: I1121 14:57:04.371806 4897 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 21 14:57:05 crc kubenswrapper[4897]: I1121 14:57:05.710569 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-8g9w2"] Nov 21 14:57:05 crc kubenswrapper[4897]: I1121 14:57:05.713427 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8g9w2" Nov 21 14:57:05 crc kubenswrapper[4897]: I1121 14:57:05.741198 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8g9w2"] Nov 21 14:57:05 crc kubenswrapper[4897]: I1121 14:57:05.870430 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-58259\" (UniqueName: \"kubernetes.io/projected/b636617a-cfba-4203-b0ac-05df18bcc511-kube-api-access-58259\") pod \"redhat-operators-8g9w2\" (UID: \"b636617a-cfba-4203-b0ac-05df18bcc511\") " pod="openshift-marketplace/redhat-operators-8g9w2" Nov 21 14:57:05 crc kubenswrapper[4897]: I1121 14:57:05.871376 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b636617a-cfba-4203-b0ac-05df18bcc511-catalog-content\") pod \"redhat-operators-8g9w2\" (UID: \"b636617a-cfba-4203-b0ac-05df18bcc511\") " pod="openshift-marketplace/redhat-operators-8g9w2" Nov 21 14:57:05 crc kubenswrapper[4897]: I1121 14:57:05.871447 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b636617a-cfba-4203-b0ac-05df18bcc511-utilities\") pod \"redhat-operators-8g9w2\" (UID: \"b636617a-cfba-4203-b0ac-05df18bcc511\") " pod="openshift-marketplace/redhat-operators-8g9w2" Nov 21 14:57:05 crc kubenswrapper[4897]: I1121 14:57:05.973918 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-58259\" (UniqueName: \"kubernetes.io/projected/b636617a-cfba-4203-b0ac-05df18bcc511-kube-api-access-58259\") pod \"redhat-operators-8g9w2\" (UID: \"b636617a-cfba-4203-b0ac-05df18bcc511\") " pod="openshift-marketplace/redhat-operators-8g9w2" Nov 21 14:57:05 crc kubenswrapper[4897]: I1121 14:57:05.974184 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b636617a-cfba-4203-b0ac-05df18bcc511-catalog-content\") pod \"redhat-operators-8g9w2\" (UID: \"b636617a-cfba-4203-b0ac-05df18bcc511\") " pod="openshift-marketplace/redhat-operators-8g9w2" Nov 21 14:57:05 crc kubenswrapper[4897]: I1121 14:57:05.974229 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b636617a-cfba-4203-b0ac-05df18bcc511-utilities\") pod \"redhat-operators-8g9w2\" (UID: \"b636617a-cfba-4203-b0ac-05df18bcc511\") " pod="openshift-marketplace/redhat-operators-8g9w2" Nov 21 14:57:05 crc kubenswrapper[4897]: I1121 14:57:05.974756 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b636617a-cfba-4203-b0ac-05df18bcc511-catalog-content\") pod \"redhat-operators-8g9w2\" (UID: \"b636617a-cfba-4203-b0ac-05df18bcc511\") " pod="openshift-marketplace/redhat-operators-8g9w2" Nov 21 14:57:05 crc kubenswrapper[4897]: I1121 14:57:05.974796 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b636617a-cfba-4203-b0ac-05df18bcc511-utilities\") pod \"redhat-operators-8g9w2\" (UID: \"b636617a-cfba-4203-b0ac-05df18bcc511\") " pod="openshift-marketplace/redhat-operators-8g9w2" Nov 21 14:57:06 crc kubenswrapper[4897]: I1121 14:57:06.002917 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-58259\" (UniqueName: \"kubernetes.io/projected/b636617a-cfba-4203-b0ac-05df18bcc511-kube-api-access-58259\") pod \"redhat-operators-8g9w2\" (UID: \"b636617a-cfba-4203-b0ac-05df18bcc511\") " pod="openshift-marketplace/redhat-operators-8g9w2" Nov 21 14:57:06 crc kubenswrapper[4897]: I1121 14:57:06.035041 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8g9w2" Nov 21 14:57:06 crc kubenswrapper[4897]: I1121 14:57:06.577950 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8g9w2"] Nov 21 14:57:07 crc kubenswrapper[4897]: I1121 14:57:07.527670 4897 generic.go:334] "Generic (PLEG): container finished" podID="b636617a-cfba-4203-b0ac-05df18bcc511" containerID="297e196a896f626226136978df6c4a65592611e29cb7faf343c1b7c66f4b5d15" exitCode=0 Nov 21 14:57:07 crc kubenswrapper[4897]: I1121 14:57:07.527843 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8g9w2" event={"ID":"b636617a-cfba-4203-b0ac-05df18bcc511","Type":"ContainerDied","Data":"297e196a896f626226136978df6c4a65592611e29cb7faf343c1b7c66f4b5d15"} Nov 21 14:57:07 crc kubenswrapper[4897]: I1121 14:57:07.528002 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8g9w2" event={"ID":"b636617a-cfba-4203-b0ac-05df18bcc511","Type":"ContainerStarted","Data":"f20070f7c0c5ed9e50447185796ff29974429de1ff90746be9861b88126a9e02"} Nov 21 14:57:10 crc kubenswrapper[4897]: I1121 14:57:10.561630 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8g9w2" event={"ID":"b636617a-cfba-4203-b0ac-05df18bcc511","Type":"ContainerStarted","Data":"52f6b003837aac81a78f9bdf86267891d3d1512dceac9832ca27b728b07aa50e"} Nov 21 14:57:13 crc kubenswrapper[4897]: I1121 14:57:13.592226 4897 generic.go:334] "Generic (PLEG): container finished" podID="4855262f-81f6-44ee-b472-7a99a94b1b7f" containerID="c27a120fd7dbeb5fa1b56c51fd92477c83a48c2a7622836d74d76739ab64e42f" exitCode=0 Nov 21 14:57:13 crc kubenswrapper[4897]: I1121 14:57:13.592497 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-x8zrj" event={"ID":"4855262f-81f6-44ee-b472-7a99a94b1b7f","Type":"ContainerDied","Data":"c27a120fd7dbeb5fa1b56c51fd92477c83a48c2a7622836d74d76739ab64e42f"} Nov 21 14:57:15 crc kubenswrapper[4897]: I1121 14:57:15.139843 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-x8zrj" Nov 21 14:57:15 crc kubenswrapper[4897]: I1121 14:57:15.196336 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/4855262f-81f6-44ee-b472-7a99a94b1b7f-ovncontroller-config-0\") pod \"4855262f-81f6-44ee-b472-7a99a94b1b7f\" (UID: \"4855262f-81f6-44ee-b472-7a99a94b1b7f\") " Nov 21 14:57:15 crc kubenswrapper[4897]: I1121 14:57:15.196450 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4855262f-81f6-44ee-b472-7a99a94b1b7f-inventory\") pod \"4855262f-81f6-44ee-b472-7a99a94b1b7f\" (UID: \"4855262f-81f6-44ee-b472-7a99a94b1b7f\") " Nov 21 14:57:15 crc kubenswrapper[4897]: I1121 14:57:15.196736 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4855262f-81f6-44ee-b472-7a99a94b1b7f-ssh-key\") pod \"4855262f-81f6-44ee-b472-7a99a94b1b7f\" (UID: \"4855262f-81f6-44ee-b472-7a99a94b1b7f\") " Nov 21 14:57:15 crc kubenswrapper[4897]: I1121 14:57:15.196768 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4855262f-81f6-44ee-b472-7a99a94b1b7f-ovn-combined-ca-bundle\") pod \"4855262f-81f6-44ee-b472-7a99a94b1b7f\" (UID: \"4855262f-81f6-44ee-b472-7a99a94b1b7f\") " Nov 21 14:57:15 crc kubenswrapper[4897]: I1121 14:57:15.196982 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8gv9v\" (UniqueName: \"kubernetes.io/projected/4855262f-81f6-44ee-b472-7a99a94b1b7f-kube-api-access-8gv9v\") pod \"4855262f-81f6-44ee-b472-7a99a94b1b7f\" (UID: \"4855262f-81f6-44ee-b472-7a99a94b1b7f\") " Nov 21 14:57:15 crc kubenswrapper[4897]: I1121 14:57:15.206695 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4855262f-81f6-44ee-b472-7a99a94b1b7f-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "4855262f-81f6-44ee-b472-7a99a94b1b7f" (UID: "4855262f-81f6-44ee-b472-7a99a94b1b7f"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:57:15 crc kubenswrapper[4897]: I1121 14:57:15.206717 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4855262f-81f6-44ee-b472-7a99a94b1b7f-kube-api-access-8gv9v" (OuterVolumeSpecName: "kube-api-access-8gv9v") pod "4855262f-81f6-44ee-b472-7a99a94b1b7f" (UID: "4855262f-81f6-44ee-b472-7a99a94b1b7f"). InnerVolumeSpecName "kube-api-access-8gv9v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:57:15 crc kubenswrapper[4897]: I1121 14:57:15.238646 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4855262f-81f6-44ee-b472-7a99a94b1b7f-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "4855262f-81f6-44ee-b472-7a99a94b1b7f" (UID: "4855262f-81f6-44ee-b472-7a99a94b1b7f"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 14:57:15 crc kubenswrapper[4897]: I1121 14:57:15.241766 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4855262f-81f6-44ee-b472-7a99a94b1b7f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4855262f-81f6-44ee-b472-7a99a94b1b7f" (UID: "4855262f-81f6-44ee-b472-7a99a94b1b7f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:57:15 crc kubenswrapper[4897]: I1121 14:57:15.244682 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4855262f-81f6-44ee-b472-7a99a94b1b7f-inventory" (OuterVolumeSpecName: "inventory") pod "4855262f-81f6-44ee-b472-7a99a94b1b7f" (UID: "4855262f-81f6-44ee-b472-7a99a94b1b7f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:57:15 crc kubenswrapper[4897]: I1121 14:57:15.306083 4897 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/4855262f-81f6-44ee-b472-7a99a94b1b7f-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Nov 21 14:57:15 crc kubenswrapper[4897]: I1121 14:57:15.306145 4897 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4855262f-81f6-44ee-b472-7a99a94b1b7f-inventory\") on node \"crc\" DevicePath \"\"" Nov 21 14:57:15 crc kubenswrapper[4897]: I1121 14:57:15.306158 4897 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4855262f-81f6-44ee-b472-7a99a94b1b7f-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 21 14:57:15 crc kubenswrapper[4897]: I1121 14:57:15.306175 4897 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4855262f-81f6-44ee-b472-7a99a94b1b7f-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 21 14:57:15 crc kubenswrapper[4897]: I1121 14:57:15.306186 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8gv9v\" (UniqueName: \"kubernetes.io/projected/4855262f-81f6-44ee-b472-7a99a94b1b7f-kube-api-access-8gv9v\") on node \"crc\" DevicePath \"\"" Nov 21 14:57:15 crc kubenswrapper[4897]: I1121 14:57:15.620712 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-x8zrj" event={"ID":"4855262f-81f6-44ee-b472-7a99a94b1b7f","Type":"ContainerDied","Data":"5a5abac618128685ac3d58e7dd63ff2fd962faf662d1018699d86d1dc194c07c"} Nov 21 14:57:15 crc kubenswrapper[4897]: I1121 14:57:15.620786 4897 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5a5abac618128685ac3d58e7dd63ff2fd962faf662d1018699d86d1dc194c07c" Nov 21 14:57:15 crc kubenswrapper[4897]: I1121 14:57:15.620857 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-x8zrj" Nov 21 14:57:15 crc kubenswrapper[4897]: I1121 14:57:15.786205 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vlwnn"] Nov 21 14:57:15 crc kubenswrapper[4897]: E1121 14:57:15.786727 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4855262f-81f6-44ee-b472-7a99a94b1b7f" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Nov 21 14:57:15 crc kubenswrapper[4897]: I1121 14:57:15.786744 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="4855262f-81f6-44ee-b472-7a99a94b1b7f" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Nov 21 14:57:15 crc kubenswrapper[4897]: I1121 14:57:15.787035 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="4855262f-81f6-44ee-b472-7a99a94b1b7f" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Nov 21 14:57:15 crc kubenswrapper[4897]: I1121 14:57:15.787972 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vlwnn" Nov 21 14:57:15 crc kubenswrapper[4897]: I1121 14:57:15.791621 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Nov 21 14:57:15 crc kubenswrapper[4897]: I1121 14:57:15.791621 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 21 14:57:15 crc kubenswrapper[4897]: I1121 14:57:15.791630 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-dj88m" Nov 21 14:57:15 crc kubenswrapper[4897]: I1121 14:57:15.791630 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Nov 21 14:57:15 crc kubenswrapper[4897]: I1121 14:57:15.791691 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 21 14:57:15 crc kubenswrapper[4897]: I1121 14:57:15.791772 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 21 14:57:15 crc kubenswrapper[4897]: I1121 14:57:15.837991 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vlwnn"] Nov 21 14:57:15 crc kubenswrapper[4897]: I1121 14:57:15.927657 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mdqkx\" (UniqueName: \"kubernetes.io/projected/b8435f6a-a8be-40c6-8a3e-38f3b31d5f99-kube-api-access-mdqkx\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-vlwnn\" (UID: \"b8435f6a-a8be-40c6-8a3e-38f3b31d5f99\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vlwnn" Nov 21 14:57:15 crc kubenswrapper[4897]: I1121 14:57:15.927889 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8435f6a-a8be-40c6-8a3e-38f3b31d5f99-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-vlwnn\" (UID: \"b8435f6a-a8be-40c6-8a3e-38f3b31d5f99\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vlwnn" Nov 21 14:57:15 crc kubenswrapper[4897]: I1121 14:57:15.928411 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b8435f6a-a8be-40c6-8a3e-38f3b31d5f99-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-vlwnn\" (UID: \"b8435f6a-a8be-40c6-8a3e-38f3b31d5f99\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vlwnn" Nov 21 14:57:15 crc kubenswrapper[4897]: I1121 14:57:15.928585 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/b8435f6a-a8be-40c6-8a3e-38f3b31d5f99-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-vlwnn\" (UID: \"b8435f6a-a8be-40c6-8a3e-38f3b31d5f99\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vlwnn" Nov 21 14:57:15 crc kubenswrapper[4897]: I1121 14:57:15.928656 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/b8435f6a-a8be-40c6-8a3e-38f3b31d5f99-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-vlwnn\" (UID: \"b8435f6a-a8be-40c6-8a3e-38f3b31d5f99\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vlwnn" Nov 21 14:57:15 crc kubenswrapper[4897]: I1121 14:57:15.928783 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b8435f6a-a8be-40c6-8a3e-38f3b31d5f99-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-vlwnn\" (UID: \"b8435f6a-a8be-40c6-8a3e-38f3b31d5f99\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vlwnn" Nov 21 14:57:16 crc kubenswrapper[4897]: I1121 14:57:16.030709 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b8435f6a-a8be-40c6-8a3e-38f3b31d5f99-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-vlwnn\" (UID: \"b8435f6a-a8be-40c6-8a3e-38f3b31d5f99\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vlwnn" Nov 21 14:57:16 crc kubenswrapper[4897]: I1121 14:57:16.030777 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/b8435f6a-a8be-40c6-8a3e-38f3b31d5f99-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-vlwnn\" (UID: \"b8435f6a-a8be-40c6-8a3e-38f3b31d5f99\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vlwnn" Nov 21 14:57:16 crc kubenswrapper[4897]: I1121 14:57:16.030805 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/b8435f6a-a8be-40c6-8a3e-38f3b31d5f99-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-vlwnn\" (UID: \"b8435f6a-a8be-40c6-8a3e-38f3b31d5f99\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vlwnn" Nov 21 14:57:16 crc kubenswrapper[4897]: I1121 14:57:16.030863 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b8435f6a-a8be-40c6-8a3e-38f3b31d5f99-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-vlwnn\" (UID: \"b8435f6a-a8be-40c6-8a3e-38f3b31d5f99\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vlwnn" Nov 21 14:57:16 crc kubenswrapper[4897]: I1121 14:57:16.030942 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mdqkx\" (UniqueName: \"kubernetes.io/projected/b8435f6a-a8be-40c6-8a3e-38f3b31d5f99-kube-api-access-mdqkx\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-vlwnn\" (UID: \"b8435f6a-a8be-40c6-8a3e-38f3b31d5f99\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vlwnn" Nov 21 14:57:16 crc kubenswrapper[4897]: I1121 14:57:16.030973 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8435f6a-a8be-40c6-8a3e-38f3b31d5f99-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-vlwnn\" (UID: \"b8435f6a-a8be-40c6-8a3e-38f3b31d5f99\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vlwnn" Nov 21 14:57:16 crc kubenswrapper[4897]: I1121 14:57:16.035658 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/b8435f6a-a8be-40c6-8a3e-38f3b31d5f99-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-vlwnn\" (UID: \"b8435f6a-a8be-40c6-8a3e-38f3b31d5f99\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vlwnn" Nov 21 14:57:16 crc kubenswrapper[4897]: I1121 14:57:16.036590 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b8435f6a-a8be-40c6-8a3e-38f3b31d5f99-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-vlwnn\" (UID: \"b8435f6a-a8be-40c6-8a3e-38f3b31d5f99\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vlwnn" Nov 21 14:57:16 crc kubenswrapper[4897]: I1121 14:57:16.037208 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8435f6a-a8be-40c6-8a3e-38f3b31d5f99-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-vlwnn\" (UID: \"b8435f6a-a8be-40c6-8a3e-38f3b31d5f99\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vlwnn" Nov 21 14:57:16 crc kubenswrapper[4897]: I1121 14:57:16.038359 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b8435f6a-a8be-40c6-8a3e-38f3b31d5f99-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-vlwnn\" (UID: \"b8435f6a-a8be-40c6-8a3e-38f3b31d5f99\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vlwnn" Nov 21 14:57:16 crc kubenswrapper[4897]: I1121 14:57:16.041242 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/b8435f6a-a8be-40c6-8a3e-38f3b31d5f99-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-vlwnn\" (UID: \"b8435f6a-a8be-40c6-8a3e-38f3b31d5f99\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vlwnn" Nov 21 14:57:16 crc kubenswrapper[4897]: I1121 14:57:16.054311 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mdqkx\" (UniqueName: \"kubernetes.io/projected/b8435f6a-a8be-40c6-8a3e-38f3b31d5f99-kube-api-access-mdqkx\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-vlwnn\" (UID: \"b8435f6a-a8be-40c6-8a3e-38f3b31d5f99\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vlwnn" Nov 21 14:57:16 crc kubenswrapper[4897]: I1121 14:57:16.106214 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vlwnn" Nov 21 14:57:16 crc kubenswrapper[4897]: I1121 14:57:16.695866 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vlwnn"] Nov 21 14:57:17 crc kubenswrapper[4897]: I1121 14:57:17.646627 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vlwnn" event={"ID":"b8435f6a-a8be-40c6-8a3e-38f3b31d5f99","Type":"ContainerStarted","Data":"16ea82dba36608e015638471716778e445270337311c3b3e07b6f13cdc692dac"} Nov 21 14:57:20 crc kubenswrapper[4897]: I1121 14:57:20.678001 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vlwnn" event={"ID":"b8435f6a-a8be-40c6-8a3e-38f3b31d5f99","Type":"ContainerStarted","Data":"0e401b6cc8f13ee94e29ce4c28255b9efe58c210c56d61ccda2397f69f665df5"} Nov 21 14:57:20 crc kubenswrapper[4897]: I1121 14:57:20.698578 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vlwnn" podStartSLOduration=2.460476335 podStartE2EDuration="5.698487684s" podCreationTimestamp="2025-11-21 14:57:15 +0000 UTC" firstStartedPulling="2025-11-21 14:57:16.711072393 +0000 UTC m=+2913.995665868" lastFinishedPulling="2025-11-21 14:57:19.949083752 +0000 UTC m=+2917.233677217" observedRunningTime="2025-11-21 14:57:20.697116197 +0000 UTC m=+2917.981709682" watchObservedRunningTime="2025-11-21 14:57:20.698487684 +0000 UTC m=+2917.983081159" Nov 21 14:57:24 crc kubenswrapper[4897]: I1121 14:57:24.728993 4897 generic.go:334] "Generic (PLEG): container finished" podID="b636617a-cfba-4203-b0ac-05df18bcc511" containerID="52f6b003837aac81a78f9bdf86267891d3d1512dceac9832ca27b728b07aa50e" exitCode=0 Nov 21 14:57:24 crc kubenswrapper[4897]: I1121 14:57:24.729114 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8g9w2" event={"ID":"b636617a-cfba-4203-b0ac-05df18bcc511","Type":"ContainerDied","Data":"52f6b003837aac81a78f9bdf86267891d3d1512dceac9832ca27b728b07aa50e"} Nov 21 14:57:26 crc kubenswrapper[4897]: I1121 14:57:26.754225 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8g9w2" event={"ID":"b636617a-cfba-4203-b0ac-05df18bcc511","Type":"ContainerStarted","Data":"829835eac8314d88b7a57bf9070a6fc38ebfd4b3c15fe68e50dcf86bc5cd544d"} Nov 21 14:57:26 crc kubenswrapper[4897]: I1121 14:57:26.782815 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-8g9w2" podStartSLOduration=4.656349836 podStartE2EDuration="21.782791674s" podCreationTimestamp="2025-11-21 14:57:05 +0000 UTC" firstStartedPulling="2025-11-21 14:57:08.543252036 +0000 UTC m=+2905.827845511" lastFinishedPulling="2025-11-21 14:57:25.669693874 +0000 UTC m=+2922.954287349" observedRunningTime="2025-11-21 14:57:26.77263548 +0000 UTC m=+2924.057228965" watchObservedRunningTime="2025-11-21 14:57:26.782791674 +0000 UTC m=+2924.067385149" Nov 21 14:57:34 crc kubenswrapper[4897]: I1121 14:57:34.370633 4897 patch_prober.go:28] interesting pod/machine-config-daemon-krv5b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 21 14:57:34 crc kubenswrapper[4897]: I1121 14:57:34.372083 4897 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 21 14:57:36 crc kubenswrapper[4897]: I1121 14:57:36.035590 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-8g9w2" Nov 21 14:57:36 crc kubenswrapper[4897]: I1121 14:57:36.036018 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-8g9w2" Nov 21 14:57:37 crc kubenswrapper[4897]: I1121 14:57:37.095363 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-8g9w2" podUID="b636617a-cfba-4203-b0ac-05df18bcc511" containerName="registry-server" probeResult="failure" output=< Nov 21 14:57:37 crc kubenswrapper[4897]: timeout: failed to connect service ":50051" within 1s Nov 21 14:57:37 crc kubenswrapper[4897]: > Nov 21 14:57:47 crc kubenswrapper[4897]: I1121 14:57:47.096567 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-8g9w2" podUID="b636617a-cfba-4203-b0ac-05df18bcc511" containerName="registry-server" probeResult="failure" output=< Nov 21 14:57:47 crc kubenswrapper[4897]: timeout: failed to connect service ":50051" within 1s Nov 21 14:57:47 crc kubenswrapper[4897]: > Nov 21 14:57:57 crc kubenswrapper[4897]: I1121 14:57:57.091807 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-8g9w2" podUID="b636617a-cfba-4203-b0ac-05df18bcc511" containerName="registry-server" probeResult="failure" output=< Nov 21 14:57:57 crc kubenswrapper[4897]: timeout: failed to connect service ":50051" within 1s Nov 21 14:57:57 crc kubenswrapper[4897]: > Nov 21 14:58:04 crc kubenswrapper[4897]: I1121 14:58:04.370830 4897 patch_prober.go:28] interesting pod/machine-config-daemon-krv5b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 21 14:58:04 crc kubenswrapper[4897]: I1121 14:58:04.371389 4897 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 21 14:58:04 crc kubenswrapper[4897]: I1121 14:58:04.371439 4897 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" Nov 21 14:58:04 crc kubenswrapper[4897]: I1121 14:58:04.372366 4897 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a17b834dbb567d490c0f7ce8a22ab5aa6ecb645fc460c063af4ca43f58348563"} pod="openshift-machine-config-operator/machine-config-daemon-krv5b" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 21 14:58:04 crc kubenswrapper[4897]: I1121 14:58:04.372415 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" containerName="machine-config-daemon" containerID="cri-o://a17b834dbb567d490c0f7ce8a22ab5aa6ecb645fc460c063af4ca43f58348563" gracePeriod=600 Nov 21 14:58:04 crc kubenswrapper[4897]: E1121 14:58:04.520871 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 14:58:05 crc kubenswrapper[4897]: I1121 14:58:05.177905 4897 generic.go:334] "Generic (PLEG): container finished" podID="e7670227-d280-4847-b882-754429f56b0f" containerID="a17b834dbb567d490c0f7ce8a22ab5aa6ecb645fc460c063af4ca43f58348563" exitCode=0 Nov 21 14:58:05 crc kubenswrapper[4897]: I1121 14:58:05.177943 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" event={"ID":"e7670227-d280-4847-b882-754429f56b0f","Type":"ContainerDied","Data":"a17b834dbb567d490c0f7ce8a22ab5aa6ecb645fc460c063af4ca43f58348563"} Nov 21 14:58:05 crc kubenswrapper[4897]: I1121 14:58:05.178015 4897 scope.go:117] "RemoveContainer" containerID="93211f44a4c3f56e3048bcba270e9ae8a65d8435c6fcfa8de9ac51f949ca9ee4" Nov 21 14:58:05 crc kubenswrapper[4897]: I1121 14:58:05.178869 4897 scope.go:117] "RemoveContainer" containerID="a17b834dbb567d490c0f7ce8a22ab5aa6ecb645fc460c063af4ca43f58348563" Nov 21 14:58:05 crc kubenswrapper[4897]: E1121 14:58:05.179252 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 14:58:07 crc kubenswrapper[4897]: I1121 14:58:07.085753 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-8g9w2" podUID="b636617a-cfba-4203-b0ac-05df18bcc511" containerName="registry-server" probeResult="failure" output=< Nov 21 14:58:07 crc kubenswrapper[4897]: timeout: failed to connect service ":50051" within 1s Nov 21 14:58:07 crc kubenswrapper[4897]: > Nov 21 14:58:13 crc kubenswrapper[4897]: I1121 14:58:13.276063 4897 generic.go:334] "Generic (PLEG): container finished" podID="b8435f6a-a8be-40c6-8a3e-38f3b31d5f99" containerID="0e401b6cc8f13ee94e29ce4c28255b9efe58c210c56d61ccda2397f69f665df5" exitCode=0 Nov 21 14:58:13 crc kubenswrapper[4897]: I1121 14:58:13.276170 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vlwnn" event={"ID":"b8435f6a-a8be-40c6-8a3e-38f3b31d5f99","Type":"ContainerDied","Data":"0e401b6cc8f13ee94e29ce4c28255b9efe58c210c56d61ccda2397f69f665df5"} Nov 21 14:58:14 crc kubenswrapper[4897]: I1121 14:58:14.753719 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vlwnn" Nov 21 14:58:14 crc kubenswrapper[4897]: I1121 14:58:14.827038 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mdqkx\" (UniqueName: \"kubernetes.io/projected/b8435f6a-a8be-40c6-8a3e-38f3b31d5f99-kube-api-access-mdqkx\") pod \"b8435f6a-a8be-40c6-8a3e-38f3b31d5f99\" (UID: \"b8435f6a-a8be-40c6-8a3e-38f3b31d5f99\") " Nov 21 14:58:14 crc kubenswrapper[4897]: I1121 14:58:14.827201 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/b8435f6a-a8be-40c6-8a3e-38f3b31d5f99-neutron-ovn-metadata-agent-neutron-config-0\") pod \"b8435f6a-a8be-40c6-8a3e-38f3b31d5f99\" (UID: \"b8435f6a-a8be-40c6-8a3e-38f3b31d5f99\") " Nov 21 14:58:14 crc kubenswrapper[4897]: I1121 14:58:14.827295 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b8435f6a-a8be-40c6-8a3e-38f3b31d5f99-inventory\") pod \"b8435f6a-a8be-40c6-8a3e-38f3b31d5f99\" (UID: \"b8435f6a-a8be-40c6-8a3e-38f3b31d5f99\") " Nov 21 14:58:14 crc kubenswrapper[4897]: I1121 14:58:14.827456 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8435f6a-a8be-40c6-8a3e-38f3b31d5f99-neutron-metadata-combined-ca-bundle\") pod \"b8435f6a-a8be-40c6-8a3e-38f3b31d5f99\" (UID: \"b8435f6a-a8be-40c6-8a3e-38f3b31d5f99\") " Nov 21 14:58:14 crc kubenswrapper[4897]: I1121 14:58:14.827564 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b8435f6a-a8be-40c6-8a3e-38f3b31d5f99-ssh-key\") pod \"b8435f6a-a8be-40c6-8a3e-38f3b31d5f99\" (UID: \"b8435f6a-a8be-40c6-8a3e-38f3b31d5f99\") " Nov 21 14:58:14 crc kubenswrapper[4897]: I1121 14:58:14.827625 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/b8435f6a-a8be-40c6-8a3e-38f3b31d5f99-nova-metadata-neutron-config-0\") pod \"b8435f6a-a8be-40c6-8a3e-38f3b31d5f99\" (UID: \"b8435f6a-a8be-40c6-8a3e-38f3b31d5f99\") " Nov 21 14:58:14 crc kubenswrapper[4897]: I1121 14:58:14.833872 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8435f6a-a8be-40c6-8a3e-38f3b31d5f99-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "b8435f6a-a8be-40c6-8a3e-38f3b31d5f99" (UID: "b8435f6a-a8be-40c6-8a3e-38f3b31d5f99"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:58:14 crc kubenswrapper[4897]: I1121 14:58:14.846356 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b8435f6a-a8be-40c6-8a3e-38f3b31d5f99-kube-api-access-mdqkx" (OuterVolumeSpecName: "kube-api-access-mdqkx") pod "b8435f6a-a8be-40c6-8a3e-38f3b31d5f99" (UID: "b8435f6a-a8be-40c6-8a3e-38f3b31d5f99"). InnerVolumeSpecName "kube-api-access-mdqkx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:58:14 crc kubenswrapper[4897]: I1121 14:58:14.865793 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8435f6a-a8be-40c6-8a3e-38f3b31d5f99-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b8435f6a-a8be-40c6-8a3e-38f3b31d5f99" (UID: "b8435f6a-a8be-40c6-8a3e-38f3b31d5f99"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:58:14 crc kubenswrapper[4897]: I1121 14:58:14.867009 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8435f6a-a8be-40c6-8a3e-38f3b31d5f99-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "b8435f6a-a8be-40c6-8a3e-38f3b31d5f99" (UID: "b8435f6a-a8be-40c6-8a3e-38f3b31d5f99"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:58:14 crc kubenswrapper[4897]: I1121 14:58:14.867552 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8435f6a-a8be-40c6-8a3e-38f3b31d5f99-inventory" (OuterVolumeSpecName: "inventory") pod "b8435f6a-a8be-40c6-8a3e-38f3b31d5f99" (UID: "b8435f6a-a8be-40c6-8a3e-38f3b31d5f99"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:58:14 crc kubenswrapper[4897]: I1121 14:58:14.869800 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8435f6a-a8be-40c6-8a3e-38f3b31d5f99-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "b8435f6a-a8be-40c6-8a3e-38f3b31d5f99" (UID: "b8435f6a-a8be-40c6-8a3e-38f3b31d5f99"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 14:58:14 crc kubenswrapper[4897]: I1121 14:58:14.931026 4897 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b8435f6a-a8be-40c6-8a3e-38f3b31d5f99-inventory\") on node \"crc\" DevicePath \"\"" Nov 21 14:58:14 crc kubenswrapper[4897]: I1121 14:58:14.931073 4897 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8435f6a-a8be-40c6-8a3e-38f3b31d5f99-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 21 14:58:14 crc kubenswrapper[4897]: I1121 14:58:14.931085 4897 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b8435f6a-a8be-40c6-8a3e-38f3b31d5f99-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 21 14:58:14 crc kubenswrapper[4897]: I1121 14:58:14.931094 4897 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/b8435f6a-a8be-40c6-8a3e-38f3b31d5f99-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Nov 21 14:58:14 crc kubenswrapper[4897]: I1121 14:58:14.931106 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mdqkx\" (UniqueName: \"kubernetes.io/projected/b8435f6a-a8be-40c6-8a3e-38f3b31d5f99-kube-api-access-mdqkx\") on node \"crc\" DevicePath \"\"" Nov 21 14:58:14 crc kubenswrapper[4897]: I1121 14:58:14.931116 4897 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/b8435f6a-a8be-40c6-8a3e-38f3b31d5f99-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Nov 21 14:58:15 crc kubenswrapper[4897]: I1121 14:58:15.298430 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vlwnn" event={"ID":"b8435f6a-a8be-40c6-8a3e-38f3b31d5f99","Type":"ContainerDied","Data":"16ea82dba36608e015638471716778e445270337311c3b3e07b6f13cdc692dac"} Nov 21 14:58:15 crc kubenswrapper[4897]: I1121 14:58:15.298480 4897 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="16ea82dba36608e015638471716778e445270337311c3b3e07b6f13cdc692dac" Nov 21 14:58:15 crc kubenswrapper[4897]: I1121 14:58:15.298580 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-vlwnn" Nov 21 14:58:15 crc kubenswrapper[4897]: I1121 14:58:15.534045 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tmr6q"] Nov 21 14:58:15 crc kubenswrapper[4897]: E1121 14:58:15.534584 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8435f6a-a8be-40c6-8a3e-38f3b31d5f99" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Nov 21 14:58:15 crc kubenswrapper[4897]: I1121 14:58:15.534600 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8435f6a-a8be-40c6-8a3e-38f3b31d5f99" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Nov 21 14:58:15 crc kubenswrapper[4897]: I1121 14:58:15.534859 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8435f6a-a8be-40c6-8a3e-38f3b31d5f99" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Nov 21 14:58:15 crc kubenswrapper[4897]: I1121 14:58:15.535720 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tmr6q" Nov 21 14:58:15 crc kubenswrapper[4897]: I1121 14:58:15.538955 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-dj88m" Nov 21 14:58:15 crc kubenswrapper[4897]: I1121 14:58:15.539169 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Nov 21 14:58:15 crc kubenswrapper[4897]: I1121 14:58:15.539337 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 21 14:58:15 crc kubenswrapper[4897]: I1121 14:58:15.539569 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 21 14:58:15 crc kubenswrapper[4897]: I1121 14:58:15.539724 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 21 14:58:15 crc kubenswrapper[4897]: I1121 14:58:15.550031 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tmr6q"] Nov 21 14:58:15 crc kubenswrapper[4897]: I1121 14:58:15.649856 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d2ft2\" (UniqueName: \"kubernetes.io/projected/9f8e267b-7f34-4793-bb66-a2066bf63ce5-kube-api-access-d2ft2\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-tmr6q\" (UID: \"9f8e267b-7f34-4793-bb66-a2066bf63ce5\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tmr6q" Nov 21 14:58:15 crc kubenswrapper[4897]: I1121 14:58:15.654687 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f8e267b-7f34-4793-bb66-a2066bf63ce5-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-tmr6q\" (UID: \"9f8e267b-7f34-4793-bb66-a2066bf63ce5\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tmr6q" Nov 21 14:58:15 crc kubenswrapper[4897]: I1121 14:58:15.654822 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9f8e267b-7f34-4793-bb66-a2066bf63ce5-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-tmr6q\" (UID: \"9f8e267b-7f34-4793-bb66-a2066bf63ce5\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tmr6q" Nov 21 14:58:15 crc kubenswrapper[4897]: I1121 14:58:15.655308 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9f8e267b-7f34-4793-bb66-a2066bf63ce5-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-tmr6q\" (UID: \"9f8e267b-7f34-4793-bb66-a2066bf63ce5\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tmr6q" Nov 21 14:58:15 crc kubenswrapper[4897]: I1121 14:58:15.655372 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/9f8e267b-7f34-4793-bb66-a2066bf63ce5-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-tmr6q\" (UID: \"9f8e267b-7f34-4793-bb66-a2066bf63ce5\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tmr6q" Nov 21 14:58:15 crc kubenswrapper[4897]: I1121 14:58:15.757600 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9f8e267b-7f34-4793-bb66-a2066bf63ce5-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-tmr6q\" (UID: \"9f8e267b-7f34-4793-bb66-a2066bf63ce5\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tmr6q" Nov 21 14:58:15 crc kubenswrapper[4897]: I1121 14:58:15.757679 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/9f8e267b-7f34-4793-bb66-a2066bf63ce5-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-tmr6q\" (UID: \"9f8e267b-7f34-4793-bb66-a2066bf63ce5\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tmr6q" Nov 21 14:58:15 crc kubenswrapper[4897]: I1121 14:58:15.757817 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d2ft2\" (UniqueName: \"kubernetes.io/projected/9f8e267b-7f34-4793-bb66-a2066bf63ce5-kube-api-access-d2ft2\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-tmr6q\" (UID: \"9f8e267b-7f34-4793-bb66-a2066bf63ce5\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tmr6q" Nov 21 14:58:15 crc kubenswrapper[4897]: I1121 14:58:15.757903 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f8e267b-7f34-4793-bb66-a2066bf63ce5-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-tmr6q\" (UID: \"9f8e267b-7f34-4793-bb66-a2066bf63ce5\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tmr6q" Nov 21 14:58:15 crc kubenswrapper[4897]: I1121 14:58:15.757954 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9f8e267b-7f34-4793-bb66-a2066bf63ce5-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-tmr6q\" (UID: \"9f8e267b-7f34-4793-bb66-a2066bf63ce5\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tmr6q" Nov 21 14:58:15 crc kubenswrapper[4897]: I1121 14:58:15.762258 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f8e267b-7f34-4793-bb66-a2066bf63ce5-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-tmr6q\" (UID: \"9f8e267b-7f34-4793-bb66-a2066bf63ce5\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tmr6q" Nov 21 14:58:15 crc kubenswrapper[4897]: I1121 14:58:15.762543 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9f8e267b-7f34-4793-bb66-a2066bf63ce5-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-tmr6q\" (UID: \"9f8e267b-7f34-4793-bb66-a2066bf63ce5\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tmr6q" Nov 21 14:58:15 crc kubenswrapper[4897]: I1121 14:58:15.764041 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9f8e267b-7f34-4793-bb66-a2066bf63ce5-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-tmr6q\" (UID: \"9f8e267b-7f34-4793-bb66-a2066bf63ce5\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tmr6q" Nov 21 14:58:15 crc kubenswrapper[4897]: I1121 14:58:15.771033 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/9f8e267b-7f34-4793-bb66-a2066bf63ce5-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-tmr6q\" (UID: \"9f8e267b-7f34-4793-bb66-a2066bf63ce5\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tmr6q" Nov 21 14:58:15 crc kubenswrapper[4897]: I1121 14:58:15.774368 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d2ft2\" (UniqueName: \"kubernetes.io/projected/9f8e267b-7f34-4793-bb66-a2066bf63ce5-kube-api-access-d2ft2\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-tmr6q\" (UID: \"9f8e267b-7f34-4793-bb66-a2066bf63ce5\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tmr6q" Nov 21 14:58:15 crc kubenswrapper[4897]: I1121 14:58:15.866539 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tmr6q" Nov 21 14:58:16 crc kubenswrapper[4897]: I1121 14:58:16.089379 4897 scope.go:117] "RemoveContainer" containerID="a17b834dbb567d490c0f7ce8a22ab5aa6ecb645fc460c063af4ca43f58348563" Nov 21 14:58:16 crc kubenswrapper[4897]: E1121 14:58:16.090017 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 14:58:16 crc kubenswrapper[4897]: I1121 14:58:16.101405 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-8g9w2" Nov 21 14:58:16 crc kubenswrapper[4897]: I1121 14:58:16.157061 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-8g9w2" Nov 21 14:58:16 crc kubenswrapper[4897]: I1121 14:58:16.343707 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8g9w2"] Nov 21 14:58:16 crc kubenswrapper[4897]: I1121 14:58:16.412594 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tmr6q"] Nov 21 14:58:17 crc kubenswrapper[4897]: I1121 14:58:17.320716 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tmr6q" event={"ID":"9f8e267b-7f34-4793-bb66-a2066bf63ce5","Type":"ContainerStarted","Data":"9100a89831ed8428be3f3d349c8730bf17aa5d8f02956e117651a49c120c78df"} Nov 21 14:58:17 crc kubenswrapper[4897]: I1121 14:58:17.320867 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-8g9w2" podUID="b636617a-cfba-4203-b0ac-05df18bcc511" containerName="registry-server" containerID="cri-o://829835eac8314d88b7a57bf9070a6fc38ebfd4b3c15fe68e50dcf86bc5cd544d" gracePeriod=2 Nov 21 14:58:18 crc kubenswrapper[4897]: I1121 14:58:18.340096 4897 generic.go:334] "Generic (PLEG): container finished" podID="b636617a-cfba-4203-b0ac-05df18bcc511" containerID="829835eac8314d88b7a57bf9070a6fc38ebfd4b3c15fe68e50dcf86bc5cd544d" exitCode=0 Nov 21 14:58:18 crc kubenswrapper[4897]: I1121 14:58:18.340171 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8g9w2" event={"ID":"b636617a-cfba-4203-b0ac-05df18bcc511","Type":"ContainerDied","Data":"829835eac8314d88b7a57bf9070a6fc38ebfd4b3c15fe68e50dcf86bc5cd544d"} Nov 21 14:58:20 crc kubenswrapper[4897]: I1121 14:58:20.590787 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8g9w2" Nov 21 14:58:20 crc kubenswrapper[4897]: I1121 14:58:20.710192 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b636617a-cfba-4203-b0ac-05df18bcc511-catalog-content\") pod \"b636617a-cfba-4203-b0ac-05df18bcc511\" (UID: \"b636617a-cfba-4203-b0ac-05df18bcc511\") " Nov 21 14:58:20 crc kubenswrapper[4897]: I1121 14:58:20.710335 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b636617a-cfba-4203-b0ac-05df18bcc511-utilities\") pod \"b636617a-cfba-4203-b0ac-05df18bcc511\" (UID: \"b636617a-cfba-4203-b0ac-05df18bcc511\") " Nov 21 14:58:20 crc kubenswrapper[4897]: I1121 14:58:20.710442 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-58259\" (UniqueName: \"kubernetes.io/projected/b636617a-cfba-4203-b0ac-05df18bcc511-kube-api-access-58259\") pod \"b636617a-cfba-4203-b0ac-05df18bcc511\" (UID: \"b636617a-cfba-4203-b0ac-05df18bcc511\") " Nov 21 14:58:20 crc kubenswrapper[4897]: I1121 14:58:20.712523 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b636617a-cfba-4203-b0ac-05df18bcc511-utilities" (OuterVolumeSpecName: "utilities") pod "b636617a-cfba-4203-b0ac-05df18bcc511" (UID: "b636617a-cfba-4203-b0ac-05df18bcc511"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:58:20 crc kubenswrapper[4897]: I1121 14:58:20.717459 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b636617a-cfba-4203-b0ac-05df18bcc511-kube-api-access-58259" (OuterVolumeSpecName: "kube-api-access-58259") pod "b636617a-cfba-4203-b0ac-05df18bcc511" (UID: "b636617a-cfba-4203-b0ac-05df18bcc511"). InnerVolumeSpecName "kube-api-access-58259". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 14:58:20 crc kubenswrapper[4897]: I1121 14:58:20.795603 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b636617a-cfba-4203-b0ac-05df18bcc511-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b636617a-cfba-4203-b0ac-05df18bcc511" (UID: "b636617a-cfba-4203-b0ac-05df18bcc511"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 14:58:20 crc kubenswrapper[4897]: I1121 14:58:20.813253 4897 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b636617a-cfba-4203-b0ac-05df18bcc511-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 21 14:58:20 crc kubenswrapper[4897]: I1121 14:58:20.813288 4897 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b636617a-cfba-4203-b0ac-05df18bcc511-utilities\") on node \"crc\" DevicePath \"\"" Nov 21 14:58:20 crc kubenswrapper[4897]: I1121 14:58:20.813297 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-58259\" (UniqueName: \"kubernetes.io/projected/b636617a-cfba-4203-b0ac-05df18bcc511-kube-api-access-58259\") on node \"crc\" DevicePath \"\"" Nov 21 14:58:21 crc kubenswrapper[4897]: I1121 14:58:21.377058 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8g9w2" event={"ID":"b636617a-cfba-4203-b0ac-05df18bcc511","Type":"ContainerDied","Data":"f20070f7c0c5ed9e50447185796ff29974429de1ff90746be9861b88126a9e02"} Nov 21 14:58:21 crc kubenswrapper[4897]: I1121 14:58:21.377093 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8g9w2" Nov 21 14:58:21 crc kubenswrapper[4897]: I1121 14:58:21.377114 4897 scope.go:117] "RemoveContainer" containerID="829835eac8314d88b7a57bf9070a6fc38ebfd4b3c15fe68e50dcf86bc5cd544d" Nov 21 14:58:21 crc kubenswrapper[4897]: I1121 14:58:21.379620 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tmr6q" event={"ID":"9f8e267b-7f34-4793-bb66-a2066bf63ce5","Type":"ContainerStarted","Data":"3745add26e7e03d442c861649da7caad1fcf48a01734c3ff971dc877553b76e0"} Nov 21 14:58:21 crc kubenswrapper[4897]: I1121 14:58:21.412389 4897 scope.go:117] "RemoveContainer" containerID="52f6b003837aac81a78f9bdf86267891d3d1512dceac9832ca27b728b07aa50e" Nov 21 14:58:21 crc kubenswrapper[4897]: I1121 14:58:21.421086 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tmr6q" podStartSLOduration=2.442410746 podStartE2EDuration="6.421065148s" podCreationTimestamp="2025-11-21 14:58:15 +0000 UTC" firstStartedPulling="2025-11-21 14:58:16.423111015 +0000 UTC m=+2973.707704490" lastFinishedPulling="2025-11-21 14:58:20.401765417 +0000 UTC m=+2977.686358892" observedRunningTime="2025-11-21 14:58:21.415762374 +0000 UTC m=+2978.700355859" watchObservedRunningTime="2025-11-21 14:58:21.421065148 +0000 UTC m=+2978.705658613" Nov 21 14:58:21 crc kubenswrapper[4897]: I1121 14:58:21.454633 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8g9w2"] Nov 21 14:58:21 crc kubenswrapper[4897]: I1121 14:58:21.478230 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-8g9w2"] Nov 21 14:58:21 crc kubenswrapper[4897]: I1121 14:58:21.483418 4897 scope.go:117] "RemoveContainer" containerID="297e196a896f626226136978df6c4a65592611e29cb7faf343c1b7c66f4b5d15" Nov 21 14:58:22 crc kubenswrapper[4897]: I1121 14:58:22.102316 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b636617a-cfba-4203-b0ac-05df18bcc511" path="/var/lib/kubelet/pods/b636617a-cfba-4203-b0ac-05df18bcc511/volumes" Nov 21 14:58:31 crc kubenswrapper[4897]: I1121 14:58:31.089445 4897 scope.go:117] "RemoveContainer" containerID="a17b834dbb567d490c0f7ce8a22ab5aa6ecb645fc460c063af4ca43f58348563" Nov 21 14:58:31 crc kubenswrapper[4897]: E1121 14:58:31.090302 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 14:58:44 crc kubenswrapper[4897]: I1121 14:58:44.100546 4897 scope.go:117] "RemoveContainer" containerID="a17b834dbb567d490c0f7ce8a22ab5aa6ecb645fc460c063af4ca43f58348563" Nov 21 14:58:44 crc kubenswrapper[4897]: E1121 14:58:44.102993 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 14:58:59 crc kubenswrapper[4897]: I1121 14:58:59.089834 4897 scope.go:117] "RemoveContainer" containerID="a17b834dbb567d490c0f7ce8a22ab5aa6ecb645fc460c063af4ca43f58348563" Nov 21 14:58:59 crc kubenswrapper[4897]: E1121 14:58:59.090603 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 14:59:12 crc kubenswrapper[4897]: I1121 14:59:12.089852 4897 scope.go:117] "RemoveContainer" containerID="a17b834dbb567d490c0f7ce8a22ab5aa6ecb645fc460c063af4ca43f58348563" Nov 21 14:59:12 crc kubenswrapper[4897]: E1121 14:59:12.091146 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 14:59:27 crc kubenswrapper[4897]: I1121 14:59:27.089480 4897 scope.go:117] "RemoveContainer" containerID="a17b834dbb567d490c0f7ce8a22ab5aa6ecb645fc460c063af4ca43f58348563" Nov 21 14:59:27 crc kubenswrapper[4897]: E1121 14:59:27.092634 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 14:59:42 crc kubenswrapper[4897]: I1121 14:59:42.089302 4897 scope.go:117] "RemoveContainer" containerID="a17b834dbb567d490c0f7ce8a22ab5aa6ecb645fc460c063af4ca43f58348563" Nov 21 14:59:42 crc kubenswrapper[4897]: E1121 14:59:42.090917 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 14:59:57 crc kubenswrapper[4897]: I1121 14:59:57.090785 4897 scope.go:117] "RemoveContainer" containerID="a17b834dbb567d490c0f7ce8a22ab5aa6ecb645fc460c063af4ca43f58348563" Nov 21 14:59:57 crc kubenswrapper[4897]: E1121 14:59:57.092726 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:00:00 crc kubenswrapper[4897]: I1121 15:00:00.159697 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29395620-84swg"] Nov 21 15:00:00 crc kubenswrapper[4897]: E1121 15:00:00.160464 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b636617a-cfba-4203-b0ac-05df18bcc511" containerName="extract-utilities" Nov 21 15:00:00 crc kubenswrapper[4897]: I1121 15:00:00.160477 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="b636617a-cfba-4203-b0ac-05df18bcc511" containerName="extract-utilities" Nov 21 15:00:00 crc kubenswrapper[4897]: E1121 15:00:00.160539 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b636617a-cfba-4203-b0ac-05df18bcc511" containerName="extract-content" Nov 21 15:00:00 crc kubenswrapper[4897]: I1121 15:00:00.160550 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="b636617a-cfba-4203-b0ac-05df18bcc511" containerName="extract-content" Nov 21 15:00:00 crc kubenswrapper[4897]: E1121 15:00:00.160575 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b636617a-cfba-4203-b0ac-05df18bcc511" containerName="registry-server" Nov 21 15:00:00 crc kubenswrapper[4897]: I1121 15:00:00.160584 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="b636617a-cfba-4203-b0ac-05df18bcc511" containerName="registry-server" Nov 21 15:00:00 crc kubenswrapper[4897]: I1121 15:00:00.160880 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="b636617a-cfba-4203-b0ac-05df18bcc511" containerName="registry-server" Nov 21 15:00:00 crc kubenswrapper[4897]: I1121 15:00:00.161702 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29395620-84swg" Nov 21 15:00:00 crc kubenswrapper[4897]: I1121 15:00:00.169267 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 21 15:00:00 crc kubenswrapper[4897]: I1121 15:00:00.170449 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 21 15:00:00 crc kubenswrapper[4897]: I1121 15:00:00.175626 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29395620-84swg"] Nov 21 15:00:00 crc kubenswrapper[4897]: I1121 15:00:00.297015 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c2518e25-e7d2-4bbe-b2fb-1aa2f933069b-config-volume\") pod \"collect-profiles-29395620-84swg\" (UID: \"c2518e25-e7d2-4bbe-b2fb-1aa2f933069b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29395620-84swg" Nov 21 15:00:00 crc kubenswrapper[4897]: I1121 15:00:00.297396 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c2518e25-e7d2-4bbe-b2fb-1aa2f933069b-secret-volume\") pod \"collect-profiles-29395620-84swg\" (UID: \"c2518e25-e7d2-4bbe-b2fb-1aa2f933069b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29395620-84swg" Nov 21 15:00:00 crc kubenswrapper[4897]: I1121 15:00:00.297665 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xchln\" (UniqueName: \"kubernetes.io/projected/c2518e25-e7d2-4bbe-b2fb-1aa2f933069b-kube-api-access-xchln\") pod \"collect-profiles-29395620-84swg\" (UID: \"c2518e25-e7d2-4bbe-b2fb-1aa2f933069b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29395620-84swg" Nov 21 15:00:00 crc kubenswrapper[4897]: I1121 15:00:00.399813 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c2518e25-e7d2-4bbe-b2fb-1aa2f933069b-config-volume\") pod \"collect-profiles-29395620-84swg\" (UID: \"c2518e25-e7d2-4bbe-b2fb-1aa2f933069b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29395620-84swg" Nov 21 15:00:00 crc kubenswrapper[4897]: I1121 15:00:00.399874 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c2518e25-e7d2-4bbe-b2fb-1aa2f933069b-secret-volume\") pod \"collect-profiles-29395620-84swg\" (UID: \"c2518e25-e7d2-4bbe-b2fb-1aa2f933069b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29395620-84swg" Nov 21 15:00:00 crc kubenswrapper[4897]: I1121 15:00:00.400023 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xchln\" (UniqueName: \"kubernetes.io/projected/c2518e25-e7d2-4bbe-b2fb-1aa2f933069b-kube-api-access-xchln\") pod \"collect-profiles-29395620-84swg\" (UID: \"c2518e25-e7d2-4bbe-b2fb-1aa2f933069b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29395620-84swg" Nov 21 15:00:00 crc kubenswrapper[4897]: I1121 15:00:00.400787 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c2518e25-e7d2-4bbe-b2fb-1aa2f933069b-config-volume\") pod \"collect-profiles-29395620-84swg\" (UID: \"c2518e25-e7d2-4bbe-b2fb-1aa2f933069b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29395620-84swg" Nov 21 15:00:00 crc kubenswrapper[4897]: I1121 15:00:00.408468 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c2518e25-e7d2-4bbe-b2fb-1aa2f933069b-secret-volume\") pod \"collect-profiles-29395620-84swg\" (UID: \"c2518e25-e7d2-4bbe-b2fb-1aa2f933069b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29395620-84swg" Nov 21 15:00:00 crc kubenswrapper[4897]: I1121 15:00:00.422327 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xchln\" (UniqueName: \"kubernetes.io/projected/c2518e25-e7d2-4bbe-b2fb-1aa2f933069b-kube-api-access-xchln\") pod \"collect-profiles-29395620-84swg\" (UID: \"c2518e25-e7d2-4bbe-b2fb-1aa2f933069b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29395620-84swg" Nov 21 15:00:00 crc kubenswrapper[4897]: I1121 15:00:00.484244 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29395620-84swg" Nov 21 15:00:00 crc kubenswrapper[4897]: I1121 15:00:00.977533 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29395620-84swg"] Nov 21 15:00:01 crc kubenswrapper[4897]: I1121 15:00:01.459725 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29395620-84swg" event={"ID":"c2518e25-e7d2-4bbe-b2fb-1aa2f933069b","Type":"ContainerStarted","Data":"fdaaddea473e11bf7a52648c4831c17c899e611002c030ddf6c9f4ad8d48c283"} Nov 21 15:00:01 crc kubenswrapper[4897]: I1121 15:00:01.460146 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29395620-84swg" event={"ID":"c2518e25-e7d2-4bbe-b2fb-1aa2f933069b","Type":"ContainerStarted","Data":"b714417e3c09f0879f27569e82ff1b79e12e75c40ae938c517bed93db6148dbd"} Nov 21 15:00:01 crc kubenswrapper[4897]: I1121 15:00:01.492768 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29395620-84swg" podStartSLOduration=1.4927451409999999 podStartE2EDuration="1.492745141s" podCreationTimestamp="2025-11-21 15:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 15:00:01.485027402 +0000 UTC m=+3078.769620877" watchObservedRunningTime="2025-11-21 15:00:01.492745141 +0000 UTC m=+3078.777338606" Nov 21 15:00:02 crc kubenswrapper[4897]: I1121 15:00:02.470579 4897 generic.go:334] "Generic (PLEG): container finished" podID="c2518e25-e7d2-4bbe-b2fb-1aa2f933069b" containerID="fdaaddea473e11bf7a52648c4831c17c899e611002c030ddf6c9f4ad8d48c283" exitCode=0 Nov 21 15:00:02 crc kubenswrapper[4897]: I1121 15:00:02.470679 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29395620-84swg" event={"ID":"c2518e25-e7d2-4bbe-b2fb-1aa2f933069b","Type":"ContainerDied","Data":"fdaaddea473e11bf7a52648c4831c17c899e611002c030ddf6c9f4ad8d48c283"} Nov 21 15:00:03 crc kubenswrapper[4897]: I1121 15:00:03.918180 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29395620-84swg" Nov 21 15:00:04 crc kubenswrapper[4897]: I1121 15:00:04.087131 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xchln\" (UniqueName: \"kubernetes.io/projected/c2518e25-e7d2-4bbe-b2fb-1aa2f933069b-kube-api-access-xchln\") pod \"c2518e25-e7d2-4bbe-b2fb-1aa2f933069b\" (UID: \"c2518e25-e7d2-4bbe-b2fb-1aa2f933069b\") " Nov 21 15:00:04 crc kubenswrapper[4897]: I1121 15:00:04.087450 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c2518e25-e7d2-4bbe-b2fb-1aa2f933069b-secret-volume\") pod \"c2518e25-e7d2-4bbe-b2fb-1aa2f933069b\" (UID: \"c2518e25-e7d2-4bbe-b2fb-1aa2f933069b\") " Nov 21 15:00:04 crc kubenswrapper[4897]: I1121 15:00:04.087627 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c2518e25-e7d2-4bbe-b2fb-1aa2f933069b-config-volume\") pod \"c2518e25-e7d2-4bbe-b2fb-1aa2f933069b\" (UID: \"c2518e25-e7d2-4bbe-b2fb-1aa2f933069b\") " Nov 21 15:00:04 crc kubenswrapper[4897]: I1121 15:00:04.088870 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c2518e25-e7d2-4bbe-b2fb-1aa2f933069b-config-volume" (OuterVolumeSpecName: "config-volume") pod "c2518e25-e7d2-4bbe-b2fb-1aa2f933069b" (UID: "c2518e25-e7d2-4bbe-b2fb-1aa2f933069b"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 15:00:04 crc kubenswrapper[4897]: I1121 15:00:04.093831 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2518e25-e7d2-4bbe-b2fb-1aa2f933069b-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "c2518e25-e7d2-4bbe-b2fb-1aa2f933069b" (UID: "c2518e25-e7d2-4bbe-b2fb-1aa2f933069b"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 15:00:04 crc kubenswrapper[4897]: I1121 15:00:04.094297 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2518e25-e7d2-4bbe-b2fb-1aa2f933069b-kube-api-access-xchln" (OuterVolumeSpecName: "kube-api-access-xchln") pod "c2518e25-e7d2-4bbe-b2fb-1aa2f933069b" (UID: "c2518e25-e7d2-4bbe-b2fb-1aa2f933069b"). InnerVolumeSpecName "kube-api-access-xchln". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 15:00:04 crc kubenswrapper[4897]: I1121 15:00:04.190392 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xchln\" (UniqueName: \"kubernetes.io/projected/c2518e25-e7d2-4bbe-b2fb-1aa2f933069b-kube-api-access-xchln\") on node \"crc\" DevicePath \"\"" Nov 21 15:00:04 crc kubenswrapper[4897]: I1121 15:00:04.190428 4897 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c2518e25-e7d2-4bbe-b2fb-1aa2f933069b-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 21 15:00:04 crc kubenswrapper[4897]: I1121 15:00:04.190442 4897 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c2518e25-e7d2-4bbe-b2fb-1aa2f933069b-config-volume\") on node \"crc\" DevicePath \"\"" Nov 21 15:00:04 crc kubenswrapper[4897]: I1121 15:00:04.495539 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29395620-84swg" event={"ID":"c2518e25-e7d2-4bbe-b2fb-1aa2f933069b","Type":"ContainerDied","Data":"b714417e3c09f0879f27569e82ff1b79e12e75c40ae938c517bed93db6148dbd"} Nov 21 15:00:04 crc kubenswrapper[4897]: I1121 15:00:04.497773 4897 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b714417e3c09f0879f27569e82ff1b79e12e75c40ae938c517bed93db6148dbd" Nov 21 15:00:04 crc kubenswrapper[4897]: I1121 15:00:04.495872 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29395620-84swg" Nov 21 15:00:04 crc kubenswrapper[4897]: I1121 15:00:04.997391 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29395575-tx6lf"] Nov 21 15:00:05 crc kubenswrapper[4897]: I1121 15:00:05.009597 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29395575-tx6lf"] Nov 21 15:00:06 crc kubenswrapper[4897]: I1121 15:00:06.109229 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7def41c-7fc4-44a5-8ae4-71ebd66ff255" path="/var/lib/kubelet/pods/e7def41c-7fc4-44a5-8ae4-71ebd66ff255/volumes" Nov 21 15:00:11 crc kubenswrapper[4897]: I1121 15:00:11.090110 4897 scope.go:117] "RemoveContainer" containerID="a17b834dbb567d490c0f7ce8a22ab5aa6ecb645fc460c063af4ca43f58348563" Nov 21 15:00:11 crc kubenswrapper[4897]: E1121 15:00:11.090963 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:00:22 crc kubenswrapper[4897]: I1121 15:00:22.290571 4897 scope.go:117] "RemoveContainer" containerID="044b03bc8a8b588143aadbdbb3974f4bf577f413dc45ae7a95dd44cd65eb0b47" Nov 21 15:00:25 crc kubenswrapper[4897]: I1121 15:00:25.089564 4897 scope.go:117] "RemoveContainer" containerID="a17b834dbb567d490c0f7ce8a22ab5aa6ecb645fc460c063af4ca43f58348563" Nov 21 15:00:25 crc kubenswrapper[4897]: E1121 15:00:25.090559 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:00:39 crc kubenswrapper[4897]: I1121 15:00:39.089912 4897 scope.go:117] "RemoveContainer" containerID="a17b834dbb567d490c0f7ce8a22ab5aa6ecb645fc460c063af4ca43f58348563" Nov 21 15:00:39 crc kubenswrapper[4897]: E1121 15:00:39.090955 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:00:50 crc kubenswrapper[4897]: I1121 15:00:50.089366 4897 scope.go:117] "RemoveContainer" containerID="a17b834dbb567d490c0f7ce8a22ab5aa6ecb645fc460c063af4ca43f58348563" Nov 21 15:00:50 crc kubenswrapper[4897]: E1121 15:00:50.090407 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:01:00 crc kubenswrapper[4897]: I1121 15:01:00.173094 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29395621-r2l64"] Nov 21 15:01:00 crc kubenswrapper[4897]: E1121 15:01:00.177400 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2518e25-e7d2-4bbe-b2fb-1aa2f933069b" containerName="collect-profiles" Nov 21 15:01:00 crc kubenswrapper[4897]: I1121 15:01:00.177424 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2518e25-e7d2-4bbe-b2fb-1aa2f933069b" containerName="collect-profiles" Nov 21 15:01:00 crc kubenswrapper[4897]: I1121 15:01:00.177661 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2518e25-e7d2-4bbe-b2fb-1aa2f933069b" containerName="collect-profiles" Nov 21 15:01:00 crc kubenswrapper[4897]: I1121 15:01:00.178601 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29395621-r2l64" Nov 21 15:01:00 crc kubenswrapper[4897]: I1121 15:01:00.186703 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29395621-r2l64"] Nov 21 15:01:00 crc kubenswrapper[4897]: I1121 15:01:00.289047 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dlhr7\" (UniqueName: \"kubernetes.io/projected/693a8cfa-1643-4928-b38f-d4708eb3e1b1-kube-api-access-dlhr7\") pod \"keystone-cron-29395621-r2l64\" (UID: \"693a8cfa-1643-4928-b38f-d4708eb3e1b1\") " pod="openstack/keystone-cron-29395621-r2l64" Nov 21 15:01:00 crc kubenswrapper[4897]: I1121 15:01:00.289105 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/693a8cfa-1643-4928-b38f-d4708eb3e1b1-combined-ca-bundle\") pod \"keystone-cron-29395621-r2l64\" (UID: \"693a8cfa-1643-4928-b38f-d4708eb3e1b1\") " pod="openstack/keystone-cron-29395621-r2l64" Nov 21 15:01:00 crc kubenswrapper[4897]: I1121 15:01:00.289295 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/693a8cfa-1643-4928-b38f-d4708eb3e1b1-fernet-keys\") pod \"keystone-cron-29395621-r2l64\" (UID: \"693a8cfa-1643-4928-b38f-d4708eb3e1b1\") " pod="openstack/keystone-cron-29395621-r2l64" Nov 21 15:01:00 crc kubenswrapper[4897]: I1121 15:01:00.289475 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/693a8cfa-1643-4928-b38f-d4708eb3e1b1-config-data\") pod \"keystone-cron-29395621-r2l64\" (UID: \"693a8cfa-1643-4928-b38f-d4708eb3e1b1\") " pod="openstack/keystone-cron-29395621-r2l64" Nov 21 15:01:00 crc kubenswrapper[4897]: I1121 15:01:00.392149 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/693a8cfa-1643-4928-b38f-d4708eb3e1b1-config-data\") pod \"keystone-cron-29395621-r2l64\" (UID: \"693a8cfa-1643-4928-b38f-d4708eb3e1b1\") " pod="openstack/keystone-cron-29395621-r2l64" Nov 21 15:01:00 crc kubenswrapper[4897]: I1121 15:01:00.392477 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dlhr7\" (UniqueName: \"kubernetes.io/projected/693a8cfa-1643-4928-b38f-d4708eb3e1b1-kube-api-access-dlhr7\") pod \"keystone-cron-29395621-r2l64\" (UID: \"693a8cfa-1643-4928-b38f-d4708eb3e1b1\") " pod="openstack/keystone-cron-29395621-r2l64" Nov 21 15:01:00 crc kubenswrapper[4897]: I1121 15:01:00.392565 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/693a8cfa-1643-4928-b38f-d4708eb3e1b1-combined-ca-bundle\") pod \"keystone-cron-29395621-r2l64\" (UID: \"693a8cfa-1643-4928-b38f-d4708eb3e1b1\") " pod="openstack/keystone-cron-29395621-r2l64" Nov 21 15:01:00 crc kubenswrapper[4897]: I1121 15:01:00.392849 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/693a8cfa-1643-4928-b38f-d4708eb3e1b1-fernet-keys\") pod \"keystone-cron-29395621-r2l64\" (UID: \"693a8cfa-1643-4928-b38f-d4708eb3e1b1\") " pod="openstack/keystone-cron-29395621-r2l64" Nov 21 15:01:00 crc kubenswrapper[4897]: I1121 15:01:00.399451 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/693a8cfa-1643-4928-b38f-d4708eb3e1b1-combined-ca-bundle\") pod \"keystone-cron-29395621-r2l64\" (UID: \"693a8cfa-1643-4928-b38f-d4708eb3e1b1\") " pod="openstack/keystone-cron-29395621-r2l64" Nov 21 15:01:00 crc kubenswrapper[4897]: I1121 15:01:00.399896 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/693a8cfa-1643-4928-b38f-d4708eb3e1b1-config-data\") pod \"keystone-cron-29395621-r2l64\" (UID: \"693a8cfa-1643-4928-b38f-d4708eb3e1b1\") " pod="openstack/keystone-cron-29395621-r2l64" Nov 21 15:01:00 crc kubenswrapper[4897]: I1121 15:01:00.400070 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/693a8cfa-1643-4928-b38f-d4708eb3e1b1-fernet-keys\") pod \"keystone-cron-29395621-r2l64\" (UID: \"693a8cfa-1643-4928-b38f-d4708eb3e1b1\") " pod="openstack/keystone-cron-29395621-r2l64" Nov 21 15:01:00 crc kubenswrapper[4897]: I1121 15:01:00.409314 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dlhr7\" (UniqueName: \"kubernetes.io/projected/693a8cfa-1643-4928-b38f-d4708eb3e1b1-kube-api-access-dlhr7\") pod \"keystone-cron-29395621-r2l64\" (UID: \"693a8cfa-1643-4928-b38f-d4708eb3e1b1\") " pod="openstack/keystone-cron-29395621-r2l64" Nov 21 15:01:00 crc kubenswrapper[4897]: I1121 15:01:00.506310 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29395621-r2l64" Nov 21 15:01:00 crc kubenswrapper[4897]: I1121 15:01:00.997471 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29395621-r2l64"] Nov 21 15:01:01 crc kubenswrapper[4897]: I1121 15:01:01.089403 4897 scope.go:117] "RemoveContainer" containerID="a17b834dbb567d490c0f7ce8a22ab5aa6ecb645fc460c063af4ca43f58348563" Nov 21 15:01:01 crc kubenswrapper[4897]: E1121 15:01:01.089774 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:01:01 crc kubenswrapper[4897]: I1121 15:01:01.162553 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29395621-r2l64" event={"ID":"693a8cfa-1643-4928-b38f-d4708eb3e1b1","Type":"ContainerStarted","Data":"619175fd88074996c0a2b711ef0b4679b776d90f58cea51510fc2eda468ce136"} Nov 21 15:01:02 crc kubenswrapper[4897]: I1121 15:01:02.173094 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29395621-r2l64" event={"ID":"693a8cfa-1643-4928-b38f-d4708eb3e1b1","Type":"ContainerStarted","Data":"5ebdea3f94d3695a8b9f350275646c0abb5df0ceb25ba52e4a9216593174e73c"} Nov 21 15:01:02 crc kubenswrapper[4897]: I1121 15:01:02.194440 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29395621-r2l64" podStartSLOduration=2.19442095 podStartE2EDuration="2.19442095s" podCreationTimestamp="2025-11-21 15:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 15:01:02.190545326 +0000 UTC m=+3139.475138821" watchObservedRunningTime="2025-11-21 15:01:02.19442095 +0000 UTC m=+3139.479014425" Nov 21 15:01:09 crc kubenswrapper[4897]: I1121 15:01:09.256537 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29395621-r2l64" event={"ID":"693a8cfa-1643-4928-b38f-d4708eb3e1b1","Type":"ContainerDied","Data":"5ebdea3f94d3695a8b9f350275646c0abb5df0ceb25ba52e4a9216593174e73c"} Nov 21 15:01:09 crc kubenswrapper[4897]: I1121 15:01:09.256486 4897 generic.go:334] "Generic (PLEG): container finished" podID="693a8cfa-1643-4928-b38f-d4708eb3e1b1" containerID="5ebdea3f94d3695a8b9f350275646c0abb5df0ceb25ba52e4a9216593174e73c" exitCode=0 Nov 21 15:01:10 crc kubenswrapper[4897]: I1121 15:01:10.677172 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29395621-r2l64" Nov 21 15:01:10 crc kubenswrapper[4897]: I1121 15:01:10.751727 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/693a8cfa-1643-4928-b38f-d4708eb3e1b1-combined-ca-bundle\") pod \"693a8cfa-1643-4928-b38f-d4708eb3e1b1\" (UID: \"693a8cfa-1643-4928-b38f-d4708eb3e1b1\") " Nov 21 15:01:10 crc kubenswrapper[4897]: I1121 15:01:10.751907 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/693a8cfa-1643-4928-b38f-d4708eb3e1b1-fernet-keys\") pod \"693a8cfa-1643-4928-b38f-d4708eb3e1b1\" (UID: \"693a8cfa-1643-4928-b38f-d4708eb3e1b1\") " Nov 21 15:01:10 crc kubenswrapper[4897]: I1121 15:01:10.752041 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/693a8cfa-1643-4928-b38f-d4708eb3e1b1-config-data\") pod \"693a8cfa-1643-4928-b38f-d4708eb3e1b1\" (UID: \"693a8cfa-1643-4928-b38f-d4708eb3e1b1\") " Nov 21 15:01:10 crc kubenswrapper[4897]: I1121 15:01:10.752183 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dlhr7\" (UniqueName: \"kubernetes.io/projected/693a8cfa-1643-4928-b38f-d4708eb3e1b1-kube-api-access-dlhr7\") pod \"693a8cfa-1643-4928-b38f-d4708eb3e1b1\" (UID: \"693a8cfa-1643-4928-b38f-d4708eb3e1b1\") " Nov 21 15:01:10 crc kubenswrapper[4897]: I1121 15:01:10.760644 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/693a8cfa-1643-4928-b38f-d4708eb3e1b1-kube-api-access-dlhr7" (OuterVolumeSpecName: "kube-api-access-dlhr7") pod "693a8cfa-1643-4928-b38f-d4708eb3e1b1" (UID: "693a8cfa-1643-4928-b38f-d4708eb3e1b1"). InnerVolumeSpecName "kube-api-access-dlhr7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 15:01:10 crc kubenswrapper[4897]: I1121 15:01:10.761758 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/693a8cfa-1643-4928-b38f-d4708eb3e1b1-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "693a8cfa-1643-4928-b38f-d4708eb3e1b1" (UID: "693a8cfa-1643-4928-b38f-d4708eb3e1b1"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 15:01:10 crc kubenswrapper[4897]: I1121 15:01:10.799197 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/693a8cfa-1643-4928-b38f-d4708eb3e1b1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "693a8cfa-1643-4928-b38f-d4708eb3e1b1" (UID: "693a8cfa-1643-4928-b38f-d4708eb3e1b1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 15:01:10 crc kubenswrapper[4897]: I1121 15:01:10.821235 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/693a8cfa-1643-4928-b38f-d4708eb3e1b1-config-data" (OuterVolumeSpecName: "config-data") pod "693a8cfa-1643-4928-b38f-d4708eb3e1b1" (UID: "693a8cfa-1643-4928-b38f-d4708eb3e1b1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 15:01:10 crc kubenswrapper[4897]: I1121 15:01:10.854982 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dlhr7\" (UniqueName: \"kubernetes.io/projected/693a8cfa-1643-4928-b38f-d4708eb3e1b1-kube-api-access-dlhr7\") on node \"crc\" DevicePath \"\"" Nov 21 15:01:10 crc kubenswrapper[4897]: I1121 15:01:10.855289 4897 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/693a8cfa-1643-4928-b38f-d4708eb3e1b1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 21 15:01:10 crc kubenswrapper[4897]: I1121 15:01:10.855306 4897 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/693a8cfa-1643-4928-b38f-d4708eb3e1b1-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 21 15:01:10 crc kubenswrapper[4897]: I1121 15:01:10.855318 4897 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/693a8cfa-1643-4928-b38f-d4708eb3e1b1-config-data\") on node \"crc\" DevicePath \"\"" Nov 21 15:01:11 crc kubenswrapper[4897]: I1121 15:01:11.279683 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29395621-r2l64" event={"ID":"693a8cfa-1643-4928-b38f-d4708eb3e1b1","Type":"ContainerDied","Data":"619175fd88074996c0a2b711ef0b4679b776d90f58cea51510fc2eda468ce136"} Nov 21 15:01:11 crc kubenswrapper[4897]: I1121 15:01:11.279734 4897 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="619175fd88074996c0a2b711ef0b4679b776d90f58cea51510fc2eda468ce136" Nov 21 15:01:11 crc kubenswrapper[4897]: I1121 15:01:11.279758 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29395621-r2l64" Nov 21 15:01:14 crc kubenswrapper[4897]: I1121 15:01:14.088858 4897 scope.go:117] "RemoveContainer" containerID="a17b834dbb567d490c0f7ce8a22ab5aa6ecb645fc460c063af4ca43f58348563" Nov 21 15:01:14 crc kubenswrapper[4897]: E1121 15:01:14.089760 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:01:28 crc kubenswrapper[4897]: I1121 15:01:28.089422 4897 scope.go:117] "RemoveContainer" containerID="a17b834dbb567d490c0f7ce8a22ab5aa6ecb645fc460c063af4ca43f58348563" Nov 21 15:01:28 crc kubenswrapper[4897]: E1121 15:01:28.090295 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:01:40 crc kubenswrapper[4897]: I1121 15:01:40.089865 4897 scope.go:117] "RemoveContainer" containerID="a17b834dbb567d490c0f7ce8a22ab5aa6ecb645fc460c063af4ca43f58348563" Nov 21 15:01:40 crc kubenswrapper[4897]: E1121 15:01:40.090883 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:01:52 crc kubenswrapper[4897]: I1121 15:01:52.088974 4897 scope.go:117] "RemoveContainer" containerID="a17b834dbb567d490c0f7ce8a22ab5aa6ecb645fc460c063af4ca43f58348563" Nov 21 15:01:52 crc kubenswrapper[4897]: E1121 15:01:52.089876 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:02:03 crc kubenswrapper[4897]: I1121 15:02:03.090454 4897 scope.go:117] "RemoveContainer" containerID="a17b834dbb567d490c0f7ce8a22ab5aa6ecb645fc460c063af4ca43f58348563" Nov 21 15:02:03 crc kubenswrapper[4897]: E1121 15:02:03.091254 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:02:09 crc kubenswrapper[4897]: I1121 15:02:09.547402 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-5p6nn"] Nov 21 15:02:09 crc kubenswrapper[4897]: E1121 15:02:09.548763 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="693a8cfa-1643-4928-b38f-d4708eb3e1b1" containerName="keystone-cron" Nov 21 15:02:09 crc kubenswrapper[4897]: I1121 15:02:09.548781 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="693a8cfa-1643-4928-b38f-d4708eb3e1b1" containerName="keystone-cron" Nov 21 15:02:09 crc kubenswrapper[4897]: I1121 15:02:09.549157 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="693a8cfa-1643-4928-b38f-d4708eb3e1b1" containerName="keystone-cron" Nov 21 15:02:09 crc kubenswrapper[4897]: I1121 15:02:09.554108 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5p6nn" Nov 21 15:02:09 crc kubenswrapper[4897]: I1121 15:02:09.566479 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5p6nn"] Nov 21 15:02:09 crc kubenswrapper[4897]: I1121 15:02:09.574471 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0453db98-4ad2-42d2-8781-947a3d03eb7c-utilities\") pod \"community-operators-5p6nn\" (UID: \"0453db98-4ad2-42d2-8781-947a3d03eb7c\") " pod="openshift-marketplace/community-operators-5p6nn" Nov 21 15:02:09 crc kubenswrapper[4897]: I1121 15:02:09.574617 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v9lb8\" (UniqueName: \"kubernetes.io/projected/0453db98-4ad2-42d2-8781-947a3d03eb7c-kube-api-access-v9lb8\") pod \"community-operators-5p6nn\" (UID: \"0453db98-4ad2-42d2-8781-947a3d03eb7c\") " pod="openshift-marketplace/community-operators-5p6nn" Nov 21 15:02:09 crc kubenswrapper[4897]: I1121 15:02:09.574764 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0453db98-4ad2-42d2-8781-947a3d03eb7c-catalog-content\") pod \"community-operators-5p6nn\" (UID: \"0453db98-4ad2-42d2-8781-947a3d03eb7c\") " pod="openshift-marketplace/community-operators-5p6nn" Nov 21 15:02:09 crc kubenswrapper[4897]: I1121 15:02:09.677314 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0453db98-4ad2-42d2-8781-947a3d03eb7c-catalog-content\") pod \"community-operators-5p6nn\" (UID: \"0453db98-4ad2-42d2-8781-947a3d03eb7c\") " pod="openshift-marketplace/community-operators-5p6nn" Nov 21 15:02:09 crc kubenswrapper[4897]: I1121 15:02:09.677565 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0453db98-4ad2-42d2-8781-947a3d03eb7c-utilities\") pod \"community-operators-5p6nn\" (UID: \"0453db98-4ad2-42d2-8781-947a3d03eb7c\") " pod="openshift-marketplace/community-operators-5p6nn" Nov 21 15:02:09 crc kubenswrapper[4897]: I1121 15:02:09.677709 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v9lb8\" (UniqueName: \"kubernetes.io/projected/0453db98-4ad2-42d2-8781-947a3d03eb7c-kube-api-access-v9lb8\") pod \"community-operators-5p6nn\" (UID: \"0453db98-4ad2-42d2-8781-947a3d03eb7c\") " pod="openshift-marketplace/community-operators-5p6nn" Nov 21 15:02:09 crc kubenswrapper[4897]: I1121 15:02:09.677984 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0453db98-4ad2-42d2-8781-947a3d03eb7c-catalog-content\") pod \"community-operators-5p6nn\" (UID: \"0453db98-4ad2-42d2-8781-947a3d03eb7c\") " pod="openshift-marketplace/community-operators-5p6nn" Nov 21 15:02:09 crc kubenswrapper[4897]: I1121 15:02:09.678048 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0453db98-4ad2-42d2-8781-947a3d03eb7c-utilities\") pod \"community-operators-5p6nn\" (UID: \"0453db98-4ad2-42d2-8781-947a3d03eb7c\") " pod="openshift-marketplace/community-operators-5p6nn" Nov 21 15:02:09 crc kubenswrapper[4897]: I1121 15:02:09.705627 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v9lb8\" (UniqueName: \"kubernetes.io/projected/0453db98-4ad2-42d2-8781-947a3d03eb7c-kube-api-access-v9lb8\") pod \"community-operators-5p6nn\" (UID: \"0453db98-4ad2-42d2-8781-947a3d03eb7c\") " pod="openshift-marketplace/community-operators-5p6nn" Nov 21 15:02:09 crc kubenswrapper[4897]: I1121 15:02:09.874969 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5p6nn" Nov 21 15:02:10 crc kubenswrapper[4897]: I1121 15:02:10.392033 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5p6nn"] Nov 21 15:02:10 crc kubenswrapper[4897]: I1121 15:02:10.996959 4897 generic.go:334] "Generic (PLEG): container finished" podID="0453db98-4ad2-42d2-8781-947a3d03eb7c" containerID="7268ac6c4a9e2087bb67f8578c2b222c1175c7cdf575906a1bb46d3e8f79dc61" exitCode=0 Nov 21 15:02:10 crc kubenswrapper[4897]: I1121 15:02:10.997253 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5p6nn" event={"ID":"0453db98-4ad2-42d2-8781-947a3d03eb7c","Type":"ContainerDied","Data":"7268ac6c4a9e2087bb67f8578c2b222c1175c7cdf575906a1bb46d3e8f79dc61"} Nov 21 15:02:10 crc kubenswrapper[4897]: I1121 15:02:10.997285 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5p6nn" event={"ID":"0453db98-4ad2-42d2-8781-947a3d03eb7c","Type":"ContainerStarted","Data":"c3efa88603a9761b6309b35a16a468a3f81c045a0ea8562416b64a2c490948f0"} Nov 21 15:02:11 crc kubenswrapper[4897]: I1121 15:02:10.999429 4897 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 21 15:02:12 crc kubenswrapper[4897]: I1121 15:02:12.011093 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5p6nn" event={"ID":"0453db98-4ad2-42d2-8781-947a3d03eb7c","Type":"ContainerStarted","Data":"27f0c5acf4fe887511df3c66309c0b7e3880a613ed0fcc5a8e6c0536eb2dd10d"} Nov 21 15:02:15 crc kubenswrapper[4897]: I1121 15:02:15.047775 4897 generic.go:334] "Generic (PLEG): container finished" podID="0453db98-4ad2-42d2-8781-947a3d03eb7c" containerID="27f0c5acf4fe887511df3c66309c0b7e3880a613ed0fcc5a8e6c0536eb2dd10d" exitCode=0 Nov 21 15:02:15 crc kubenswrapper[4897]: I1121 15:02:15.047905 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5p6nn" event={"ID":"0453db98-4ad2-42d2-8781-947a3d03eb7c","Type":"ContainerDied","Data":"27f0c5acf4fe887511df3c66309c0b7e3880a613ed0fcc5a8e6c0536eb2dd10d"} Nov 21 15:02:16 crc kubenswrapper[4897]: I1121 15:02:16.059975 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5p6nn" event={"ID":"0453db98-4ad2-42d2-8781-947a3d03eb7c","Type":"ContainerStarted","Data":"3d062632583d242643381a7616929aba9efb179f4149eb463b407b2c7b068763"} Nov 21 15:02:16 crc kubenswrapper[4897]: I1121 15:02:16.093055 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-5p6nn" podStartSLOduration=2.498284317 podStartE2EDuration="7.093032847s" podCreationTimestamp="2025-11-21 15:02:09 +0000 UTC" firstStartedPulling="2025-11-21 15:02:10.999168564 +0000 UTC m=+3208.283762039" lastFinishedPulling="2025-11-21 15:02:15.593917084 +0000 UTC m=+3212.878510569" observedRunningTime="2025-11-21 15:02:16.078443849 +0000 UTC m=+3213.363037334" watchObservedRunningTime="2025-11-21 15:02:16.093032847 +0000 UTC m=+3213.377626322" Nov 21 15:02:18 crc kubenswrapper[4897]: I1121 15:02:18.089976 4897 scope.go:117] "RemoveContainer" containerID="a17b834dbb567d490c0f7ce8a22ab5aa6ecb645fc460c063af4ca43f58348563" Nov 21 15:02:18 crc kubenswrapper[4897]: E1121 15:02:18.090982 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:02:19 crc kubenswrapper[4897]: I1121 15:02:19.875795 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-5p6nn" Nov 21 15:02:19 crc kubenswrapper[4897]: I1121 15:02:19.876392 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-5p6nn" Nov 21 15:02:19 crc kubenswrapper[4897]: I1121 15:02:19.935199 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-5p6nn" Nov 21 15:02:20 crc kubenswrapper[4897]: I1121 15:02:20.146154 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-5p6nn" Nov 21 15:02:20 crc kubenswrapper[4897]: I1121 15:02:20.194906 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5p6nn"] Nov 21 15:02:22 crc kubenswrapper[4897]: I1121 15:02:22.117257 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-5p6nn" podUID="0453db98-4ad2-42d2-8781-947a3d03eb7c" containerName="registry-server" containerID="cri-o://3d062632583d242643381a7616929aba9efb179f4149eb463b407b2c7b068763" gracePeriod=2 Nov 21 15:02:22 crc kubenswrapper[4897]: I1121 15:02:22.747927 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5p6nn" Nov 21 15:02:22 crc kubenswrapper[4897]: I1121 15:02:22.803048 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v9lb8\" (UniqueName: \"kubernetes.io/projected/0453db98-4ad2-42d2-8781-947a3d03eb7c-kube-api-access-v9lb8\") pod \"0453db98-4ad2-42d2-8781-947a3d03eb7c\" (UID: \"0453db98-4ad2-42d2-8781-947a3d03eb7c\") " Nov 21 15:02:22 crc kubenswrapper[4897]: I1121 15:02:22.803418 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0453db98-4ad2-42d2-8781-947a3d03eb7c-utilities\") pod \"0453db98-4ad2-42d2-8781-947a3d03eb7c\" (UID: \"0453db98-4ad2-42d2-8781-947a3d03eb7c\") " Nov 21 15:02:22 crc kubenswrapper[4897]: I1121 15:02:22.803590 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0453db98-4ad2-42d2-8781-947a3d03eb7c-catalog-content\") pod \"0453db98-4ad2-42d2-8781-947a3d03eb7c\" (UID: \"0453db98-4ad2-42d2-8781-947a3d03eb7c\") " Nov 21 15:02:22 crc kubenswrapper[4897]: I1121 15:02:22.804291 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0453db98-4ad2-42d2-8781-947a3d03eb7c-utilities" (OuterVolumeSpecName: "utilities") pod "0453db98-4ad2-42d2-8781-947a3d03eb7c" (UID: "0453db98-4ad2-42d2-8781-947a3d03eb7c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 15:02:22 crc kubenswrapper[4897]: I1121 15:02:22.809930 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0453db98-4ad2-42d2-8781-947a3d03eb7c-kube-api-access-v9lb8" (OuterVolumeSpecName: "kube-api-access-v9lb8") pod "0453db98-4ad2-42d2-8781-947a3d03eb7c" (UID: "0453db98-4ad2-42d2-8781-947a3d03eb7c"). InnerVolumeSpecName "kube-api-access-v9lb8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 15:02:22 crc kubenswrapper[4897]: I1121 15:02:22.871967 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0453db98-4ad2-42d2-8781-947a3d03eb7c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0453db98-4ad2-42d2-8781-947a3d03eb7c" (UID: "0453db98-4ad2-42d2-8781-947a3d03eb7c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 15:02:22 crc kubenswrapper[4897]: I1121 15:02:22.905690 4897 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0453db98-4ad2-42d2-8781-947a3d03eb7c-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 21 15:02:22 crc kubenswrapper[4897]: I1121 15:02:22.905732 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v9lb8\" (UniqueName: \"kubernetes.io/projected/0453db98-4ad2-42d2-8781-947a3d03eb7c-kube-api-access-v9lb8\") on node \"crc\" DevicePath \"\"" Nov 21 15:02:22 crc kubenswrapper[4897]: I1121 15:02:22.905752 4897 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0453db98-4ad2-42d2-8781-947a3d03eb7c-utilities\") on node \"crc\" DevicePath \"\"" Nov 21 15:02:23 crc kubenswrapper[4897]: I1121 15:02:23.143971 4897 generic.go:334] "Generic (PLEG): container finished" podID="0453db98-4ad2-42d2-8781-947a3d03eb7c" containerID="3d062632583d242643381a7616929aba9efb179f4149eb463b407b2c7b068763" exitCode=0 Nov 21 15:02:23 crc kubenswrapper[4897]: I1121 15:02:23.145722 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5p6nn" event={"ID":"0453db98-4ad2-42d2-8781-947a3d03eb7c","Type":"ContainerDied","Data":"3d062632583d242643381a7616929aba9efb179f4149eb463b407b2c7b068763"} Nov 21 15:02:23 crc kubenswrapper[4897]: I1121 15:02:23.145889 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5p6nn" event={"ID":"0453db98-4ad2-42d2-8781-947a3d03eb7c","Type":"ContainerDied","Data":"c3efa88603a9761b6309b35a16a468a3f81c045a0ea8562416b64a2c490948f0"} Nov 21 15:02:23 crc kubenswrapper[4897]: I1121 15:02:23.145988 4897 scope.go:117] "RemoveContainer" containerID="3d062632583d242643381a7616929aba9efb179f4149eb463b407b2c7b068763" Nov 21 15:02:23 crc kubenswrapper[4897]: I1121 15:02:23.146456 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5p6nn" Nov 21 15:02:23 crc kubenswrapper[4897]: I1121 15:02:23.173331 4897 scope.go:117] "RemoveContainer" containerID="27f0c5acf4fe887511df3c66309c0b7e3880a613ed0fcc5a8e6c0536eb2dd10d" Nov 21 15:02:23 crc kubenswrapper[4897]: I1121 15:02:23.193979 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5p6nn"] Nov 21 15:02:23 crc kubenswrapper[4897]: I1121 15:02:23.204468 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-5p6nn"] Nov 21 15:02:23 crc kubenswrapper[4897]: I1121 15:02:23.231337 4897 scope.go:117] "RemoveContainer" containerID="7268ac6c4a9e2087bb67f8578c2b222c1175c7cdf575906a1bb46d3e8f79dc61" Nov 21 15:02:23 crc kubenswrapper[4897]: I1121 15:02:23.259703 4897 scope.go:117] "RemoveContainer" containerID="3d062632583d242643381a7616929aba9efb179f4149eb463b407b2c7b068763" Nov 21 15:02:23 crc kubenswrapper[4897]: E1121 15:02:23.260286 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d062632583d242643381a7616929aba9efb179f4149eb463b407b2c7b068763\": container with ID starting with 3d062632583d242643381a7616929aba9efb179f4149eb463b407b2c7b068763 not found: ID does not exist" containerID="3d062632583d242643381a7616929aba9efb179f4149eb463b407b2c7b068763" Nov 21 15:02:23 crc kubenswrapper[4897]: I1121 15:02:23.260319 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d062632583d242643381a7616929aba9efb179f4149eb463b407b2c7b068763"} err="failed to get container status \"3d062632583d242643381a7616929aba9efb179f4149eb463b407b2c7b068763\": rpc error: code = NotFound desc = could not find container \"3d062632583d242643381a7616929aba9efb179f4149eb463b407b2c7b068763\": container with ID starting with 3d062632583d242643381a7616929aba9efb179f4149eb463b407b2c7b068763 not found: ID does not exist" Nov 21 15:02:23 crc kubenswrapper[4897]: I1121 15:02:23.260340 4897 scope.go:117] "RemoveContainer" containerID="27f0c5acf4fe887511df3c66309c0b7e3880a613ed0fcc5a8e6c0536eb2dd10d" Nov 21 15:02:23 crc kubenswrapper[4897]: E1121 15:02:23.260730 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"27f0c5acf4fe887511df3c66309c0b7e3880a613ed0fcc5a8e6c0536eb2dd10d\": container with ID starting with 27f0c5acf4fe887511df3c66309c0b7e3880a613ed0fcc5a8e6c0536eb2dd10d not found: ID does not exist" containerID="27f0c5acf4fe887511df3c66309c0b7e3880a613ed0fcc5a8e6c0536eb2dd10d" Nov 21 15:02:23 crc kubenswrapper[4897]: I1121 15:02:23.260752 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"27f0c5acf4fe887511df3c66309c0b7e3880a613ed0fcc5a8e6c0536eb2dd10d"} err="failed to get container status \"27f0c5acf4fe887511df3c66309c0b7e3880a613ed0fcc5a8e6c0536eb2dd10d\": rpc error: code = NotFound desc = could not find container \"27f0c5acf4fe887511df3c66309c0b7e3880a613ed0fcc5a8e6c0536eb2dd10d\": container with ID starting with 27f0c5acf4fe887511df3c66309c0b7e3880a613ed0fcc5a8e6c0536eb2dd10d not found: ID does not exist" Nov 21 15:02:23 crc kubenswrapper[4897]: I1121 15:02:23.260765 4897 scope.go:117] "RemoveContainer" containerID="7268ac6c4a9e2087bb67f8578c2b222c1175c7cdf575906a1bb46d3e8f79dc61" Nov 21 15:02:23 crc kubenswrapper[4897]: E1121 15:02:23.261070 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7268ac6c4a9e2087bb67f8578c2b222c1175c7cdf575906a1bb46d3e8f79dc61\": container with ID starting with 7268ac6c4a9e2087bb67f8578c2b222c1175c7cdf575906a1bb46d3e8f79dc61 not found: ID does not exist" containerID="7268ac6c4a9e2087bb67f8578c2b222c1175c7cdf575906a1bb46d3e8f79dc61" Nov 21 15:02:23 crc kubenswrapper[4897]: I1121 15:02:23.261100 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7268ac6c4a9e2087bb67f8578c2b222c1175c7cdf575906a1bb46d3e8f79dc61"} err="failed to get container status \"7268ac6c4a9e2087bb67f8578c2b222c1175c7cdf575906a1bb46d3e8f79dc61\": rpc error: code = NotFound desc = could not find container \"7268ac6c4a9e2087bb67f8578c2b222c1175c7cdf575906a1bb46d3e8f79dc61\": container with ID starting with 7268ac6c4a9e2087bb67f8578c2b222c1175c7cdf575906a1bb46d3e8f79dc61 not found: ID does not exist" Nov 21 15:02:24 crc kubenswrapper[4897]: I1121 15:02:24.103674 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0453db98-4ad2-42d2-8781-947a3d03eb7c" path="/var/lib/kubelet/pods/0453db98-4ad2-42d2-8781-947a3d03eb7c/volumes" Nov 21 15:02:33 crc kubenswrapper[4897]: I1121 15:02:33.089688 4897 scope.go:117] "RemoveContainer" containerID="a17b834dbb567d490c0f7ce8a22ab5aa6ecb645fc460c063af4ca43f58348563" Nov 21 15:02:33 crc kubenswrapper[4897]: E1121 15:02:33.090720 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:02:45 crc kubenswrapper[4897]: I1121 15:02:45.089531 4897 scope.go:117] "RemoveContainer" containerID="a17b834dbb567d490c0f7ce8a22ab5aa6ecb645fc460c063af4ca43f58348563" Nov 21 15:02:45 crc kubenswrapper[4897]: E1121 15:02:45.090477 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:02:57 crc kubenswrapper[4897]: I1121 15:02:57.091619 4897 scope.go:117] "RemoveContainer" containerID="a17b834dbb567d490c0f7ce8a22ab5aa6ecb645fc460c063af4ca43f58348563" Nov 21 15:02:57 crc kubenswrapper[4897]: E1121 15:02:57.092944 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:02:57 crc kubenswrapper[4897]: I1121 15:02:57.552246 4897 generic.go:334] "Generic (PLEG): container finished" podID="9f8e267b-7f34-4793-bb66-a2066bf63ce5" containerID="3745add26e7e03d442c861649da7caad1fcf48a01734c3ff971dc877553b76e0" exitCode=0 Nov 21 15:02:57 crc kubenswrapper[4897]: I1121 15:02:57.552305 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tmr6q" event={"ID":"9f8e267b-7f34-4793-bb66-a2066bf63ce5","Type":"ContainerDied","Data":"3745add26e7e03d442c861649da7caad1fcf48a01734c3ff971dc877553b76e0"} Nov 21 15:02:59 crc kubenswrapper[4897]: I1121 15:02:59.014851 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tmr6q" Nov 21 15:02:59 crc kubenswrapper[4897]: I1121 15:02:59.126718 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d2ft2\" (UniqueName: \"kubernetes.io/projected/9f8e267b-7f34-4793-bb66-a2066bf63ce5-kube-api-access-d2ft2\") pod \"9f8e267b-7f34-4793-bb66-a2066bf63ce5\" (UID: \"9f8e267b-7f34-4793-bb66-a2066bf63ce5\") " Nov 21 15:02:59 crc kubenswrapper[4897]: I1121 15:02:59.126831 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9f8e267b-7f34-4793-bb66-a2066bf63ce5-ssh-key\") pod \"9f8e267b-7f34-4793-bb66-a2066bf63ce5\" (UID: \"9f8e267b-7f34-4793-bb66-a2066bf63ce5\") " Nov 21 15:02:59 crc kubenswrapper[4897]: I1121 15:02:59.126943 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/9f8e267b-7f34-4793-bb66-a2066bf63ce5-libvirt-secret-0\") pod \"9f8e267b-7f34-4793-bb66-a2066bf63ce5\" (UID: \"9f8e267b-7f34-4793-bb66-a2066bf63ce5\") " Nov 21 15:02:59 crc kubenswrapper[4897]: I1121 15:02:59.127042 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9f8e267b-7f34-4793-bb66-a2066bf63ce5-inventory\") pod \"9f8e267b-7f34-4793-bb66-a2066bf63ce5\" (UID: \"9f8e267b-7f34-4793-bb66-a2066bf63ce5\") " Nov 21 15:02:59 crc kubenswrapper[4897]: I1121 15:02:59.127159 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f8e267b-7f34-4793-bb66-a2066bf63ce5-libvirt-combined-ca-bundle\") pod \"9f8e267b-7f34-4793-bb66-a2066bf63ce5\" (UID: \"9f8e267b-7f34-4793-bb66-a2066bf63ce5\") " Nov 21 15:02:59 crc kubenswrapper[4897]: I1121 15:02:59.134425 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f8e267b-7f34-4793-bb66-a2066bf63ce5-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "9f8e267b-7f34-4793-bb66-a2066bf63ce5" (UID: "9f8e267b-7f34-4793-bb66-a2066bf63ce5"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 15:02:59 crc kubenswrapper[4897]: I1121 15:02:59.138440 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f8e267b-7f34-4793-bb66-a2066bf63ce5-kube-api-access-d2ft2" (OuterVolumeSpecName: "kube-api-access-d2ft2") pod "9f8e267b-7f34-4793-bb66-a2066bf63ce5" (UID: "9f8e267b-7f34-4793-bb66-a2066bf63ce5"). InnerVolumeSpecName "kube-api-access-d2ft2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 15:02:59 crc kubenswrapper[4897]: I1121 15:02:59.165363 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f8e267b-7f34-4793-bb66-a2066bf63ce5-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "9f8e267b-7f34-4793-bb66-a2066bf63ce5" (UID: "9f8e267b-7f34-4793-bb66-a2066bf63ce5"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 15:02:59 crc kubenswrapper[4897]: I1121 15:02:59.170730 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f8e267b-7f34-4793-bb66-a2066bf63ce5-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "9f8e267b-7f34-4793-bb66-a2066bf63ce5" (UID: "9f8e267b-7f34-4793-bb66-a2066bf63ce5"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 15:02:59 crc kubenswrapper[4897]: I1121 15:02:59.170846 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f8e267b-7f34-4793-bb66-a2066bf63ce5-inventory" (OuterVolumeSpecName: "inventory") pod "9f8e267b-7f34-4793-bb66-a2066bf63ce5" (UID: "9f8e267b-7f34-4793-bb66-a2066bf63ce5"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 15:02:59 crc kubenswrapper[4897]: I1121 15:02:59.230146 4897 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9f8e267b-7f34-4793-bb66-a2066bf63ce5-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 21 15:02:59 crc kubenswrapper[4897]: I1121 15:02:59.230175 4897 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/9f8e267b-7f34-4793-bb66-a2066bf63ce5-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Nov 21 15:02:59 crc kubenswrapper[4897]: I1121 15:02:59.230186 4897 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9f8e267b-7f34-4793-bb66-a2066bf63ce5-inventory\") on node \"crc\" DevicePath \"\"" Nov 21 15:02:59 crc kubenswrapper[4897]: I1121 15:02:59.230203 4897 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f8e267b-7f34-4793-bb66-a2066bf63ce5-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 21 15:02:59 crc kubenswrapper[4897]: I1121 15:02:59.230214 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d2ft2\" (UniqueName: \"kubernetes.io/projected/9f8e267b-7f34-4793-bb66-a2066bf63ce5-kube-api-access-d2ft2\") on node \"crc\" DevicePath \"\"" Nov 21 15:02:59 crc kubenswrapper[4897]: I1121 15:02:59.574762 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tmr6q" event={"ID":"9f8e267b-7f34-4793-bb66-a2066bf63ce5","Type":"ContainerDied","Data":"9100a89831ed8428be3f3d349c8730bf17aa5d8f02956e117651a49c120c78df"} Nov 21 15:02:59 crc kubenswrapper[4897]: I1121 15:02:59.575061 4897 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9100a89831ed8428be3f3d349c8730bf17aa5d8f02956e117651a49c120c78df" Nov 21 15:02:59 crc kubenswrapper[4897]: I1121 15:02:59.574828 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tmr6q" Nov 21 15:02:59 crc kubenswrapper[4897]: I1121 15:02:59.671574 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-8vsfm"] Nov 21 15:02:59 crc kubenswrapper[4897]: E1121 15:02:59.672018 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f8e267b-7f34-4793-bb66-a2066bf63ce5" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Nov 21 15:02:59 crc kubenswrapper[4897]: I1121 15:02:59.672037 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f8e267b-7f34-4793-bb66-a2066bf63ce5" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Nov 21 15:02:59 crc kubenswrapper[4897]: E1121 15:02:59.672053 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0453db98-4ad2-42d2-8781-947a3d03eb7c" containerName="registry-server" Nov 21 15:02:59 crc kubenswrapper[4897]: I1121 15:02:59.672060 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="0453db98-4ad2-42d2-8781-947a3d03eb7c" containerName="registry-server" Nov 21 15:02:59 crc kubenswrapper[4897]: E1121 15:02:59.672092 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0453db98-4ad2-42d2-8781-947a3d03eb7c" containerName="extract-content" Nov 21 15:02:59 crc kubenswrapper[4897]: I1121 15:02:59.672101 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="0453db98-4ad2-42d2-8781-947a3d03eb7c" containerName="extract-content" Nov 21 15:02:59 crc kubenswrapper[4897]: E1121 15:02:59.672138 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0453db98-4ad2-42d2-8781-947a3d03eb7c" containerName="extract-utilities" Nov 21 15:02:59 crc kubenswrapper[4897]: I1121 15:02:59.672145 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="0453db98-4ad2-42d2-8781-947a3d03eb7c" containerName="extract-utilities" Nov 21 15:02:59 crc kubenswrapper[4897]: I1121 15:02:59.672387 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f8e267b-7f34-4793-bb66-a2066bf63ce5" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Nov 21 15:02:59 crc kubenswrapper[4897]: I1121 15:02:59.672423 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="0453db98-4ad2-42d2-8781-947a3d03eb7c" containerName="registry-server" Nov 21 15:02:59 crc kubenswrapper[4897]: I1121 15:02:59.673416 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8vsfm" Nov 21 15:02:59 crc kubenswrapper[4897]: I1121 15:02:59.675711 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Nov 21 15:02:59 crc kubenswrapper[4897]: I1121 15:02:59.675744 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 21 15:02:59 crc kubenswrapper[4897]: I1121 15:02:59.675973 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Nov 21 15:02:59 crc kubenswrapper[4897]: I1121 15:02:59.676097 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-dj88m" Nov 21 15:02:59 crc kubenswrapper[4897]: I1121 15:02:59.676144 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Nov 21 15:02:59 crc kubenswrapper[4897]: I1121 15:02:59.676164 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 21 15:02:59 crc kubenswrapper[4897]: I1121 15:02:59.676276 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 21 15:02:59 crc kubenswrapper[4897]: I1121 15:02:59.686108 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-8vsfm"] Nov 21 15:02:59 crc kubenswrapper[4897]: I1121 15:02:59.843261 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/a65ca580-bb29-4224-bdb1-7178ca9f7a45-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8vsfm\" (UID: \"a65ca580-bb29-4224-bdb1-7178ca9f7a45\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8vsfm" Nov 21 15:02:59 crc kubenswrapper[4897]: I1121 15:02:59.843327 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a65ca580-bb29-4224-bdb1-7178ca9f7a45-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8vsfm\" (UID: \"a65ca580-bb29-4224-bdb1-7178ca9f7a45\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8vsfm" Nov 21 15:02:59 crc kubenswrapper[4897]: I1121 15:02:59.843391 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/a65ca580-bb29-4224-bdb1-7178ca9f7a45-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8vsfm\" (UID: \"a65ca580-bb29-4224-bdb1-7178ca9f7a45\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8vsfm" Nov 21 15:02:59 crc kubenswrapper[4897]: I1121 15:02:59.843413 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/a65ca580-bb29-4224-bdb1-7178ca9f7a45-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8vsfm\" (UID: \"a65ca580-bb29-4224-bdb1-7178ca9f7a45\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8vsfm" Nov 21 15:02:59 crc kubenswrapper[4897]: I1121 15:02:59.843756 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a65ca580-bb29-4224-bdb1-7178ca9f7a45-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8vsfm\" (UID: \"a65ca580-bb29-4224-bdb1-7178ca9f7a45\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8vsfm" Nov 21 15:02:59 crc kubenswrapper[4897]: I1121 15:02:59.843980 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lj8cc\" (UniqueName: \"kubernetes.io/projected/a65ca580-bb29-4224-bdb1-7178ca9f7a45-kube-api-access-lj8cc\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8vsfm\" (UID: \"a65ca580-bb29-4224-bdb1-7178ca9f7a45\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8vsfm" Nov 21 15:02:59 crc kubenswrapper[4897]: I1121 15:02:59.844083 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/a65ca580-bb29-4224-bdb1-7178ca9f7a45-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8vsfm\" (UID: \"a65ca580-bb29-4224-bdb1-7178ca9f7a45\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8vsfm" Nov 21 15:02:59 crc kubenswrapper[4897]: I1121 15:02:59.844126 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a65ca580-bb29-4224-bdb1-7178ca9f7a45-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8vsfm\" (UID: \"a65ca580-bb29-4224-bdb1-7178ca9f7a45\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8vsfm" Nov 21 15:02:59 crc kubenswrapper[4897]: I1121 15:02:59.844178 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/a65ca580-bb29-4224-bdb1-7178ca9f7a45-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8vsfm\" (UID: \"a65ca580-bb29-4224-bdb1-7178ca9f7a45\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8vsfm" Nov 21 15:02:59 crc kubenswrapper[4897]: I1121 15:02:59.946972 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lj8cc\" (UniqueName: \"kubernetes.io/projected/a65ca580-bb29-4224-bdb1-7178ca9f7a45-kube-api-access-lj8cc\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8vsfm\" (UID: \"a65ca580-bb29-4224-bdb1-7178ca9f7a45\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8vsfm" Nov 21 15:02:59 crc kubenswrapper[4897]: I1121 15:02:59.947033 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/a65ca580-bb29-4224-bdb1-7178ca9f7a45-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8vsfm\" (UID: \"a65ca580-bb29-4224-bdb1-7178ca9f7a45\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8vsfm" Nov 21 15:02:59 crc kubenswrapper[4897]: I1121 15:02:59.947057 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a65ca580-bb29-4224-bdb1-7178ca9f7a45-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8vsfm\" (UID: \"a65ca580-bb29-4224-bdb1-7178ca9f7a45\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8vsfm" Nov 21 15:02:59 crc kubenswrapper[4897]: I1121 15:02:59.947080 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/a65ca580-bb29-4224-bdb1-7178ca9f7a45-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8vsfm\" (UID: \"a65ca580-bb29-4224-bdb1-7178ca9f7a45\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8vsfm" Nov 21 15:02:59 crc kubenswrapper[4897]: I1121 15:02:59.947174 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/a65ca580-bb29-4224-bdb1-7178ca9f7a45-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8vsfm\" (UID: \"a65ca580-bb29-4224-bdb1-7178ca9f7a45\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8vsfm" Nov 21 15:02:59 crc kubenswrapper[4897]: I1121 15:02:59.947206 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a65ca580-bb29-4224-bdb1-7178ca9f7a45-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8vsfm\" (UID: \"a65ca580-bb29-4224-bdb1-7178ca9f7a45\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8vsfm" Nov 21 15:02:59 crc kubenswrapper[4897]: I1121 15:02:59.947246 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/a65ca580-bb29-4224-bdb1-7178ca9f7a45-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8vsfm\" (UID: \"a65ca580-bb29-4224-bdb1-7178ca9f7a45\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8vsfm" Nov 21 15:02:59 crc kubenswrapper[4897]: I1121 15:02:59.947270 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/a65ca580-bb29-4224-bdb1-7178ca9f7a45-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8vsfm\" (UID: \"a65ca580-bb29-4224-bdb1-7178ca9f7a45\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8vsfm" Nov 21 15:02:59 crc kubenswrapper[4897]: I1121 15:02:59.947363 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a65ca580-bb29-4224-bdb1-7178ca9f7a45-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8vsfm\" (UID: \"a65ca580-bb29-4224-bdb1-7178ca9f7a45\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8vsfm" Nov 21 15:02:59 crc kubenswrapper[4897]: I1121 15:02:59.949045 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/a65ca580-bb29-4224-bdb1-7178ca9f7a45-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8vsfm\" (UID: \"a65ca580-bb29-4224-bdb1-7178ca9f7a45\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8vsfm" Nov 21 15:02:59 crc kubenswrapper[4897]: I1121 15:02:59.953019 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a65ca580-bb29-4224-bdb1-7178ca9f7a45-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8vsfm\" (UID: \"a65ca580-bb29-4224-bdb1-7178ca9f7a45\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8vsfm" Nov 21 15:02:59 crc kubenswrapper[4897]: I1121 15:02:59.953824 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a65ca580-bb29-4224-bdb1-7178ca9f7a45-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8vsfm\" (UID: \"a65ca580-bb29-4224-bdb1-7178ca9f7a45\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8vsfm" Nov 21 15:02:59 crc kubenswrapper[4897]: I1121 15:02:59.953964 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/a65ca580-bb29-4224-bdb1-7178ca9f7a45-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8vsfm\" (UID: \"a65ca580-bb29-4224-bdb1-7178ca9f7a45\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8vsfm" Nov 21 15:02:59 crc kubenswrapper[4897]: I1121 15:02:59.954093 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/a65ca580-bb29-4224-bdb1-7178ca9f7a45-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8vsfm\" (UID: \"a65ca580-bb29-4224-bdb1-7178ca9f7a45\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8vsfm" Nov 21 15:02:59 crc kubenswrapper[4897]: I1121 15:02:59.955922 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/a65ca580-bb29-4224-bdb1-7178ca9f7a45-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8vsfm\" (UID: \"a65ca580-bb29-4224-bdb1-7178ca9f7a45\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8vsfm" Nov 21 15:02:59 crc kubenswrapper[4897]: I1121 15:02:59.961404 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a65ca580-bb29-4224-bdb1-7178ca9f7a45-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8vsfm\" (UID: \"a65ca580-bb29-4224-bdb1-7178ca9f7a45\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8vsfm" Nov 21 15:02:59 crc kubenswrapper[4897]: I1121 15:02:59.963703 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/a65ca580-bb29-4224-bdb1-7178ca9f7a45-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8vsfm\" (UID: \"a65ca580-bb29-4224-bdb1-7178ca9f7a45\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8vsfm" Nov 21 15:02:59 crc kubenswrapper[4897]: I1121 15:02:59.971814 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lj8cc\" (UniqueName: \"kubernetes.io/projected/a65ca580-bb29-4224-bdb1-7178ca9f7a45-kube-api-access-lj8cc\") pod \"nova-edpm-deployment-openstack-edpm-ipam-8vsfm\" (UID: \"a65ca580-bb29-4224-bdb1-7178ca9f7a45\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8vsfm" Nov 21 15:02:59 crc kubenswrapper[4897]: I1121 15:02:59.996230 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8vsfm" Nov 21 15:03:00 crc kubenswrapper[4897]: I1121 15:03:00.560426 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-8vsfm"] Nov 21 15:03:00 crc kubenswrapper[4897]: I1121 15:03:00.585904 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8vsfm" event={"ID":"a65ca580-bb29-4224-bdb1-7178ca9f7a45","Type":"ContainerStarted","Data":"347cff58a0060704026bc233089dbf7c3fb957adfb7b732f29e56a2ae78706ea"} Nov 21 15:03:02 crc kubenswrapper[4897]: I1121 15:03:02.611990 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8vsfm" event={"ID":"a65ca580-bb29-4224-bdb1-7178ca9f7a45","Type":"ContainerStarted","Data":"c0623976ad28d5049929c3e6e9145600949a2fb353def2c251860f2710acbd9e"} Nov 21 15:03:02 crc kubenswrapper[4897]: I1121 15:03:02.635389 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8vsfm" podStartSLOduration=2.655535795 podStartE2EDuration="3.635366139s" podCreationTimestamp="2025-11-21 15:02:59 +0000 UTC" firstStartedPulling="2025-11-21 15:03:00.570001093 +0000 UTC m=+3257.854594568" lastFinishedPulling="2025-11-21 15:03:01.549831437 +0000 UTC m=+3258.834424912" observedRunningTime="2025-11-21 15:03:02.630438184 +0000 UTC m=+3259.915031659" watchObservedRunningTime="2025-11-21 15:03:02.635366139 +0000 UTC m=+3259.919959614" Nov 21 15:03:11 crc kubenswrapper[4897]: I1121 15:03:11.089803 4897 scope.go:117] "RemoveContainer" containerID="a17b834dbb567d490c0f7ce8a22ab5aa6ecb645fc460c063af4ca43f58348563" Nov 21 15:03:11 crc kubenswrapper[4897]: I1121 15:03:11.707687 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" event={"ID":"e7670227-d280-4847-b882-754429f56b0f","Type":"ContainerStarted","Data":"11456ed300543798476a9b68095120ef6d6073c876d338f9212ec0e3865c7a1a"} Nov 21 15:05:30 crc kubenswrapper[4897]: I1121 15:05:30.027829 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-qrz6x"] Nov 21 15:05:30 crc kubenswrapper[4897]: I1121 15:05:30.031417 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qrz6x" Nov 21 15:05:30 crc kubenswrapper[4897]: I1121 15:05:30.042790 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qrz6x"] Nov 21 15:05:30 crc kubenswrapper[4897]: I1121 15:05:30.136654 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08353c71-61dc-46a2-80bf-a83f1207b572-catalog-content\") pod \"redhat-marketplace-qrz6x\" (UID: \"08353c71-61dc-46a2-80bf-a83f1207b572\") " pod="openshift-marketplace/redhat-marketplace-qrz6x" Nov 21 15:05:30 crc kubenswrapper[4897]: I1121 15:05:30.136737 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08353c71-61dc-46a2-80bf-a83f1207b572-utilities\") pod \"redhat-marketplace-qrz6x\" (UID: \"08353c71-61dc-46a2-80bf-a83f1207b572\") " pod="openshift-marketplace/redhat-marketplace-qrz6x" Nov 21 15:05:30 crc kubenswrapper[4897]: I1121 15:05:30.136978 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jls4j\" (UniqueName: \"kubernetes.io/projected/08353c71-61dc-46a2-80bf-a83f1207b572-kube-api-access-jls4j\") pod \"redhat-marketplace-qrz6x\" (UID: \"08353c71-61dc-46a2-80bf-a83f1207b572\") " pod="openshift-marketplace/redhat-marketplace-qrz6x" Nov 21 15:05:30 crc kubenswrapper[4897]: I1121 15:05:30.240285 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08353c71-61dc-46a2-80bf-a83f1207b572-catalog-content\") pod \"redhat-marketplace-qrz6x\" (UID: \"08353c71-61dc-46a2-80bf-a83f1207b572\") " pod="openshift-marketplace/redhat-marketplace-qrz6x" Nov 21 15:05:30 crc kubenswrapper[4897]: I1121 15:05:30.240369 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08353c71-61dc-46a2-80bf-a83f1207b572-utilities\") pod \"redhat-marketplace-qrz6x\" (UID: \"08353c71-61dc-46a2-80bf-a83f1207b572\") " pod="openshift-marketplace/redhat-marketplace-qrz6x" Nov 21 15:05:30 crc kubenswrapper[4897]: I1121 15:05:30.240628 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jls4j\" (UniqueName: \"kubernetes.io/projected/08353c71-61dc-46a2-80bf-a83f1207b572-kube-api-access-jls4j\") pod \"redhat-marketplace-qrz6x\" (UID: \"08353c71-61dc-46a2-80bf-a83f1207b572\") " pod="openshift-marketplace/redhat-marketplace-qrz6x" Nov 21 15:05:30 crc kubenswrapper[4897]: I1121 15:05:30.240921 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08353c71-61dc-46a2-80bf-a83f1207b572-catalog-content\") pod \"redhat-marketplace-qrz6x\" (UID: \"08353c71-61dc-46a2-80bf-a83f1207b572\") " pod="openshift-marketplace/redhat-marketplace-qrz6x" Nov 21 15:05:30 crc kubenswrapper[4897]: I1121 15:05:30.240984 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08353c71-61dc-46a2-80bf-a83f1207b572-utilities\") pod \"redhat-marketplace-qrz6x\" (UID: \"08353c71-61dc-46a2-80bf-a83f1207b572\") " pod="openshift-marketplace/redhat-marketplace-qrz6x" Nov 21 15:05:30 crc kubenswrapper[4897]: I1121 15:05:30.270830 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jls4j\" (UniqueName: \"kubernetes.io/projected/08353c71-61dc-46a2-80bf-a83f1207b572-kube-api-access-jls4j\") pod \"redhat-marketplace-qrz6x\" (UID: \"08353c71-61dc-46a2-80bf-a83f1207b572\") " pod="openshift-marketplace/redhat-marketplace-qrz6x" Nov 21 15:05:30 crc kubenswrapper[4897]: I1121 15:05:30.361887 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qrz6x" Nov 21 15:05:30 crc kubenswrapper[4897]: I1121 15:05:30.828826 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qrz6x"] Nov 21 15:05:31 crc kubenswrapper[4897]: I1121 15:05:31.267301 4897 generic.go:334] "Generic (PLEG): container finished" podID="08353c71-61dc-46a2-80bf-a83f1207b572" containerID="c9d2bd65d356c360cdc3ddc49dd7617d15a52ed1e1fd268d5c41e24c761d792a" exitCode=0 Nov 21 15:05:31 crc kubenswrapper[4897]: I1121 15:05:31.267361 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qrz6x" event={"ID":"08353c71-61dc-46a2-80bf-a83f1207b572","Type":"ContainerDied","Data":"c9d2bd65d356c360cdc3ddc49dd7617d15a52ed1e1fd268d5c41e24c761d792a"} Nov 21 15:05:31 crc kubenswrapper[4897]: I1121 15:05:31.267444 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qrz6x" event={"ID":"08353c71-61dc-46a2-80bf-a83f1207b572","Type":"ContainerStarted","Data":"0192efca1c1d5cf0d9499717278712510101039dff8e1f2b12738b472facfd23"} Nov 21 15:05:32 crc kubenswrapper[4897]: I1121 15:05:32.280590 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qrz6x" event={"ID":"08353c71-61dc-46a2-80bf-a83f1207b572","Type":"ContainerStarted","Data":"8ce575b2df8d5dede791e4752aa3f413a1ace83bc22ed9e4c634ba2812fa0980"} Nov 21 15:05:32 crc kubenswrapper[4897]: I1121 15:05:32.424197 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-r5gz8"] Nov 21 15:05:32 crc kubenswrapper[4897]: I1121 15:05:32.427183 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r5gz8" Nov 21 15:05:32 crc kubenswrapper[4897]: I1121 15:05:32.436594 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-r5gz8"] Nov 21 15:05:32 crc kubenswrapper[4897]: I1121 15:05:32.506422 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vwc7z\" (UniqueName: \"kubernetes.io/projected/e5469f78-a615-45fe-8a8a-30d76071e075-kube-api-access-vwc7z\") pod \"certified-operators-r5gz8\" (UID: \"e5469f78-a615-45fe-8a8a-30d76071e075\") " pod="openshift-marketplace/certified-operators-r5gz8" Nov 21 15:05:32 crc kubenswrapper[4897]: I1121 15:05:32.506650 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5469f78-a615-45fe-8a8a-30d76071e075-catalog-content\") pod \"certified-operators-r5gz8\" (UID: \"e5469f78-a615-45fe-8a8a-30d76071e075\") " pod="openshift-marketplace/certified-operators-r5gz8" Nov 21 15:05:32 crc kubenswrapper[4897]: I1121 15:05:32.506716 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5469f78-a615-45fe-8a8a-30d76071e075-utilities\") pod \"certified-operators-r5gz8\" (UID: \"e5469f78-a615-45fe-8a8a-30d76071e075\") " pod="openshift-marketplace/certified-operators-r5gz8" Nov 21 15:05:32 crc kubenswrapper[4897]: I1121 15:05:32.609138 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5469f78-a615-45fe-8a8a-30d76071e075-catalog-content\") pod \"certified-operators-r5gz8\" (UID: \"e5469f78-a615-45fe-8a8a-30d76071e075\") " pod="openshift-marketplace/certified-operators-r5gz8" Nov 21 15:05:32 crc kubenswrapper[4897]: I1121 15:05:32.609223 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5469f78-a615-45fe-8a8a-30d76071e075-utilities\") pod \"certified-operators-r5gz8\" (UID: \"e5469f78-a615-45fe-8a8a-30d76071e075\") " pod="openshift-marketplace/certified-operators-r5gz8" Nov 21 15:05:32 crc kubenswrapper[4897]: I1121 15:05:32.609401 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vwc7z\" (UniqueName: \"kubernetes.io/projected/e5469f78-a615-45fe-8a8a-30d76071e075-kube-api-access-vwc7z\") pod \"certified-operators-r5gz8\" (UID: \"e5469f78-a615-45fe-8a8a-30d76071e075\") " pod="openshift-marketplace/certified-operators-r5gz8" Nov 21 15:05:32 crc kubenswrapper[4897]: I1121 15:05:32.610278 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5469f78-a615-45fe-8a8a-30d76071e075-catalog-content\") pod \"certified-operators-r5gz8\" (UID: \"e5469f78-a615-45fe-8a8a-30d76071e075\") " pod="openshift-marketplace/certified-operators-r5gz8" Nov 21 15:05:32 crc kubenswrapper[4897]: I1121 15:05:32.610491 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5469f78-a615-45fe-8a8a-30d76071e075-utilities\") pod \"certified-operators-r5gz8\" (UID: \"e5469f78-a615-45fe-8a8a-30d76071e075\") " pod="openshift-marketplace/certified-operators-r5gz8" Nov 21 15:05:32 crc kubenswrapper[4897]: I1121 15:05:32.631460 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vwc7z\" (UniqueName: \"kubernetes.io/projected/e5469f78-a615-45fe-8a8a-30d76071e075-kube-api-access-vwc7z\") pod \"certified-operators-r5gz8\" (UID: \"e5469f78-a615-45fe-8a8a-30d76071e075\") " pod="openshift-marketplace/certified-operators-r5gz8" Nov 21 15:05:32 crc kubenswrapper[4897]: I1121 15:05:32.755972 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r5gz8" Nov 21 15:05:33 crc kubenswrapper[4897]: I1121 15:05:33.256182 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-r5gz8"] Nov 21 15:05:33 crc kubenswrapper[4897]: I1121 15:05:33.293120 4897 generic.go:334] "Generic (PLEG): container finished" podID="08353c71-61dc-46a2-80bf-a83f1207b572" containerID="8ce575b2df8d5dede791e4752aa3f413a1ace83bc22ed9e4c634ba2812fa0980" exitCode=0 Nov 21 15:05:33 crc kubenswrapper[4897]: I1121 15:05:33.293206 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qrz6x" event={"ID":"08353c71-61dc-46a2-80bf-a83f1207b572","Type":"ContainerDied","Data":"8ce575b2df8d5dede791e4752aa3f413a1ace83bc22ed9e4c634ba2812fa0980"} Nov 21 15:05:33 crc kubenswrapper[4897]: I1121 15:05:33.297006 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r5gz8" event={"ID":"e5469f78-a615-45fe-8a8a-30d76071e075","Type":"ContainerStarted","Data":"661a228233e3cd42da4683cee3efba6ce5943bb9830b471bef97bf2fa763e338"} Nov 21 15:05:34 crc kubenswrapper[4897]: I1121 15:05:34.310372 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qrz6x" event={"ID":"08353c71-61dc-46a2-80bf-a83f1207b572","Type":"ContainerStarted","Data":"e1d537b48ade5ff5ef307ca346d6c470cc2896e1f6e14a574c3469e451390ff9"} Nov 21 15:05:34 crc kubenswrapper[4897]: I1121 15:05:34.313956 4897 generic.go:334] "Generic (PLEG): container finished" podID="e5469f78-a615-45fe-8a8a-30d76071e075" containerID="01075a1a2c71d0866293b5815871c389c8357d07f8134e7a17b8f20e23858c90" exitCode=0 Nov 21 15:05:34 crc kubenswrapper[4897]: I1121 15:05:34.314025 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r5gz8" event={"ID":"e5469f78-a615-45fe-8a8a-30d76071e075","Type":"ContainerDied","Data":"01075a1a2c71d0866293b5815871c389c8357d07f8134e7a17b8f20e23858c90"} Nov 21 15:05:34 crc kubenswrapper[4897]: I1121 15:05:34.337683 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-qrz6x" podStartSLOduration=1.7775846849999999 podStartE2EDuration="4.337664024s" podCreationTimestamp="2025-11-21 15:05:30 +0000 UTC" firstStartedPulling="2025-11-21 15:05:31.270411723 +0000 UTC m=+3408.555005198" lastFinishedPulling="2025-11-21 15:05:33.830491062 +0000 UTC m=+3411.115084537" observedRunningTime="2025-11-21 15:05:34.331446315 +0000 UTC m=+3411.616039790" watchObservedRunningTime="2025-11-21 15:05:34.337664024 +0000 UTC m=+3411.622257499" Nov 21 15:05:34 crc kubenswrapper[4897]: I1121 15:05:34.370569 4897 patch_prober.go:28] interesting pod/machine-config-daemon-krv5b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 21 15:05:34 crc kubenswrapper[4897]: I1121 15:05:34.370656 4897 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 21 15:05:35 crc kubenswrapper[4897]: I1121 15:05:35.333087 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r5gz8" event={"ID":"e5469f78-a615-45fe-8a8a-30d76071e075","Type":"ContainerStarted","Data":"213f9c6fecee0eedd09c5b056725deb68fcefb32f1b40124a6f57cedfe1b577a"} Nov 21 15:05:38 crc kubenswrapper[4897]: I1121 15:05:38.364493 4897 generic.go:334] "Generic (PLEG): container finished" podID="e5469f78-a615-45fe-8a8a-30d76071e075" containerID="213f9c6fecee0eedd09c5b056725deb68fcefb32f1b40124a6f57cedfe1b577a" exitCode=0 Nov 21 15:05:38 crc kubenswrapper[4897]: I1121 15:05:38.364538 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r5gz8" event={"ID":"e5469f78-a615-45fe-8a8a-30d76071e075","Type":"ContainerDied","Data":"213f9c6fecee0eedd09c5b056725deb68fcefb32f1b40124a6f57cedfe1b577a"} Nov 21 15:05:39 crc kubenswrapper[4897]: I1121 15:05:39.380962 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r5gz8" event={"ID":"e5469f78-a615-45fe-8a8a-30d76071e075","Type":"ContainerStarted","Data":"d5cdfd1a1f79bc8d39f42880163e4df7b54a12630b79b887db2190bc25d69cc0"} Nov 21 15:05:39 crc kubenswrapper[4897]: I1121 15:05:39.407314 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-r5gz8" podStartSLOduration=2.896716988 podStartE2EDuration="7.407294207s" podCreationTimestamp="2025-11-21 15:05:32 +0000 UTC" firstStartedPulling="2025-11-21 15:05:34.317217598 +0000 UTC m=+3411.601811073" lastFinishedPulling="2025-11-21 15:05:38.827794817 +0000 UTC m=+3416.112388292" observedRunningTime="2025-11-21 15:05:39.40296978 +0000 UTC m=+3416.687563255" watchObservedRunningTime="2025-11-21 15:05:39.407294207 +0000 UTC m=+3416.691887682" Nov 21 15:05:40 crc kubenswrapper[4897]: I1121 15:05:40.362812 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-qrz6x" Nov 21 15:05:40 crc kubenswrapper[4897]: I1121 15:05:40.362892 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-qrz6x" Nov 21 15:05:40 crc kubenswrapper[4897]: I1121 15:05:40.428087 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-qrz6x" Nov 21 15:05:40 crc kubenswrapper[4897]: I1121 15:05:40.482310 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-qrz6x" Nov 21 15:05:41 crc kubenswrapper[4897]: I1121 15:05:41.417426 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qrz6x"] Nov 21 15:05:42 crc kubenswrapper[4897]: I1121 15:05:42.414336 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-qrz6x" podUID="08353c71-61dc-46a2-80bf-a83f1207b572" containerName="registry-server" containerID="cri-o://e1d537b48ade5ff5ef307ca346d6c470cc2896e1f6e14a574c3469e451390ff9" gracePeriod=2 Nov 21 15:05:42 crc kubenswrapper[4897]: I1121 15:05:42.756845 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-r5gz8" Nov 21 15:05:42 crc kubenswrapper[4897]: I1121 15:05:42.757022 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-r5gz8" Nov 21 15:05:42 crc kubenswrapper[4897]: I1121 15:05:42.811270 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-r5gz8" Nov 21 15:05:42 crc kubenswrapper[4897]: I1121 15:05:42.921647 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qrz6x" Nov 21 15:05:43 crc kubenswrapper[4897]: I1121 15:05:43.090379 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08353c71-61dc-46a2-80bf-a83f1207b572-catalog-content\") pod \"08353c71-61dc-46a2-80bf-a83f1207b572\" (UID: \"08353c71-61dc-46a2-80bf-a83f1207b572\") " Nov 21 15:05:43 crc kubenswrapper[4897]: I1121 15:05:43.090889 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08353c71-61dc-46a2-80bf-a83f1207b572-utilities\") pod \"08353c71-61dc-46a2-80bf-a83f1207b572\" (UID: \"08353c71-61dc-46a2-80bf-a83f1207b572\") " Nov 21 15:05:43 crc kubenswrapper[4897]: I1121 15:05:43.091469 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/08353c71-61dc-46a2-80bf-a83f1207b572-utilities" (OuterVolumeSpecName: "utilities") pod "08353c71-61dc-46a2-80bf-a83f1207b572" (UID: "08353c71-61dc-46a2-80bf-a83f1207b572"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 15:05:43 crc kubenswrapper[4897]: I1121 15:05:43.091548 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jls4j\" (UniqueName: \"kubernetes.io/projected/08353c71-61dc-46a2-80bf-a83f1207b572-kube-api-access-jls4j\") pod \"08353c71-61dc-46a2-80bf-a83f1207b572\" (UID: \"08353c71-61dc-46a2-80bf-a83f1207b572\") " Nov 21 15:05:43 crc kubenswrapper[4897]: I1121 15:05:43.094377 4897 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08353c71-61dc-46a2-80bf-a83f1207b572-utilities\") on node \"crc\" DevicePath \"\"" Nov 21 15:05:43 crc kubenswrapper[4897]: I1121 15:05:43.098240 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08353c71-61dc-46a2-80bf-a83f1207b572-kube-api-access-jls4j" (OuterVolumeSpecName: "kube-api-access-jls4j") pod "08353c71-61dc-46a2-80bf-a83f1207b572" (UID: "08353c71-61dc-46a2-80bf-a83f1207b572"). InnerVolumeSpecName "kube-api-access-jls4j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 15:05:43 crc kubenswrapper[4897]: I1121 15:05:43.106552 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/08353c71-61dc-46a2-80bf-a83f1207b572-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "08353c71-61dc-46a2-80bf-a83f1207b572" (UID: "08353c71-61dc-46a2-80bf-a83f1207b572"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 15:05:43 crc kubenswrapper[4897]: I1121 15:05:43.197723 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jls4j\" (UniqueName: \"kubernetes.io/projected/08353c71-61dc-46a2-80bf-a83f1207b572-kube-api-access-jls4j\") on node \"crc\" DevicePath \"\"" Nov 21 15:05:43 crc kubenswrapper[4897]: I1121 15:05:43.198085 4897 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08353c71-61dc-46a2-80bf-a83f1207b572-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 21 15:05:43 crc kubenswrapper[4897]: I1121 15:05:43.428959 4897 generic.go:334] "Generic (PLEG): container finished" podID="08353c71-61dc-46a2-80bf-a83f1207b572" containerID="e1d537b48ade5ff5ef307ca346d6c470cc2896e1f6e14a574c3469e451390ff9" exitCode=0 Nov 21 15:05:43 crc kubenswrapper[4897]: I1121 15:05:43.430147 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qrz6x" Nov 21 15:05:43 crc kubenswrapper[4897]: I1121 15:05:43.434694 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qrz6x" event={"ID":"08353c71-61dc-46a2-80bf-a83f1207b572","Type":"ContainerDied","Data":"e1d537b48ade5ff5ef307ca346d6c470cc2896e1f6e14a574c3469e451390ff9"} Nov 21 15:05:43 crc kubenswrapper[4897]: I1121 15:05:43.434760 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qrz6x" event={"ID":"08353c71-61dc-46a2-80bf-a83f1207b572","Type":"ContainerDied","Data":"0192efca1c1d5cf0d9499717278712510101039dff8e1f2b12738b472facfd23"} Nov 21 15:05:43 crc kubenswrapper[4897]: I1121 15:05:43.434781 4897 scope.go:117] "RemoveContainer" containerID="e1d537b48ade5ff5ef307ca346d6c470cc2896e1f6e14a574c3469e451390ff9" Nov 21 15:05:43 crc kubenswrapper[4897]: I1121 15:05:43.476201 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qrz6x"] Nov 21 15:05:43 crc kubenswrapper[4897]: I1121 15:05:43.483337 4897 scope.go:117] "RemoveContainer" containerID="8ce575b2df8d5dede791e4752aa3f413a1ace83bc22ed9e4c634ba2812fa0980" Nov 21 15:05:43 crc kubenswrapper[4897]: I1121 15:05:43.498586 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-qrz6x"] Nov 21 15:05:43 crc kubenswrapper[4897]: I1121 15:05:43.517073 4897 scope.go:117] "RemoveContainer" containerID="c9d2bd65d356c360cdc3ddc49dd7617d15a52ed1e1fd268d5c41e24c761d792a" Nov 21 15:05:43 crc kubenswrapper[4897]: I1121 15:05:43.597378 4897 scope.go:117] "RemoveContainer" containerID="e1d537b48ade5ff5ef307ca346d6c470cc2896e1f6e14a574c3469e451390ff9" Nov 21 15:05:43 crc kubenswrapper[4897]: E1121 15:05:43.598528 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e1d537b48ade5ff5ef307ca346d6c470cc2896e1f6e14a574c3469e451390ff9\": container with ID starting with e1d537b48ade5ff5ef307ca346d6c470cc2896e1f6e14a574c3469e451390ff9 not found: ID does not exist" containerID="e1d537b48ade5ff5ef307ca346d6c470cc2896e1f6e14a574c3469e451390ff9" Nov 21 15:05:43 crc kubenswrapper[4897]: I1121 15:05:43.598572 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1d537b48ade5ff5ef307ca346d6c470cc2896e1f6e14a574c3469e451390ff9"} err="failed to get container status \"e1d537b48ade5ff5ef307ca346d6c470cc2896e1f6e14a574c3469e451390ff9\": rpc error: code = NotFound desc = could not find container \"e1d537b48ade5ff5ef307ca346d6c470cc2896e1f6e14a574c3469e451390ff9\": container with ID starting with e1d537b48ade5ff5ef307ca346d6c470cc2896e1f6e14a574c3469e451390ff9 not found: ID does not exist" Nov 21 15:05:43 crc kubenswrapper[4897]: I1121 15:05:43.598603 4897 scope.go:117] "RemoveContainer" containerID="8ce575b2df8d5dede791e4752aa3f413a1ace83bc22ed9e4c634ba2812fa0980" Nov 21 15:05:43 crc kubenswrapper[4897]: E1121 15:05:43.598914 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8ce575b2df8d5dede791e4752aa3f413a1ace83bc22ed9e4c634ba2812fa0980\": container with ID starting with 8ce575b2df8d5dede791e4752aa3f413a1ace83bc22ed9e4c634ba2812fa0980 not found: ID does not exist" containerID="8ce575b2df8d5dede791e4752aa3f413a1ace83bc22ed9e4c634ba2812fa0980" Nov 21 15:05:43 crc kubenswrapper[4897]: I1121 15:05:43.598945 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8ce575b2df8d5dede791e4752aa3f413a1ace83bc22ed9e4c634ba2812fa0980"} err="failed to get container status \"8ce575b2df8d5dede791e4752aa3f413a1ace83bc22ed9e4c634ba2812fa0980\": rpc error: code = NotFound desc = could not find container \"8ce575b2df8d5dede791e4752aa3f413a1ace83bc22ed9e4c634ba2812fa0980\": container with ID starting with 8ce575b2df8d5dede791e4752aa3f413a1ace83bc22ed9e4c634ba2812fa0980 not found: ID does not exist" Nov 21 15:05:43 crc kubenswrapper[4897]: I1121 15:05:43.598962 4897 scope.go:117] "RemoveContainer" containerID="c9d2bd65d356c360cdc3ddc49dd7617d15a52ed1e1fd268d5c41e24c761d792a" Nov 21 15:05:43 crc kubenswrapper[4897]: E1121 15:05:43.599242 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c9d2bd65d356c360cdc3ddc49dd7617d15a52ed1e1fd268d5c41e24c761d792a\": container with ID starting with c9d2bd65d356c360cdc3ddc49dd7617d15a52ed1e1fd268d5c41e24c761d792a not found: ID does not exist" containerID="c9d2bd65d356c360cdc3ddc49dd7617d15a52ed1e1fd268d5c41e24c761d792a" Nov 21 15:05:43 crc kubenswrapper[4897]: I1121 15:05:43.599272 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c9d2bd65d356c360cdc3ddc49dd7617d15a52ed1e1fd268d5c41e24c761d792a"} err="failed to get container status \"c9d2bd65d356c360cdc3ddc49dd7617d15a52ed1e1fd268d5c41e24c761d792a\": rpc error: code = NotFound desc = could not find container \"c9d2bd65d356c360cdc3ddc49dd7617d15a52ed1e1fd268d5c41e24c761d792a\": container with ID starting with c9d2bd65d356c360cdc3ddc49dd7617d15a52ed1e1fd268d5c41e24c761d792a not found: ID does not exist" Nov 21 15:05:44 crc kubenswrapper[4897]: I1121 15:05:44.108584 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="08353c71-61dc-46a2-80bf-a83f1207b572" path="/var/lib/kubelet/pods/08353c71-61dc-46a2-80bf-a83f1207b572/volumes" Nov 21 15:05:52 crc kubenswrapper[4897]: I1121 15:05:52.816812 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-r5gz8" Nov 21 15:05:52 crc kubenswrapper[4897]: I1121 15:05:52.867019 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-r5gz8"] Nov 21 15:05:53 crc kubenswrapper[4897]: I1121 15:05:53.538477 4897 generic.go:334] "Generic (PLEG): container finished" podID="a65ca580-bb29-4224-bdb1-7178ca9f7a45" containerID="c0623976ad28d5049929c3e6e9145600949a2fb353def2c251860f2710acbd9e" exitCode=0 Nov 21 15:05:53 crc kubenswrapper[4897]: I1121 15:05:53.538719 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-r5gz8" podUID="e5469f78-a615-45fe-8a8a-30d76071e075" containerName="registry-server" containerID="cri-o://d5cdfd1a1f79bc8d39f42880163e4df7b54a12630b79b887db2190bc25d69cc0" gracePeriod=2 Nov 21 15:05:53 crc kubenswrapper[4897]: I1121 15:05:53.538831 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8vsfm" event={"ID":"a65ca580-bb29-4224-bdb1-7178ca9f7a45","Type":"ContainerDied","Data":"c0623976ad28d5049929c3e6e9145600949a2fb353def2c251860f2710acbd9e"} Nov 21 15:05:54 crc kubenswrapper[4897]: I1121 15:05:54.088543 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r5gz8" Nov 21 15:05:54 crc kubenswrapper[4897]: I1121 15:05:54.169067 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5469f78-a615-45fe-8a8a-30d76071e075-catalog-content\") pod \"e5469f78-a615-45fe-8a8a-30d76071e075\" (UID: \"e5469f78-a615-45fe-8a8a-30d76071e075\") " Nov 21 15:05:54 crc kubenswrapper[4897]: I1121 15:05:54.169149 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vwc7z\" (UniqueName: \"kubernetes.io/projected/e5469f78-a615-45fe-8a8a-30d76071e075-kube-api-access-vwc7z\") pod \"e5469f78-a615-45fe-8a8a-30d76071e075\" (UID: \"e5469f78-a615-45fe-8a8a-30d76071e075\") " Nov 21 15:05:54 crc kubenswrapper[4897]: I1121 15:05:54.169231 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5469f78-a615-45fe-8a8a-30d76071e075-utilities\") pod \"e5469f78-a615-45fe-8a8a-30d76071e075\" (UID: \"e5469f78-a615-45fe-8a8a-30d76071e075\") " Nov 21 15:05:54 crc kubenswrapper[4897]: I1121 15:05:54.170971 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e5469f78-a615-45fe-8a8a-30d76071e075-utilities" (OuterVolumeSpecName: "utilities") pod "e5469f78-a615-45fe-8a8a-30d76071e075" (UID: "e5469f78-a615-45fe-8a8a-30d76071e075"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 15:05:54 crc kubenswrapper[4897]: I1121 15:05:54.177610 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5469f78-a615-45fe-8a8a-30d76071e075-kube-api-access-vwc7z" (OuterVolumeSpecName: "kube-api-access-vwc7z") pod "e5469f78-a615-45fe-8a8a-30d76071e075" (UID: "e5469f78-a615-45fe-8a8a-30d76071e075"). InnerVolumeSpecName "kube-api-access-vwc7z". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 15:05:54 crc kubenswrapper[4897]: I1121 15:05:54.219474 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e5469f78-a615-45fe-8a8a-30d76071e075-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e5469f78-a615-45fe-8a8a-30d76071e075" (UID: "e5469f78-a615-45fe-8a8a-30d76071e075"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 15:05:54 crc kubenswrapper[4897]: I1121 15:05:54.273030 4897 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5469f78-a615-45fe-8a8a-30d76071e075-utilities\") on node \"crc\" DevicePath \"\"" Nov 21 15:05:54 crc kubenswrapper[4897]: I1121 15:05:54.273074 4897 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5469f78-a615-45fe-8a8a-30d76071e075-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 21 15:05:54 crc kubenswrapper[4897]: I1121 15:05:54.273086 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vwc7z\" (UniqueName: \"kubernetes.io/projected/e5469f78-a615-45fe-8a8a-30d76071e075-kube-api-access-vwc7z\") on node \"crc\" DevicePath \"\"" Nov 21 15:05:54 crc kubenswrapper[4897]: I1121 15:05:54.561122 4897 generic.go:334] "Generic (PLEG): container finished" podID="e5469f78-a615-45fe-8a8a-30d76071e075" containerID="d5cdfd1a1f79bc8d39f42880163e4df7b54a12630b79b887db2190bc25d69cc0" exitCode=0 Nov 21 15:05:54 crc kubenswrapper[4897]: I1121 15:05:54.561177 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r5gz8" event={"ID":"e5469f78-a615-45fe-8a8a-30d76071e075","Type":"ContainerDied","Data":"d5cdfd1a1f79bc8d39f42880163e4df7b54a12630b79b887db2190bc25d69cc0"} Nov 21 15:05:54 crc kubenswrapper[4897]: I1121 15:05:54.561218 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r5gz8" Nov 21 15:05:54 crc kubenswrapper[4897]: I1121 15:05:54.561239 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r5gz8" event={"ID":"e5469f78-a615-45fe-8a8a-30d76071e075","Type":"ContainerDied","Data":"661a228233e3cd42da4683cee3efba6ce5943bb9830b471bef97bf2fa763e338"} Nov 21 15:05:54 crc kubenswrapper[4897]: I1121 15:05:54.561265 4897 scope.go:117] "RemoveContainer" containerID="d5cdfd1a1f79bc8d39f42880163e4df7b54a12630b79b887db2190bc25d69cc0" Nov 21 15:05:54 crc kubenswrapper[4897]: I1121 15:05:54.600728 4897 scope.go:117] "RemoveContainer" containerID="213f9c6fecee0eedd09c5b056725deb68fcefb32f1b40124a6f57cedfe1b577a" Nov 21 15:05:54 crc kubenswrapper[4897]: I1121 15:05:54.634564 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-r5gz8"] Nov 21 15:05:54 crc kubenswrapper[4897]: I1121 15:05:54.653562 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-r5gz8"] Nov 21 15:05:54 crc kubenswrapper[4897]: I1121 15:05:54.698558 4897 scope.go:117] "RemoveContainer" containerID="01075a1a2c71d0866293b5815871c389c8357d07f8134e7a17b8f20e23858c90" Nov 21 15:05:54 crc kubenswrapper[4897]: I1121 15:05:54.782995 4897 scope.go:117] "RemoveContainer" containerID="d5cdfd1a1f79bc8d39f42880163e4df7b54a12630b79b887db2190bc25d69cc0" Nov 21 15:05:54 crc kubenswrapper[4897]: E1121 15:05:54.787947 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d5cdfd1a1f79bc8d39f42880163e4df7b54a12630b79b887db2190bc25d69cc0\": container with ID starting with d5cdfd1a1f79bc8d39f42880163e4df7b54a12630b79b887db2190bc25d69cc0 not found: ID does not exist" containerID="d5cdfd1a1f79bc8d39f42880163e4df7b54a12630b79b887db2190bc25d69cc0" Nov 21 15:05:54 crc kubenswrapper[4897]: I1121 15:05:54.788001 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d5cdfd1a1f79bc8d39f42880163e4df7b54a12630b79b887db2190bc25d69cc0"} err="failed to get container status \"d5cdfd1a1f79bc8d39f42880163e4df7b54a12630b79b887db2190bc25d69cc0\": rpc error: code = NotFound desc = could not find container \"d5cdfd1a1f79bc8d39f42880163e4df7b54a12630b79b887db2190bc25d69cc0\": container with ID starting with d5cdfd1a1f79bc8d39f42880163e4df7b54a12630b79b887db2190bc25d69cc0 not found: ID does not exist" Nov 21 15:05:54 crc kubenswrapper[4897]: I1121 15:05:54.788035 4897 scope.go:117] "RemoveContainer" containerID="213f9c6fecee0eedd09c5b056725deb68fcefb32f1b40124a6f57cedfe1b577a" Nov 21 15:05:54 crc kubenswrapper[4897]: E1121 15:05:54.791017 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"213f9c6fecee0eedd09c5b056725deb68fcefb32f1b40124a6f57cedfe1b577a\": container with ID starting with 213f9c6fecee0eedd09c5b056725deb68fcefb32f1b40124a6f57cedfe1b577a not found: ID does not exist" containerID="213f9c6fecee0eedd09c5b056725deb68fcefb32f1b40124a6f57cedfe1b577a" Nov 21 15:05:54 crc kubenswrapper[4897]: I1121 15:05:54.791063 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"213f9c6fecee0eedd09c5b056725deb68fcefb32f1b40124a6f57cedfe1b577a"} err="failed to get container status \"213f9c6fecee0eedd09c5b056725deb68fcefb32f1b40124a6f57cedfe1b577a\": rpc error: code = NotFound desc = could not find container \"213f9c6fecee0eedd09c5b056725deb68fcefb32f1b40124a6f57cedfe1b577a\": container with ID starting with 213f9c6fecee0eedd09c5b056725deb68fcefb32f1b40124a6f57cedfe1b577a not found: ID does not exist" Nov 21 15:05:54 crc kubenswrapper[4897]: I1121 15:05:54.791098 4897 scope.go:117] "RemoveContainer" containerID="01075a1a2c71d0866293b5815871c389c8357d07f8134e7a17b8f20e23858c90" Nov 21 15:05:54 crc kubenswrapper[4897]: E1121 15:05:54.795677 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"01075a1a2c71d0866293b5815871c389c8357d07f8134e7a17b8f20e23858c90\": container with ID starting with 01075a1a2c71d0866293b5815871c389c8357d07f8134e7a17b8f20e23858c90 not found: ID does not exist" containerID="01075a1a2c71d0866293b5815871c389c8357d07f8134e7a17b8f20e23858c90" Nov 21 15:05:54 crc kubenswrapper[4897]: I1121 15:05:54.795727 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01075a1a2c71d0866293b5815871c389c8357d07f8134e7a17b8f20e23858c90"} err="failed to get container status \"01075a1a2c71d0866293b5815871c389c8357d07f8134e7a17b8f20e23858c90\": rpc error: code = NotFound desc = could not find container \"01075a1a2c71d0866293b5815871c389c8357d07f8134e7a17b8f20e23858c90\": container with ID starting with 01075a1a2c71d0866293b5815871c389c8357d07f8134e7a17b8f20e23858c90 not found: ID does not exist" Nov 21 15:05:55 crc kubenswrapper[4897]: I1121 15:05:55.166210 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8vsfm" Nov 21 15:05:55 crc kubenswrapper[4897]: I1121 15:05:55.299567 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/a65ca580-bb29-4224-bdb1-7178ca9f7a45-nova-cell1-compute-config-1\") pod \"a65ca580-bb29-4224-bdb1-7178ca9f7a45\" (UID: \"a65ca580-bb29-4224-bdb1-7178ca9f7a45\") " Nov 21 15:05:55 crc kubenswrapper[4897]: I1121 15:05:55.299688 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/a65ca580-bb29-4224-bdb1-7178ca9f7a45-nova-migration-ssh-key-1\") pod \"a65ca580-bb29-4224-bdb1-7178ca9f7a45\" (UID: \"a65ca580-bb29-4224-bdb1-7178ca9f7a45\") " Nov 21 15:05:55 crc kubenswrapper[4897]: I1121 15:05:55.299842 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a65ca580-bb29-4224-bdb1-7178ca9f7a45-nova-combined-ca-bundle\") pod \"a65ca580-bb29-4224-bdb1-7178ca9f7a45\" (UID: \"a65ca580-bb29-4224-bdb1-7178ca9f7a45\") " Nov 21 15:05:55 crc kubenswrapper[4897]: I1121 15:05:55.299913 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lj8cc\" (UniqueName: \"kubernetes.io/projected/a65ca580-bb29-4224-bdb1-7178ca9f7a45-kube-api-access-lj8cc\") pod \"a65ca580-bb29-4224-bdb1-7178ca9f7a45\" (UID: \"a65ca580-bb29-4224-bdb1-7178ca9f7a45\") " Nov 21 15:05:55 crc kubenswrapper[4897]: I1121 15:05:55.299976 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a65ca580-bb29-4224-bdb1-7178ca9f7a45-ssh-key\") pod \"a65ca580-bb29-4224-bdb1-7178ca9f7a45\" (UID: \"a65ca580-bb29-4224-bdb1-7178ca9f7a45\") " Nov 21 15:05:55 crc kubenswrapper[4897]: I1121 15:05:55.300035 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/a65ca580-bb29-4224-bdb1-7178ca9f7a45-nova-migration-ssh-key-0\") pod \"a65ca580-bb29-4224-bdb1-7178ca9f7a45\" (UID: \"a65ca580-bb29-4224-bdb1-7178ca9f7a45\") " Nov 21 15:05:55 crc kubenswrapper[4897]: I1121 15:05:55.300090 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a65ca580-bb29-4224-bdb1-7178ca9f7a45-inventory\") pod \"a65ca580-bb29-4224-bdb1-7178ca9f7a45\" (UID: \"a65ca580-bb29-4224-bdb1-7178ca9f7a45\") " Nov 21 15:05:55 crc kubenswrapper[4897]: I1121 15:05:55.300121 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/a65ca580-bb29-4224-bdb1-7178ca9f7a45-nova-extra-config-0\") pod \"a65ca580-bb29-4224-bdb1-7178ca9f7a45\" (UID: \"a65ca580-bb29-4224-bdb1-7178ca9f7a45\") " Nov 21 15:05:55 crc kubenswrapper[4897]: I1121 15:05:55.300140 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/a65ca580-bb29-4224-bdb1-7178ca9f7a45-nova-cell1-compute-config-0\") pod \"a65ca580-bb29-4224-bdb1-7178ca9f7a45\" (UID: \"a65ca580-bb29-4224-bdb1-7178ca9f7a45\") " Nov 21 15:05:55 crc kubenswrapper[4897]: I1121 15:05:55.305173 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a65ca580-bb29-4224-bdb1-7178ca9f7a45-kube-api-access-lj8cc" (OuterVolumeSpecName: "kube-api-access-lj8cc") pod "a65ca580-bb29-4224-bdb1-7178ca9f7a45" (UID: "a65ca580-bb29-4224-bdb1-7178ca9f7a45"). InnerVolumeSpecName "kube-api-access-lj8cc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 15:05:55 crc kubenswrapper[4897]: I1121 15:05:55.307767 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a65ca580-bb29-4224-bdb1-7178ca9f7a45-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "a65ca580-bb29-4224-bdb1-7178ca9f7a45" (UID: "a65ca580-bb29-4224-bdb1-7178ca9f7a45"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 15:05:55 crc kubenswrapper[4897]: I1121 15:05:55.337473 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a65ca580-bb29-4224-bdb1-7178ca9f7a45-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "a65ca580-bb29-4224-bdb1-7178ca9f7a45" (UID: "a65ca580-bb29-4224-bdb1-7178ca9f7a45"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 15:05:55 crc kubenswrapper[4897]: I1121 15:05:55.337531 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a65ca580-bb29-4224-bdb1-7178ca9f7a45-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "a65ca580-bb29-4224-bdb1-7178ca9f7a45" (UID: "a65ca580-bb29-4224-bdb1-7178ca9f7a45"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 15:05:55 crc kubenswrapper[4897]: I1121 15:05:55.338594 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a65ca580-bb29-4224-bdb1-7178ca9f7a45-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "a65ca580-bb29-4224-bdb1-7178ca9f7a45" (UID: "a65ca580-bb29-4224-bdb1-7178ca9f7a45"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 15:05:55 crc kubenswrapper[4897]: I1121 15:05:55.339020 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a65ca580-bb29-4224-bdb1-7178ca9f7a45-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "a65ca580-bb29-4224-bdb1-7178ca9f7a45" (UID: "a65ca580-bb29-4224-bdb1-7178ca9f7a45"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 15:05:55 crc kubenswrapper[4897]: I1121 15:05:55.341838 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a65ca580-bb29-4224-bdb1-7178ca9f7a45-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "a65ca580-bb29-4224-bdb1-7178ca9f7a45" (UID: "a65ca580-bb29-4224-bdb1-7178ca9f7a45"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 15:05:55 crc kubenswrapper[4897]: I1121 15:05:55.344368 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a65ca580-bb29-4224-bdb1-7178ca9f7a45-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "a65ca580-bb29-4224-bdb1-7178ca9f7a45" (UID: "a65ca580-bb29-4224-bdb1-7178ca9f7a45"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 15:05:55 crc kubenswrapper[4897]: I1121 15:05:55.345097 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a65ca580-bb29-4224-bdb1-7178ca9f7a45-inventory" (OuterVolumeSpecName: "inventory") pod "a65ca580-bb29-4224-bdb1-7178ca9f7a45" (UID: "a65ca580-bb29-4224-bdb1-7178ca9f7a45"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 15:05:55 crc kubenswrapper[4897]: I1121 15:05:55.402657 4897 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a65ca580-bb29-4224-bdb1-7178ca9f7a45-inventory\") on node \"crc\" DevicePath \"\"" Nov 21 15:05:55 crc kubenswrapper[4897]: I1121 15:05:55.402692 4897 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/a65ca580-bb29-4224-bdb1-7178ca9f7a45-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Nov 21 15:05:55 crc kubenswrapper[4897]: I1121 15:05:55.402703 4897 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/a65ca580-bb29-4224-bdb1-7178ca9f7a45-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Nov 21 15:05:55 crc kubenswrapper[4897]: I1121 15:05:55.402714 4897 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/a65ca580-bb29-4224-bdb1-7178ca9f7a45-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Nov 21 15:05:55 crc kubenswrapper[4897]: I1121 15:05:55.402723 4897 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/a65ca580-bb29-4224-bdb1-7178ca9f7a45-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Nov 21 15:05:55 crc kubenswrapper[4897]: I1121 15:05:55.402731 4897 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a65ca580-bb29-4224-bdb1-7178ca9f7a45-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 21 15:05:55 crc kubenswrapper[4897]: I1121 15:05:55.402743 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lj8cc\" (UniqueName: \"kubernetes.io/projected/a65ca580-bb29-4224-bdb1-7178ca9f7a45-kube-api-access-lj8cc\") on node \"crc\" DevicePath \"\"" Nov 21 15:05:55 crc kubenswrapper[4897]: I1121 15:05:55.402750 4897 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a65ca580-bb29-4224-bdb1-7178ca9f7a45-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 21 15:05:55 crc kubenswrapper[4897]: I1121 15:05:55.402759 4897 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/a65ca580-bb29-4224-bdb1-7178ca9f7a45-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Nov 21 15:05:55 crc kubenswrapper[4897]: I1121 15:05:55.575233 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8vsfm" event={"ID":"a65ca580-bb29-4224-bdb1-7178ca9f7a45","Type":"ContainerDied","Data":"347cff58a0060704026bc233089dbf7c3fb957adfb7b732f29e56a2ae78706ea"} Nov 21 15:05:55 crc kubenswrapper[4897]: I1121 15:05:55.575529 4897 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="347cff58a0060704026bc233089dbf7c3fb957adfb7b732f29e56a2ae78706ea" Nov 21 15:05:55 crc kubenswrapper[4897]: I1121 15:05:55.575316 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-8vsfm" Nov 21 15:05:55 crc kubenswrapper[4897]: I1121 15:05:55.674783 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mkn7j"] Nov 21 15:05:55 crc kubenswrapper[4897]: E1121 15:05:55.675657 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08353c71-61dc-46a2-80bf-a83f1207b572" containerName="extract-content" Nov 21 15:05:55 crc kubenswrapper[4897]: I1121 15:05:55.675682 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="08353c71-61dc-46a2-80bf-a83f1207b572" containerName="extract-content" Nov 21 15:05:55 crc kubenswrapper[4897]: E1121 15:05:55.675703 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08353c71-61dc-46a2-80bf-a83f1207b572" containerName="registry-server" Nov 21 15:05:55 crc kubenswrapper[4897]: I1121 15:05:55.675712 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="08353c71-61dc-46a2-80bf-a83f1207b572" containerName="registry-server" Nov 21 15:05:55 crc kubenswrapper[4897]: E1121 15:05:55.675731 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08353c71-61dc-46a2-80bf-a83f1207b572" containerName="extract-utilities" Nov 21 15:05:55 crc kubenswrapper[4897]: I1121 15:05:55.675740 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="08353c71-61dc-46a2-80bf-a83f1207b572" containerName="extract-utilities" Nov 21 15:05:55 crc kubenswrapper[4897]: E1121 15:05:55.675756 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5469f78-a615-45fe-8a8a-30d76071e075" containerName="registry-server" Nov 21 15:05:55 crc kubenswrapper[4897]: I1121 15:05:55.675765 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5469f78-a615-45fe-8a8a-30d76071e075" containerName="registry-server" Nov 21 15:05:55 crc kubenswrapper[4897]: E1121 15:05:55.675790 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5469f78-a615-45fe-8a8a-30d76071e075" containerName="extract-utilities" Nov 21 15:05:55 crc kubenswrapper[4897]: I1121 15:05:55.675796 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5469f78-a615-45fe-8a8a-30d76071e075" containerName="extract-utilities" Nov 21 15:05:55 crc kubenswrapper[4897]: E1121 15:05:55.675809 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a65ca580-bb29-4224-bdb1-7178ca9f7a45" containerName="nova-edpm-deployment-openstack-edpm-ipam" Nov 21 15:05:55 crc kubenswrapper[4897]: I1121 15:05:55.675816 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="a65ca580-bb29-4224-bdb1-7178ca9f7a45" containerName="nova-edpm-deployment-openstack-edpm-ipam" Nov 21 15:05:55 crc kubenswrapper[4897]: E1121 15:05:55.675833 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5469f78-a615-45fe-8a8a-30d76071e075" containerName="extract-content" Nov 21 15:05:55 crc kubenswrapper[4897]: I1121 15:05:55.675841 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5469f78-a615-45fe-8a8a-30d76071e075" containerName="extract-content" Nov 21 15:05:55 crc kubenswrapper[4897]: I1121 15:05:55.676078 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="08353c71-61dc-46a2-80bf-a83f1207b572" containerName="registry-server" Nov 21 15:05:55 crc kubenswrapper[4897]: I1121 15:05:55.676095 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5469f78-a615-45fe-8a8a-30d76071e075" containerName="registry-server" Nov 21 15:05:55 crc kubenswrapper[4897]: I1121 15:05:55.676103 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="a65ca580-bb29-4224-bdb1-7178ca9f7a45" containerName="nova-edpm-deployment-openstack-edpm-ipam" Nov 21 15:05:55 crc kubenswrapper[4897]: I1121 15:05:55.677214 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mkn7j" Nov 21 15:05:55 crc kubenswrapper[4897]: I1121 15:05:55.679242 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Nov 21 15:05:55 crc kubenswrapper[4897]: I1121 15:05:55.679501 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-dj88m" Nov 21 15:05:55 crc kubenswrapper[4897]: I1121 15:05:55.679791 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 21 15:05:55 crc kubenswrapper[4897]: I1121 15:05:55.681918 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 21 15:05:55 crc kubenswrapper[4897]: I1121 15:05:55.682737 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 21 15:05:55 crc kubenswrapper[4897]: I1121 15:05:55.691213 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mkn7j"] Nov 21 15:05:55 crc kubenswrapper[4897]: I1121 15:05:55.818880 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-92vkz\" (UniqueName: \"kubernetes.io/projected/5a84ad98-d97c-4702-9a34-12102daca98c-kube-api-access-92vkz\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-mkn7j\" (UID: \"5a84ad98-d97c-4702-9a34-12102daca98c\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mkn7j" Nov 21 15:05:55 crc kubenswrapper[4897]: I1121 15:05:55.819089 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a84ad98-d97c-4702-9a34-12102daca98c-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-mkn7j\" (UID: \"5a84ad98-d97c-4702-9a34-12102daca98c\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mkn7j" Nov 21 15:05:55 crc kubenswrapper[4897]: I1121 15:05:55.819209 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/5a84ad98-d97c-4702-9a34-12102daca98c-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-mkn7j\" (UID: \"5a84ad98-d97c-4702-9a34-12102daca98c\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mkn7j" Nov 21 15:05:55 crc kubenswrapper[4897]: I1121 15:05:55.819434 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5a84ad98-d97c-4702-9a34-12102daca98c-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-mkn7j\" (UID: \"5a84ad98-d97c-4702-9a34-12102daca98c\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mkn7j" Nov 21 15:05:55 crc kubenswrapper[4897]: I1121 15:05:55.819531 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/5a84ad98-d97c-4702-9a34-12102daca98c-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-mkn7j\" (UID: \"5a84ad98-d97c-4702-9a34-12102daca98c\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mkn7j" Nov 21 15:05:55 crc kubenswrapper[4897]: I1121 15:05:55.819680 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5a84ad98-d97c-4702-9a34-12102daca98c-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-mkn7j\" (UID: \"5a84ad98-d97c-4702-9a34-12102daca98c\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mkn7j" Nov 21 15:05:55 crc kubenswrapper[4897]: I1121 15:05:55.819736 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/5a84ad98-d97c-4702-9a34-12102daca98c-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-mkn7j\" (UID: \"5a84ad98-d97c-4702-9a34-12102daca98c\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mkn7j" Nov 21 15:05:55 crc kubenswrapper[4897]: I1121 15:05:55.922629 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-92vkz\" (UniqueName: \"kubernetes.io/projected/5a84ad98-d97c-4702-9a34-12102daca98c-kube-api-access-92vkz\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-mkn7j\" (UID: \"5a84ad98-d97c-4702-9a34-12102daca98c\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mkn7j" Nov 21 15:05:55 crc kubenswrapper[4897]: I1121 15:05:55.922812 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a84ad98-d97c-4702-9a34-12102daca98c-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-mkn7j\" (UID: \"5a84ad98-d97c-4702-9a34-12102daca98c\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mkn7j" Nov 21 15:05:55 crc kubenswrapper[4897]: I1121 15:05:55.922895 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/5a84ad98-d97c-4702-9a34-12102daca98c-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-mkn7j\" (UID: \"5a84ad98-d97c-4702-9a34-12102daca98c\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mkn7j" Nov 21 15:05:55 crc kubenswrapper[4897]: I1121 15:05:55.922992 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5a84ad98-d97c-4702-9a34-12102daca98c-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-mkn7j\" (UID: \"5a84ad98-d97c-4702-9a34-12102daca98c\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mkn7j" Nov 21 15:05:55 crc kubenswrapper[4897]: I1121 15:05:55.923030 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/5a84ad98-d97c-4702-9a34-12102daca98c-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-mkn7j\" (UID: \"5a84ad98-d97c-4702-9a34-12102daca98c\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mkn7j" Nov 21 15:05:55 crc kubenswrapper[4897]: I1121 15:05:55.923113 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5a84ad98-d97c-4702-9a34-12102daca98c-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-mkn7j\" (UID: \"5a84ad98-d97c-4702-9a34-12102daca98c\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mkn7j" Nov 21 15:05:55 crc kubenswrapper[4897]: I1121 15:05:55.923170 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/5a84ad98-d97c-4702-9a34-12102daca98c-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-mkn7j\" (UID: \"5a84ad98-d97c-4702-9a34-12102daca98c\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mkn7j" Nov 21 15:05:55 crc kubenswrapper[4897]: I1121 15:05:55.928520 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5a84ad98-d97c-4702-9a34-12102daca98c-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-mkn7j\" (UID: \"5a84ad98-d97c-4702-9a34-12102daca98c\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mkn7j" Nov 21 15:05:55 crc kubenswrapper[4897]: I1121 15:05:55.928520 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/5a84ad98-d97c-4702-9a34-12102daca98c-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-mkn7j\" (UID: \"5a84ad98-d97c-4702-9a34-12102daca98c\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mkn7j" Nov 21 15:05:55 crc kubenswrapper[4897]: I1121 15:05:55.928941 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/5a84ad98-d97c-4702-9a34-12102daca98c-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-mkn7j\" (UID: \"5a84ad98-d97c-4702-9a34-12102daca98c\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mkn7j" Nov 21 15:05:55 crc kubenswrapper[4897]: I1121 15:05:55.929065 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/5a84ad98-d97c-4702-9a34-12102daca98c-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-mkn7j\" (UID: \"5a84ad98-d97c-4702-9a34-12102daca98c\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mkn7j" Nov 21 15:05:55 crc kubenswrapper[4897]: I1121 15:05:55.929233 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5a84ad98-d97c-4702-9a34-12102daca98c-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-mkn7j\" (UID: \"5a84ad98-d97c-4702-9a34-12102daca98c\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mkn7j" Nov 21 15:05:55 crc kubenswrapper[4897]: I1121 15:05:55.930879 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a84ad98-d97c-4702-9a34-12102daca98c-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-mkn7j\" (UID: \"5a84ad98-d97c-4702-9a34-12102daca98c\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mkn7j" Nov 21 15:05:55 crc kubenswrapper[4897]: I1121 15:05:55.944364 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-92vkz\" (UniqueName: \"kubernetes.io/projected/5a84ad98-d97c-4702-9a34-12102daca98c-kube-api-access-92vkz\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-mkn7j\" (UID: \"5a84ad98-d97c-4702-9a34-12102daca98c\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mkn7j" Nov 21 15:05:56 crc kubenswrapper[4897]: I1121 15:05:56.008082 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mkn7j" Nov 21 15:05:56 crc kubenswrapper[4897]: I1121 15:05:56.106696 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e5469f78-a615-45fe-8a8a-30d76071e075" path="/var/lib/kubelet/pods/e5469f78-a615-45fe-8a8a-30d76071e075/volumes" Nov 21 15:05:56 crc kubenswrapper[4897]: W1121 15:05:56.569054 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5a84ad98_d97c_4702_9a34_12102daca98c.slice/crio-df9c4e0b38fc8a500b645dc978080244bcda65a1401e0493d5e49786c436a85a WatchSource:0}: Error finding container df9c4e0b38fc8a500b645dc978080244bcda65a1401e0493d5e49786c436a85a: Status 404 returned error can't find the container with id df9c4e0b38fc8a500b645dc978080244bcda65a1401e0493d5e49786c436a85a Nov 21 15:05:56 crc kubenswrapper[4897]: I1121 15:05:56.570766 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mkn7j"] Nov 21 15:05:56 crc kubenswrapper[4897]: I1121 15:05:56.607202 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mkn7j" event={"ID":"5a84ad98-d97c-4702-9a34-12102daca98c","Type":"ContainerStarted","Data":"df9c4e0b38fc8a500b645dc978080244bcda65a1401e0493d5e49786c436a85a"} Nov 21 15:05:57 crc kubenswrapper[4897]: I1121 15:05:57.620664 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mkn7j" event={"ID":"5a84ad98-d97c-4702-9a34-12102daca98c","Type":"ContainerStarted","Data":"54c54c5fbf644fe845c73f8a1eb591e4ac37a6725a02a655a971847bac1f1add"} Nov 21 15:05:57 crc kubenswrapper[4897]: I1121 15:05:57.646861 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mkn7j" podStartSLOduration=2.191211412 podStartE2EDuration="2.646827412s" podCreationTimestamp="2025-11-21 15:05:55 +0000 UTC" firstStartedPulling="2025-11-21 15:05:56.571933869 +0000 UTC m=+3433.856527344" lastFinishedPulling="2025-11-21 15:05:57.027549869 +0000 UTC m=+3434.312143344" observedRunningTime="2025-11-21 15:05:57.640225672 +0000 UTC m=+3434.924819167" watchObservedRunningTime="2025-11-21 15:05:57.646827412 +0000 UTC m=+3434.931420887" Nov 21 15:06:04 crc kubenswrapper[4897]: I1121 15:06:04.371856 4897 patch_prober.go:28] interesting pod/machine-config-daemon-krv5b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 21 15:06:04 crc kubenswrapper[4897]: I1121 15:06:04.372496 4897 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 21 15:06:34 crc kubenswrapper[4897]: I1121 15:06:34.371187 4897 patch_prober.go:28] interesting pod/machine-config-daemon-krv5b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 21 15:06:34 crc kubenswrapper[4897]: I1121 15:06:34.371686 4897 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 21 15:06:34 crc kubenswrapper[4897]: I1121 15:06:34.371745 4897 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" Nov 21 15:06:34 crc kubenswrapper[4897]: I1121 15:06:34.372635 4897 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"11456ed300543798476a9b68095120ef6d6073c876d338f9212ec0e3865c7a1a"} pod="openshift-machine-config-operator/machine-config-daemon-krv5b" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 21 15:06:34 crc kubenswrapper[4897]: I1121 15:06:34.372687 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" containerName="machine-config-daemon" containerID="cri-o://11456ed300543798476a9b68095120ef6d6073c876d338f9212ec0e3865c7a1a" gracePeriod=600 Nov 21 15:06:35 crc kubenswrapper[4897]: I1121 15:06:35.073337 4897 generic.go:334] "Generic (PLEG): container finished" podID="e7670227-d280-4847-b882-754429f56b0f" containerID="11456ed300543798476a9b68095120ef6d6073c876d338f9212ec0e3865c7a1a" exitCode=0 Nov 21 15:06:35 crc kubenswrapper[4897]: I1121 15:06:35.073438 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" event={"ID":"e7670227-d280-4847-b882-754429f56b0f","Type":"ContainerDied","Data":"11456ed300543798476a9b68095120ef6d6073c876d338f9212ec0e3865c7a1a"} Nov 21 15:06:35 crc kubenswrapper[4897]: I1121 15:06:35.073924 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" event={"ID":"e7670227-d280-4847-b882-754429f56b0f","Type":"ContainerStarted","Data":"31498b4cf1d12960e0bd3369fd9c8702f2e4dc6c2b6a5487ecab62d136f4bc3c"} Nov 21 15:06:35 crc kubenswrapper[4897]: I1121 15:06:35.073949 4897 scope.go:117] "RemoveContainer" containerID="a17b834dbb567d490c0f7ce8a22ab5aa6ecb645fc460c063af4ca43f58348563" Nov 21 15:07:55 crc kubenswrapper[4897]: I1121 15:07:55.091287 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-5dqs7"] Nov 21 15:07:55 crc kubenswrapper[4897]: I1121 15:07:55.094601 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5dqs7" Nov 21 15:07:55 crc kubenswrapper[4897]: I1121 15:07:55.104687 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5dqs7"] Nov 21 15:07:55 crc kubenswrapper[4897]: I1121 15:07:55.133344 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vt9pw\" (UniqueName: \"kubernetes.io/projected/61f3ca01-1e4b-4899-8ec3-25376d44695f-kube-api-access-vt9pw\") pod \"redhat-operators-5dqs7\" (UID: \"61f3ca01-1e4b-4899-8ec3-25376d44695f\") " pod="openshift-marketplace/redhat-operators-5dqs7" Nov 21 15:07:55 crc kubenswrapper[4897]: I1121 15:07:55.133412 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/61f3ca01-1e4b-4899-8ec3-25376d44695f-catalog-content\") pod \"redhat-operators-5dqs7\" (UID: \"61f3ca01-1e4b-4899-8ec3-25376d44695f\") " pod="openshift-marketplace/redhat-operators-5dqs7" Nov 21 15:07:55 crc kubenswrapper[4897]: I1121 15:07:55.133724 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/61f3ca01-1e4b-4899-8ec3-25376d44695f-utilities\") pod \"redhat-operators-5dqs7\" (UID: \"61f3ca01-1e4b-4899-8ec3-25376d44695f\") " pod="openshift-marketplace/redhat-operators-5dqs7" Nov 21 15:07:55 crc kubenswrapper[4897]: I1121 15:07:55.235874 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/61f3ca01-1e4b-4899-8ec3-25376d44695f-utilities\") pod \"redhat-operators-5dqs7\" (UID: \"61f3ca01-1e4b-4899-8ec3-25376d44695f\") " pod="openshift-marketplace/redhat-operators-5dqs7" Nov 21 15:07:55 crc kubenswrapper[4897]: I1121 15:07:55.236053 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vt9pw\" (UniqueName: \"kubernetes.io/projected/61f3ca01-1e4b-4899-8ec3-25376d44695f-kube-api-access-vt9pw\") pod \"redhat-operators-5dqs7\" (UID: \"61f3ca01-1e4b-4899-8ec3-25376d44695f\") " pod="openshift-marketplace/redhat-operators-5dqs7" Nov 21 15:07:55 crc kubenswrapper[4897]: I1121 15:07:55.236089 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/61f3ca01-1e4b-4899-8ec3-25376d44695f-catalog-content\") pod \"redhat-operators-5dqs7\" (UID: \"61f3ca01-1e4b-4899-8ec3-25376d44695f\") " pod="openshift-marketplace/redhat-operators-5dqs7" Nov 21 15:07:55 crc kubenswrapper[4897]: I1121 15:07:55.236445 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/61f3ca01-1e4b-4899-8ec3-25376d44695f-utilities\") pod \"redhat-operators-5dqs7\" (UID: \"61f3ca01-1e4b-4899-8ec3-25376d44695f\") " pod="openshift-marketplace/redhat-operators-5dqs7" Nov 21 15:07:55 crc kubenswrapper[4897]: I1121 15:07:55.236556 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/61f3ca01-1e4b-4899-8ec3-25376d44695f-catalog-content\") pod \"redhat-operators-5dqs7\" (UID: \"61f3ca01-1e4b-4899-8ec3-25376d44695f\") " pod="openshift-marketplace/redhat-operators-5dqs7" Nov 21 15:07:55 crc kubenswrapper[4897]: I1121 15:07:55.257023 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vt9pw\" (UniqueName: \"kubernetes.io/projected/61f3ca01-1e4b-4899-8ec3-25376d44695f-kube-api-access-vt9pw\") pod \"redhat-operators-5dqs7\" (UID: \"61f3ca01-1e4b-4899-8ec3-25376d44695f\") " pod="openshift-marketplace/redhat-operators-5dqs7" Nov 21 15:07:55 crc kubenswrapper[4897]: I1121 15:07:55.417814 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5dqs7" Nov 21 15:07:55 crc kubenswrapper[4897]: I1121 15:07:55.977723 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5dqs7"] Nov 21 15:07:56 crc kubenswrapper[4897]: I1121 15:07:56.018107 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5dqs7" event={"ID":"61f3ca01-1e4b-4899-8ec3-25376d44695f","Type":"ContainerStarted","Data":"e3e3a8c4a0f30d6532a7603c8c5123f01265c2809557ffa7868f470455d9dbd3"} Nov 21 15:07:57 crc kubenswrapper[4897]: I1121 15:07:57.032322 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5dqs7" event={"ID":"61f3ca01-1e4b-4899-8ec3-25376d44695f","Type":"ContainerStarted","Data":"8c1a80fbe6cd446617866f25e7d331f466b6a6b3afcde7001884de53f9271d4b"} Nov 21 15:07:58 crc kubenswrapper[4897]: I1121 15:07:58.046126 4897 generic.go:334] "Generic (PLEG): container finished" podID="61f3ca01-1e4b-4899-8ec3-25376d44695f" containerID="8c1a80fbe6cd446617866f25e7d331f466b6a6b3afcde7001884de53f9271d4b" exitCode=0 Nov 21 15:07:58 crc kubenswrapper[4897]: I1121 15:07:58.046205 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5dqs7" event={"ID":"61f3ca01-1e4b-4899-8ec3-25376d44695f","Type":"ContainerDied","Data":"8c1a80fbe6cd446617866f25e7d331f466b6a6b3afcde7001884de53f9271d4b"} Nov 21 15:07:58 crc kubenswrapper[4897]: I1121 15:07:58.048873 4897 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 21 15:08:00 crc kubenswrapper[4897]: I1121 15:08:00.070307 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5dqs7" event={"ID":"61f3ca01-1e4b-4899-8ec3-25376d44695f","Type":"ContainerStarted","Data":"1a2903e49cb98c8856ca389190b2d8c5316ae06ee2a2bbb5bf523fb22b1970da"} Nov 21 15:08:07 crc kubenswrapper[4897]: I1121 15:08:07.148681 4897 generic.go:334] "Generic (PLEG): container finished" podID="61f3ca01-1e4b-4899-8ec3-25376d44695f" containerID="1a2903e49cb98c8856ca389190b2d8c5316ae06ee2a2bbb5bf523fb22b1970da" exitCode=0 Nov 21 15:08:07 crc kubenswrapper[4897]: I1121 15:08:07.148742 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5dqs7" event={"ID":"61f3ca01-1e4b-4899-8ec3-25376d44695f","Type":"ContainerDied","Data":"1a2903e49cb98c8856ca389190b2d8c5316ae06ee2a2bbb5bf523fb22b1970da"} Nov 21 15:08:08 crc kubenswrapper[4897]: I1121 15:08:08.170235 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5dqs7" event={"ID":"61f3ca01-1e4b-4899-8ec3-25376d44695f","Type":"ContainerStarted","Data":"2982c7fbdd5a9f455c8397a3bfddc92b9b293106c9ae1fb8d58a35319dc5d797"} Nov 21 15:08:08 crc kubenswrapper[4897]: I1121 15:08:08.191732 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-5dqs7" podStartSLOduration=3.535055227 podStartE2EDuration="13.191691282s" podCreationTimestamp="2025-11-21 15:07:55 +0000 UTC" firstStartedPulling="2025-11-21 15:07:58.048477755 +0000 UTC m=+3555.333071230" lastFinishedPulling="2025-11-21 15:08:07.70511381 +0000 UTC m=+3564.989707285" observedRunningTime="2025-11-21 15:08:08.18647158 +0000 UTC m=+3565.471065075" watchObservedRunningTime="2025-11-21 15:08:08.191691282 +0000 UTC m=+3565.476284757" Nov 21 15:08:15 crc kubenswrapper[4897]: I1121 15:08:15.419851 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-5dqs7" Nov 21 15:08:15 crc kubenswrapper[4897]: I1121 15:08:15.420353 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-5dqs7" Nov 21 15:08:16 crc kubenswrapper[4897]: I1121 15:08:16.468650 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-5dqs7" podUID="61f3ca01-1e4b-4899-8ec3-25376d44695f" containerName="registry-server" probeResult="failure" output=< Nov 21 15:08:16 crc kubenswrapper[4897]: timeout: failed to connect service ":50051" within 1s Nov 21 15:08:16 crc kubenswrapper[4897]: > Nov 21 15:08:26 crc kubenswrapper[4897]: I1121 15:08:26.470807 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-5dqs7" podUID="61f3ca01-1e4b-4899-8ec3-25376d44695f" containerName="registry-server" probeResult="failure" output=< Nov 21 15:08:26 crc kubenswrapper[4897]: timeout: failed to connect service ":50051" within 1s Nov 21 15:08:26 crc kubenswrapper[4897]: > Nov 21 15:08:27 crc kubenswrapper[4897]: I1121 15:08:27.373618 4897 generic.go:334] "Generic (PLEG): container finished" podID="5a84ad98-d97c-4702-9a34-12102daca98c" containerID="54c54c5fbf644fe845c73f8a1eb591e4ac37a6725a02a655a971847bac1f1add" exitCode=0 Nov 21 15:08:27 crc kubenswrapper[4897]: I1121 15:08:27.373748 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mkn7j" event={"ID":"5a84ad98-d97c-4702-9a34-12102daca98c","Type":"ContainerDied","Data":"54c54c5fbf644fe845c73f8a1eb591e4ac37a6725a02a655a971847bac1f1add"} Nov 21 15:08:28 crc kubenswrapper[4897]: I1121 15:08:28.947467 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mkn7j" Nov 21 15:08:29 crc kubenswrapper[4897]: I1121 15:08:29.008660 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/5a84ad98-d97c-4702-9a34-12102daca98c-ceilometer-compute-config-data-2\") pod \"5a84ad98-d97c-4702-9a34-12102daca98c\" (UID: \"5a84ad98-d97c-4702-9a34-12102daca98c\") " Nov 21 15:08:29 crc kubenswrapper[4897]: I1121 15:08:29.008802 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a84ad98-d97c-4702-9a34-12102daca98c-telemetry-combined-ca-bundle\") pod \"5a84ad98-d97c-4702-9a34-12102daca98c\" (UID: \"5a84ad98-d97c-4702-9a34-12102daca98c\") " Nov 21 15:08:29 crc kubenswrapper[4897]: I1121 15:08:29.008830 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5a84ad98-d97c-4702-9a34-12102daca98c-inventory\") pod \"5a84ad98-d97c-4702-9a34-12102daca98c\" (UID: \"5a84ad98-d97c-4702-9a34-12102daca98c\") " Nov 21 15:08:29 crc kubenswrapper[4897]: I1121 15:08:29.009146 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5a84ad98-d97c-4702-9a34-12102daca98c-ssh-key\") pod \"5a84ad98-d97c-4702-9a34-12102daca98c\" (UID: \"5a84ad98-d97c-4702-9a34-12102daca98c\") " Nov 21 15:08:29 crc kubenswrapper[4897]: I1121 15:08:29.009215 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-92vkz\" (UniqueName: \"kubernetes.io/projected/5a84ad98-d97c-4702-9a34-12102daca98c-kube-api-access-92vkz\") pod \"5a84ad98-d97c-4702-9a34-12102daca98c\" (UID: \"5a84ad98-d97c-4702-9a34-12102daca98c\") " Nov 21 15:08:29 crc kubenswrapper[4897]: I1121 15:08:29.009310 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/5a84ad98-d97c-4702-9a34-12102daca98c-ceilometer-compute-config-data-0\") pod \"5a84ad98-d97c-4702-9a34-12102daca98c\" (UID: \"5a84ad98-d97c-4702-9a34-12102daca98c\") " Nov 21 15:08:29 crc kubenswrapper[4897]: I1121 15:08:29.009363 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/5a84ad98-d97c-4702-9a34-12102daca98c-ceilometer-compute-config-data-1\") pod \"5a84ad98-d97c-4702-9a34-12102daca98c\" (UID: \"5a84ad98-d97c-4702-9a34-12102daca98c\") " Nov 21 15:08:29 crc kubenswrapper[4897]: I1121 15:08:29.015625 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a84ad98-d97c-4702-9a34-12102daca98c-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "5a84ad98-d97c-4702-9a34-12102daca98c" (UID: "5a84ad98-d97c-4702-9a34-12102daca98c"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 15:08:29 crc kubenswrapper[4897]: I1121 15:08:29.015780 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a84ad98-d97c-4702-9a34-12102daca98c-kube-api-access-92vkz" (OuterVolumeSpecName: "kube-api-access-92vkz") pod "5a84ad98-d97c-4702-9a34-12102daca98c" (UID: "5a84ad98-d97c-4702-9a34-12102daca98c"). InnerVolumeSpecName "kube-api-access-92vkz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 15:08:29 crc kubenswrapper[4897]: I1121 15:08:29.046293 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a84ad98-d97c-4702-9a34-12102daca98c-inventory" (OuterVolumeSpecName: "inventory") pod "5a84ad98-d97c-4702-9a34-12102daca98c" (UID: "5a84ad98-d97c-4702-9a34-12102daca98c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 15:08:29 crc kubenswrapper[4897]: I1121 15:08:29.051080 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a84ad98-d97c-4702-9a34-12102daca98c-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "5a84ad98-d97c-4702-9a34-12102daca98c" (UID: "5a84ad98-d97c-4702-9a34-12102daca98c"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 15:08:29 crc kubenswrapper[4897]: I1121 15:08:29.052625 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a84ad98-d97c-4702-9a34-12102daca98c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "5a84ad98-d97c-4702-9a34-12102daca98c" (UID: "5a84ad98-d97c-4702-9a34-12102daca98c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 15:08:29 crc kubenswrapper[4897]: I1121 15:08:29.053784 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a84ad98-d97c-4702-9a34-12102daca98c-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "5a84ad98-d97c-4702-9a34-12102daca98c" (UID: "5a84ad98-d97c-4702-9a34-12102daca98c"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 15:08:29 crc kubenswrapper[4897]: I1121 15:08:29.057638 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a84ad98-d97c-4702-9a34-12102daca98c-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "5a84ad98-d97c-4702-9a34-12102daca98c" (UID: "5a84ad98-d97c-4702-9a34-12102daca98c"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 15:08:29 crc kubenswrapper[4897]: I1121 15:08:29.113034 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-92vkz\" (UniqueName: \"kubernetes.io/projected/5a84ad98-d97c-4702-9a34-12102daca98c-kube-api-access-92vkz\") on node \"crc\" DevicePath \"\"" Nov 21 15:08:29 crc kubenswrapper[4897]: I1121 15:08:29.113069 4897 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/5a84ad98-d97c-4702-9a34-12102daca98c-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Nov 21 15:08:29 crc kubenswrapper[4897]: I1121 15:08:29.113082 4897 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/5a84ad98-d97c-4702-9a34-12102daca98c-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Nov 21 15:08:29 crc kubenswrapper[4897]: I1121 15:08:29.113092 4897 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/5a84ad98-d97c-4702-9a34-12102daca98c-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Nov 21 15:08:29 crc kubenswrapper[4897]: I1121 15:08:29.113102 4897 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a84ad98-d97c-4702-9a34-12102daca98c-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 21 15:08:29 crc kubenswrapper[4897]: I1121 15:08:29.113113 4897 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5a84ad98-d97c-4702-9a34-12102daca98c-inventory\") on node \"crc\" DevicePath \"\"" Nov 21 15:08:29 crc kubenswrapper[4897]: I1121 15:08:29.113121 4897 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5a84ad98-d97c-4702-9a34-12102daca98c-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 21 15:08:29 crc kubenswrapper[4897]: I1121 15:08:29.402852 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mkn7j" event={"ID":"5a84ad98-d97c-4702-9a34-12102daca98c","Type":"ContainerDied","Data":"df9c4e0b38fc8a500b645dc978080244bcda65a1401e0493d5e49786c436a85a"} Nov 21 15:08:29 crc kubenswrapper[4897]: I1121 15:08:29.403227 4897 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="df9c4e0b38fc8a500b645dc978080244bcda65a1401e0493d5e49786c436a85a" Nov 21 15:08:29 crc kubenswrapper[4897]: I1121 15:08:29.402910 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-mkn7j" Nov 21 15:08:29 crc kubenswrapper[4897]: I1121 15:08:29.497090 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-n2lpc"] Nov 21 15:08:29 crc kubenswrapper[4897]: E1121 15:08:29.497625 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a84ad98-d97c-4702-9a34-12102daca98c" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Nov 21 15:08:29 crc kubenswrapper[4897]: I1121 15:08:29.497644 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a84ad98-d97c-4702-9a34-12102daca98c" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Nov 21 15:08:29 crc kubenswrapper[4897]: I1121 15:08:29.497885 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a84ad98-d97c-4702-9a34-12102daca98c" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Nov 21 15:08:29 crc kubenswrapper[4897]: I1121 15:08:29.498748 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-n2lpc" Nov 21 15:08:29 crc kubenswrapper[4897]: I1121 15:08:29.501967 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 21 15:08:29 crc kubenswrapper[4897]: I1121 15:08:29.502129 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-dj88m" Nov 21 15:08:29 crc kubenswrapper[4897]: I1121 15:08:29.503856 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-ipmi-config-data" Nov 21 15:08:29 crc kubenswrapper[4897]: I1121 15:08:29.504169 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 21 15:08:29 crc kubenswrapper[4897]: I1121 15:08:29.504707 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 21 15:08:29 crc kubenswrapper[4897]: I1121 15:08:29.523164 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-n2lpc"] Nov 21 15:08:29 crc kubenswrapper[4897]: I1121 15:08:29.630644 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae0e489e-6389-4ae0-947d-c75280d74e38-telemetry-power-monitoring-combined-ca-bundle\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-n2lpc\" (UID: \"ae0e489e-6389-4ae0-947d-c75280d74e38\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-n2lpc" Nov 21 15:08:29 crc kubenswrapper[4897]: I1121 15:08:29.630709 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8d2hj\" (UniqueName: \"kubernetes.io/projected/ae0e489e-6389-4ae0-947d-c75280d74e38-kube-api-access-8d2hj\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-n2lpc\" (UID: \"ae0e489e-6389-4ae0-947d-c75280d74e38\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-n2lpc" Nov 21 15:08:29 crc kubenswrapper[4897]: I1121 15:08:29.630823 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-ipmi-config-data-1\" (UniqueName: \"kubernetes.io/secret/ae0e489e-6389-4ae0-947d-c75280d74e38-ceilometer-ipmi-config-data-1\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-n2lpc\" (UID: \"ae0e489e-6389-4ae0-947d-c75280d74e38\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-n2lpc" Nov 21 15:08:29 crc kubenswrapper[4897]: I1121 15:08:29.630872 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ae0e489e-6389-4ae0-947d-c75280d74e38-inventory\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-n2lpc\" (UID: \"ae0e489e-6389-4ae0-947d-c75280d74e38\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-n2lpc" Nov 21 15:08:29 crc kubenswrapper[4897]: I1121 15:08:29.631004 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-ipmi-config-data-2\" (UniqueName: \"kubernetes.io/secret/ae0e489e-6389-4ae0-947d-c75280d74e38-ceilometer-ipmi-config-data-2\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-n2lpc\" (UID: \"ae0e489e-6389-4ae0-947d-c75280d74e38\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-n2lpc" Nov 21 15:08:29 crc kubenswrapper[4897]: I1121 15:08:29.631089 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-ipmi-config-data-0\" (UniqueName: \"kubernetes.io/secret/ae0e489e-6389-4ae0-947d-c75280d74e38-ceilometer-ipmi-config-data-0\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-n2lpc\" (UID: \"ae0e489e-6389-4ae0-947d-c75280d74e38\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-n2lpc" Nov 21 15:08:29 crc kubenswrapper[4897]: I1121 15:08:29.631124 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ae0e489e-6389-4ae0-947d-c75280d74e38-ssh-key\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-n2lpc\" (UID: \"ae0e489e-6389-4ae0-947d-c75280d74e38\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-n2lpc" Nov 21 15:08:29 crc kubenswrapper[4897]: I1121 15:08:29.733627 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-ipmi-config-data-2\" (UniqueName: \"kubernetes.io/secret/ae0e489e-6389-4ae0-947d-c75280d74e38-ceilometer-ipmi-config-data-2\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-n2lpc\" (UID: \"ae0e489e-6389-4ae0-947d-c75280d74e38\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-n2lpc" Nov 21 15:08:29 crc kubenswrapper[4897]: I1121 15:08:29.733806 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-ipmi-config-data-0\" (UniqueName: \"kubernetes.io/secret/ae0e489e-6389-4ae0-947d-c75280d74e38-ceilometer-ipmi-config-data-0\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-n2lpc\" (UID: \"ae0e489e-6389-4ae0-947d-c75280d74e38\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-n2lpc" Nov 21 15:08:29 crc kubenswrapper[4897]: I1121 15:08:29.733854 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ae0e489e-6389-4ae0-947d-c75280d74e38-ssh-key\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-n2lpc\" (UID: \"ae0e489e-6389-4ae0-947d-c75280d74e38\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-n2lpc" Nov 21 15:08:29 crc kubenswrapper[4897]: I1121 15:08:29.733926 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae0e489e-6389-4ae0-947d-c75280d74e38-telemetry-power-monitoring-combined-ca-bundle\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-n2lpc\" (UID: \"ae0e489e-6389-4ae0-947d-c75280d74e38\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-n2lpc" Nov 21 15:08:29 crc kubenswrapper[4897]: I1121 15:08:29.733973 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8d2hj\" (UniqueName: \"kubernetes.io/projected/ae0e489e-6389-4ae0-947d-c75280d74e38-kube-api-access-8d2hj\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-n2lpc\" (UID: \"ae0e489e-6389-4ae0-947d-c75280d74e38\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-n2lpc" Nov 21 15:08:29 crc kubenswrapper[4897]: I1121 15:08:29.734068 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-ipmi-config-data-1\" (UniqueName: \"kubernetes.io/secret/ae0e489e-6389-4ae0-947d-c75280d74e38-ceilometer-ipmi-config-data-1\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-n2lpc\" (UID: \"ae0e489e-6389-4ae0-947d-c75280d74e38\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-n2lpc" Nov 21 15:08:29 crc kubenswrapper[4897]: I1121 15:08:29.734134 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ae0e489e-6389-4ae0-947d-c75280d74e38-inventory\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-n2lpc\" (UID: \"ae0e489e-6389-4ae0-947d-c75280d74e38\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-n2lpc" Nov 21 15:08:29 crc kubenswrapper[4897]: I1121 15:08:29.738405 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-ipmi-config-data-1\" (UniqueName: \"kubernetes.io/secret/ae0e489e-6389-4ae0-947d-c75280d74e38-ceilometer-ipmi-config-data-1\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-n2lpc\" (UID: \"ae0e489e-6389-4ae0-947d-c75280d74e38\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-n2lpc" Nov 21 15:08:29 crc kubenswrapper[4897]: I1121 15:08:29.738719 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-ipmi-config-data-0\" (UniqueName: \"kubernetes.io/secret/ae0e489e-6389-4ae0-947d-c75280d74e38-ceilometer-ipmi-config-data-0\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-n2lpc\" (UID: \"ae0e489e-6389-4ae0-947d-c75280d74e38\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-n2lpc" Nov 21 15:08:29 crc kubenswrapper[4897]: I1121 15:08:29.738919 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ae0e489e-6389-4ae0-947d-c75280d74e38-inventory\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-n2lpc\" (UID: \"ae0e489e-6389-4ae0-947d-c75280d74e38\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-n2lpc" Nov 21 15:08:29 crc kubenswrapper[4897]: I1121 15:08:29.739550 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae0e489e-6389-4ae0-947d-c75280d74e38-telemetry-power-monitoring-combined-ca-bundle\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-n2lpc\" (UID: \"ae0e489e-6389-4ae0-947d-c75280d74e38\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-n2lpc" Nov 21 15:08:29 crc kubenswrapper[4897]: I1121 15:08:29.740423 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ae0e489e-6389-4ae0-947d-c75280d74e38-ssh-key\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-n2lpc\" (UID: \"ae0e489e-6389-4ae0-947d-c75280d74e38\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-n2lpc" Nov 21 15:08:29 crc kubenswrapper[4897]: I1121 15:08:29.740889 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-ipmi-config-data-2\" (UniqueName: \"kubernetes.io/secret/ae0e489e-6389-4ae0-947d-c75280d74e38-ceilometer-ipmi-config-data-2\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-n2lpc\" (UID: \"ae0e489e-6389-4ae0-947d-c75280d74e38\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-n2lpc" Nov 21 15:08:29 crc kubenswrapper[4897]: I1121 15:08:29.764736 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8d2hj\" (UniqueName: \"kubernetes.io/projected/ae0e489e-6389-4ae0-947d-c75280d74e38-kube-api-access-8d2hj\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-n2lpc\" (UID: \"ae0e489e-6389-4ae0-947d-c75280d74e38\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-n2lpc" Nov 21 15:08:29 crc kubenswrapper[4897]: I1121 15:08:29.836422 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-n2lpc" Nov 21 15:08:30 crc kubenswrapper[4897]: I1121 15:08:30.402572 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-n2lpc"] Nov 21 15:08:31 crc kubenswrapper[4897]: I1121 15:08:31.442583 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-n2lpc" event={"ID":"ae0e489e-6389-4ae0-947d-c75280d74e38","Type":"ContainerStarted","Data":"fd5c3bf8cd40748e4b49be3be91910ef4569b0fabcdb4e5c31afa2f18decf3f2"} Nov 21 15:08:32 crc kubenswrapper[4897]: I1121 15:08:32.459348 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-n2lpc" event={"ID":"ae0e489e-6389-4ae0-947d-c75280d74e38","Type":"ContainerStarted","Data":"952f057da4348c5e7c08a69819932881063626e548d4d32cf78a82a4077f83cc"} Nov 21 15:08:32 crc kubenswrapper[4897]: I1121 15:08:32.490302 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-n2lpc" podStartSLOduration=2.498375912 podStartE2EDuration="3.490285815s" podCreationTimestamp="2025-11-21 15:08:29 +0000 UTC" firstStartedPulling="2025-11-21 15:08:30.411969087 +0000 UTC m=+3587.696562562" lastFinishedPulling="2025-11-21 15:08:31.40387899 +0000 UTC m=+3588.688472465" observedRunningTime="2025-11-21 15:08:32.486525522 +0000 UTC m=+3589.771118997" watchObservedRunningTime="2025-11-21 15:08:32.490285815 +0000 UTC m=+3589.774879290" Nov 21 15:08:34 crc kubenswrapper[4897]: I1121 15:08:34.371170 4897 patch_prober.go:28] interesting pod/machine-config-daemon-krv5b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 21 15:08:34 crc kubenswrapper[4897]: I1121 15:08:34.371551 4897 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 21 15:08:36 crc kubenswrapper[4897]: I1121 15:08:36.470629 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-5dqs7" podUID="61f3ca01-1e4b-4899-8ec3-25376d44695f" containerName="registry-server" probeResult="failure" output=< Nov 21 15:08:36 crc kubenswrapper[4897]: timeout: failed to connect service ":50051" within 1s Nov 21 15:08:36 crc kubenswrapper[4897]: > Nov 21 15:08:46 crc kubenswrapper[4897]: I1121 15:08:46.471120 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-5dqs7" podUID="61f3ca01-1e4b-4899-8ec3-25376d44695f" containerName="registry-server" probeResult="failure" output=< Nov 21 15:08:46 crc kubenswrapper[4897]: timeout: failed to connect service ":50051" within 1s Nov 21 15:08:46 crc kubenswrapper[4897]: > Nov 21 15:08:56 crc kubenswrapper[4897]: I1121 15:08:56.468421 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-5dqs7" podUID="61f3ca01-1e4b-4899-8ec3-25376d44695f" containerName="registry-server" probeResult="failure" output=< Nov 21 15:08:56 crc kubenswrapper[4897]: timeout: failed to connect service ":50051" within 1s Nov 21 15:08:56 crc kubenswrapper[4897]: > Nov 21 15:09:04 crc kubenswrapper[4897]: I1121 15:09:04.371117 4897 patch_prober.go:28] interesting pod/machine-config-daemon-krv5b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 21 15:09:04 crc kubenswrapper[4897]: I1121 15:09:04.371905 4897 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 21 15:09:06 crc kubenswrapper[4897]: I1121 15:09:06.473990 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-5dqs7" podUID="61f3ca01-1e4b-4899-8ec3-25376d44695f" containerName="registry-server" probeResult="failure" output=< Nov 21 15:09:06 crc kubenswrapper[4897]: timeout: failed to connect service ":50051" within 1s Nov 21 15:09:06 crc kubenswrapper[4897]: > Nov 21 15:09:15 crc kubenswrapper[4897]: I1121 15:09:15.476318 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-5dqs7" Nov 21 15:09:15 crc kubenswrapper[4897]: I1121 15:09:15.538193 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-5dqs7" Nov 21 15:09:15 crc kubenswrapper[4897]: I1121 15:09:15.734104 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5dqs7"] Nov 21 15:09:16 crc kubenswrapper[4897]: I1121 15:09:16.952456 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-5dqs7" podUID="61f3ca01-1e4b-4899-8ec3-25376d44695f" containerName="registry-server" containerID="cri-o://2982c7fbdd5a9f455c8397a3bfddc92b9b293106c9ae1fb8d58a35319dc5d797" gracePeriod=2 Nov 21 15:09:17 crc kubenswrapper[4897]: I1121 15:09:17.475725 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5dqs7" Nov 21 15:09:17 crc kubenswrapper[4897]: I1121 15:09:17.580080 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt9pw\" (UniqueName: \"kubernetes.io/projected/61f3ca01-1e4b-4899-8ec3-25376d44695f-kube-api-access-vt9pw\") pod \"61f3ca01-1e4b-4899-8ec3-25376d44695f\" (UID: \"61f3ca01-1e4b-4899-8ec3-25376d44695f\") " Nov 21 15:09:17 crc kubenswrapper[4897]: I1121 15:09:17.580212 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/61f3ca01-1e4b-4899-8ec3-25376d44695f-catalog-content\") pod \"61f3ca01-1e4b-4899-8ec3-25376d44695f\" (UID: \"61f3ca01-1e4b-4899-8ec3-25376d44695f\") " Nov 21 15:09:17 crc kubenswrapper[4897]: I1121 15:09:17.580592 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/61f3ca01-1e4b-4899-8ec3-25376d44695f-utilities\") pod \"61f3ca01-1e4b-4899-8ec3-25376d44695f\" (UID: \"61f3ca01-1e4b-4899-8ec3-25376d44695f\") " Nov 21 15:09:17 crc kubenswrapper[4897]: I1121 15:09:17.581204 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/61f3ca01-1e4b-4899-8ec3-25376d44695f-utilities" (OuterVolumeSpecName: "utilities") pod "61f3ca01-1e4b-4899-8ec3-25376d44695f" (UID: "61f3ca01-1e4b-4899-8ec3-25376d44695f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 15:09:17 crc kubenswrapper[4897]: I1121 15:09:17.661117 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/61f3ca01-1e4b-4899-8ec3-25376d44695f-kube-api-access-vt9pw" (OuterVolumeSpecName: "kube-api-access-vt9pw") pod "61f3ca01-1e4b-4899-8ec3-25376d44695f" (UID: "61f3ca01-1e4b-4899-8ec3-25376d44695f"). InnerVolumeSpecName "kube-api-access-vt9pw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 15:09:17 crc kubenswrapper[4897]: I1121 15:09:17.666560 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/61f3ca01-1e4b-4899-8ec3-25376d44695f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "61f3ca01-1e4b-4899-8ec3-25376d44695f" (UID: "61f3ca01-1e4b-4899-8ec3-25376d44695f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 15:09:17 crc kubenswrapper[4897]: I1121 15:09:17.683818 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt9pw\" (UniqueName: \"kubernetes.io/projected/61f3ca01-1e4b-4899-8ec3-25376d44695f-kube-api-access-vt9pw\") on node \"crc\" DevicePath \"\"" Nov 21 15:09:17 crc kubenswrapper[4897]: I1121 15:09:17.683864 4897 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/61f3ca01-1e4b-4899-8ec3-25376d44695f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 21 15:09:17 crc kubenswrapper[4897]: I1121 15:09:17.683875 4897 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/61f3ca01-1e4b-4899-8ec3-25376d44695f-utilities\") on node \"crc\" DevicePath \"\"" Nov 21 15:09:17 crc kubenswrapper[4897]: I1121 15:09:17.966697 4897 generic.go:334] "Generic (PLEG): container finished" podID="61f3ca01-1e4b-4899-8ec3-25376d44695f" containerID="2982c7fbdd5a9f455c8397a3bfddc92b9b293106c9ae1fb8d58a35319dc5d797" exitCode=0 Nov 21 15:09:17 crc kubenswrapper[4897]: I1121 15:09:17.966783 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5dqs7" event={"ID":"61f3ca01-1e4b-4899-8ec3-25376d44695f","Type":"ContainerDied","Data":"2982c7fbdd5a9f455c8397a3bfddc92b9b293106c9ae1fb8d58a35319dc5d797"} Nov 21 15:09:17 crc kubenswrapper[4897]: I1121 15:09:17.966849 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5dqs7" Nov 21 15:09:17 crc kubenswrapper[4897]: I1121 15:09:17.967255 4897 scope.go:117] "RemoveContainer" containerID="2982c7fbdd5a9f455c8397a3bfddc92b9b293106c9ae1fb8d58a35319dc5d797" Nov 21 15:09:17 crc kubenswrapper[4897]: I1121 15:09:17.967126 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5dqs7" event={"ID":"61f3ca01-1e4b-4899-8ec3-25376d44695f","Type":"ContainerDied","Data":"e3e3a8c4a0f30d6532a7603c8c5123f01265c2809557ffa7868f470455d9dbd3"} Nov 21 15:09:17 crc kubenswrapper[4897]: I1121 15:09:17.999718 4897 scope.go:117] "RemoveContainer" containerID="1a2903e49cb98c8856ca389190b2d8c5316ae06ee2a2bbb5bf523fb22b1970da" Nov 21 15:09:18 crc kubenswrapper[4897]: I1121 15:09:18.002719 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5dqs7"] Nov 21 15:09:18 crc kubenswrapper[4897]: I1121 15:09:18.013437 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-5dqs7"] Nov 21 15:09:18 crc kubenswrapper[4897]: I1121 15:09:18.028194 4897 scope.go:117] "RemoveContainer" containerID="8c1a80fbe6cd446617866f25e7d331f466b6a6b3afcde7001884de53f9271d4b" Nov 21 15:09:18 crc kubenswrapper[4897]: I1121 15:09:18.088569 4897 scope.go:117] "RemoveContainer" containerID="2982c7fbdd5a9f455c8397a3bfddc92b9b293106c9ae1fb8d58a35319dc5d797" Nov 21 15:09:18 crc kubenswrapper[4897]: E1121 15:09:18.090650 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2982c7fbdd5a9f455c8397a3bfddc92b9b293106c9ae1fb8d58a35319dc5d797\": container with ID starting with 2982c7fbdd5a9f455c8397a3bfddc92b9b293106c9ae1fb8d58a35319dc5d797 not found: ID does not exist" containerID="2982c7fbdd5a9f455c8397a3bfddc92b9b293106c9ae1fb8d58a35319dc5d797" Nov 21 15:09:18 crc kubenswrapper[4897]: I1121 15:09:18.090707 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2982c7fbdd5a9f455c8397a3bfddc92b9b293106c9ae1fb8d58a35319dc5d797"} err="failed to get container status \"2982c7fbdd5a9f455c8397a3bfddc92b9b293106c9ae1fb8d58a35319dc5d797\": rpc error: code = NotFound desc = could not find container \"2982c7fbdd5a9f455c8397a3bfddc92b9b293106c9ae1fb8d58a35319dc5d797\": container with ID starting with 2982c7fbdd5a9f455c8397a3bfddc92b9b293106c9ae1fb8d58a35319dc5d797 not found: ID does not exist" Nov 21 15:09:18 crc kubenswrapper[4897]: I1121 15:09:18.090739 4897 scope.go:117] "RemoveContainer" containerID="1a2903e49cb98c8856ca389190b2d8c5316ae06ee2a2bbb5bf523fb22b1970da" Nov 21 15:09:18 crc kubenswrapper[4897]: E1121 15:09:18.091049 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a2903e49cb98c8856ca389190b2d8c5316ae06ee2a2bbb5bf523fb22b1970da\": container with ID starting with 1a2903e49cb98c8856ca389190b2d8c5316ae06ee2a2bbb5bf523fb22b1970da not found: ID does not exist" containerID="1a2903e49cb98c8856ca389190b2d8c5316ae06ee2a2bbb5bf523fb22b1970da" Nov 21 15:09:18 crc kubenswrapper[4897]: I1121 15:09:18.091083 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a2903e49cb98c8856ca389190b2d8c5316ae06ee2a2bbb5bf523fb22b1970da"} err="failed to get container status \"1a2903e49cb98c8856ca389190b2d8c5316ae06ee2a2bbb5bf523fb22b1970da\": rpc error: code = NotFound desc = could not find container \"1a2903e49cb98c8856ca389190b2d8c5316ae06ee2a2bbb5bf523fb22b1970da\": container with ID starting with 1a2903e49cb98c8856ca389190b2d8c5316ae06ee2a2bbb5bf523fb22b1970da not found: ID does not exist" Nov 21 15:09:18 crc kubenswrapper[4897]: I1121 15:09:18.091102 4897 scope.go:117] "RemoveContainer" containerID="8c1a80fbe6cd446617866f25e7d331f466b6a6b3afcde7001884de53f9271d4b" Nov 21 15:09:18 crc kubenswrapper[4897]: E1121 15:09:18.091344 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8c1a80fbe6cd446617866f25e7d331f466b6a6b3afcde7001884de53f9271d4b\": container with ID starting with 8c1a80fbe6cd446617866f25e7d331f466b6a6b3afcde7001884de53f9271d4b not found: ID does not exist" containerID="8c1a80fbe6cd446617866f25e7d331f466b6a6b3afcde7001884de53f9271d4b" Nov 21 15:09:18 crc kubenswrapper[4897]: I1121 15:09:18.091380 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8c1a80fbe6cd446617866f25e7d331f466b6a6b3afcde7001884de53f9271d4b"} err="failed to get container status \"8c1a80fbe6cd446617866f25e7d331f466b6a6b3afcde7001884de53f9271d4b\": rpc error: code = NotFound desc = could not find container \"8c1a80fbe6cd446617866f25e7d331f466b6a6b3afcde7001884de53f9271d4b\": container with ID starting with 8c1a80fbe6cd446617866f25e7d331f466b6a6b3afcde7001884de53f9271d4b not found: ID does not exist" Nov 21 15:09:18 crc kubenswrapper[4897]: I1121 15:09:18.104724 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="61f3ca01-1e4b-4899-8ec3-25376d44695f" path="/var/lib/kubelet/pods/61f3ca01-1e4b-4899-8ec3-25376d44695f/volumes" Nov 21 15:09:34 crc kubenswrapper[4897]: I1121 15:09:34.370899 4897 patch_prober.go:28] interesting pod/machine-config-daemon-krv5b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 21 15:09:34 crc kubenswrapper[4897]: I1121 15:09:34.372372 4897 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 21 15:09:34 crc kubenswrapper[4897]: I1121 15:09:34.372560 4897 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" Nov 21 15:09:34 crc kubenswrapper[4897]: I1121 15:09:34.373268 4897 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"31498b4cf1d12960e0bd3369fd9c8702f2e4dc6c2b6a5487ecab62d136f4bc3c"} pod="openshift-machine-config-operator/machine-config-daemon-krv5b" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 21 15:09:34 crc kubenswrapper[4897]: I1121 15:09:34.373397 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" containerName="machine-config-daemon" containerID="cri-o://31498b4cf1d12960e0bd3369fd9c8702f2e4dc6c2b6a5487ecab62d136f4bc3c" gracePeriod=600 Nov 21 15:09:35 crc kubenswrapper[4897]: E1121 15:09:35.060334 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:09:35 crc kubenswrapper[4897]: I1121 15:09:35.214116 4897 generic.go:334] "Generic (PLEG): container finished" podID="e7670227-d280-4847-b882-754429f56b0f" containerID="31498b4cf1d12960e0bd3369fd9c8702f2e4dc6c2b6a5487ecab62d136f4bc3c" exitCode=0 Nov 21 15:09:35 crc kubenswrapper[4897]: I1121 15:09:35.214165 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" event={"ID":"e7670227-d280-4847-b882-754429f56b0f","Type":"ContainerDied","Data":"31498b4cf1d12960e0bd3369fd9c8702f2e4dc6c2b6a5487ecab62d136f4bc3c"} Nov 21 15:09:35 crc kubenswrapper[4897]: I1121 15:09:35.214198 4897 scope.go:117] "RemoveContainer" containerID="11456ed300543798476a9b68095120ef6d6073c876d338f9212ec0e3865c7a1a" Nov 21 15:09:35 crc kubenswrapper[4897]: I1121 15:09:35.214958 4897 scope.go:117] "RemoveContainer" containerID="31498b4cf1d12960e0bd3369fd9c8702f2e4dc6c2b6a5487ecab62d136f4bc3c" Nov 21 15:09:35 crc kubenswrapper[4897]: E1121 15:09:35.215250 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:09:49 crc kubenswrapper[4897]: I1121 15:09:49.090672 4897 scope.go:117] "RemoveContainer" containerID="31498b4cf1d12960e0bd3369fd9c8702f2e4dc6c2b6a5487ecab62d136f4bc3c" Nov 21 15:09:49 crc kubenswrapper[4897]: E1121 15:09:49.092142 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:10:02 crc kubenswrapper[4897]: I1121 15:10:02.089928 4897 scope.go:117] "RemoveContainer" containerID="31498b4cf1d12960e0bd3369fd9c8702f2e4dc6c2b6a5487ecab62d136f4bc3c" Nov 21 15:10:02 crc kubenswrapper[4897]: E1121 15:10:02.090877 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:10:16 crc kubenswrapper[4897]: I1121 15:10:16.091079 4897 scope.go:117] "RemoveContainer" containerID="31498b4cf1d12960e0bd3369fd9c8702f2e4dc6c2b6a5487ecab62d136f4bc3c" Nov 21 15:10:16 crc kubenswrapper[4897]: E1121 15:10:16.092240 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:10:31 crc kubenswrapper[4897]: I1121 15:10:31.090441 4897 scope.go:117] "RemoveContainer" containerID="31498b4cf1d12960e0bd3369fd9c8702f2e4dc6c2b6a5487ecab62d136f4bc3c" Nov 21 15:10:31 crc kubenswrapper[4897]: E1121 15:10:31.091875 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:10:32 crc kubenswrapper[4897]: I1121 15:10:32.856944 4897 generic.go:334] "Generic (PLEG): container finished" podID="ae0e489e-6389-4ae0-947d-c75280d74e38" containerID="952f057da4348c5e7c08a69819932881063626e548d4d32cf78a82a4077f83cc" exitCode=0 Nov 21 15:10:32 crc kubenswrapper[4897]: I1121 15:10:32.858043 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-n2lpc" event={"ID":"ae0e489e-6389-4ae0-947d-c75280d74e38","Type":"ContainerDied","Data":"952f057da4348c5e7c08a69819932881063626e548d4d32cf78a82a4077f83cc"} Nov 21 15:10:34 crc kubenswrapper[4897]: I1121 15:10:34.373908 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-n2lpc" Nov 21 15:10:34 crc kubenswrapper[4897]: I1121 15:10:34.540021 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ae0e489e-6389-4ae0-947d-c75280d74e38-inventory\") pod \"ae0e489e-6389-4ae0-947d-c75280d74e38\" (UID: \"ae0e489e-6389-4ae0-947d-c75280d74e38\") " Nov 21 15:10:34 crc kubenswrapper[4897]: I1121 15:10:34.540073 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-ipmi-config-data-2\" (UniqueName: \"kubernetes.io/secret/ae0e489e-6389-4ae0-947d-c75280d74e38-ceilometer-ipmi-config-data-2\") pod \"ae0e489e-6389-4ae0-947d-c75280d74e38\" (UID: \"ae0e489e-6389-4ae0-947d-c75280d74e38\") " Nov 21 15:10:34 crc kubenswrapper[4897]: I1121 15:10:34.540223 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae0e489e-6389-4ae0-947d-c75280d74e38-telemetry-power-monitoring-combined-ca-bundle\") pod \"ae0e489e-6389-4ae0-947d-c75280d74e38\" (UID: \"ae0e489e-6389-4ae0-947d-c75280d74e38\") " Nov 21 15:10:34 crc kubenswrapper[4897]: I1121 15:10:34.540369 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-ipmi-config-data-0\" (UniqueName: \"kubernetes.io/secret/ae0e489e-6389-4ae0-947d-c75280d74e38-ceilometer-ipmi-config-data-0\") pod \"ae0e489e-6389-4ae0-947d-c75280d74e38\" (UID: \"ae0e489e-6389-4ae0-947d-c75280d74e38\") " Nov 21 15:10:34 crc kubenswrapper[4897]: I1121 15:10:34.540401 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8d2hj\" (UniqueName: \"kubernetes.io/projected/ae0e489e-6389-4ae0-947d-c75280d74e38-kube-api-access-8d2hj\") pod \"ae0e489e-6389-4ae0-947d-c75280d74e38\" (UID: \"ae0e489e-6389-4ae0-947d-c75280d74e38\") " Nov 21 15:10:34 crc kubenswrapper[4897]: I1121 15:10:34.540421 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-ipmi-config-data-1\" (UniqueName: \"kubernetes.io/secret/ae0e489e-6389-4ae0-947d-c75280d74e38-ceilometer-ipmi-config-data-1\") pod \"ae0e489e-6389-4ae0-947d-c75280d74e38\" (UID: \"ae0e489e-6389-4ae0-947d-c75280d74e38\") " Nov 21 15:10:34 crc kubenswrapper[4897]: I1121 15:10:34.540476 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ae0e489e-6389-4ae0-947d-c75280d74e38-ssh-key\") pod \"ae0e489e-6389-4ae0-947d-c75280d74e38\" (UID: \"ae0e489e-6389-4ae0-947d-c75280d74e38\") " Nov 21 15:10:34 crc kubenswrapper[4897]: I1121 15:10:34.547888 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae0e489e-6389-4ae0-947d-c75280d74e38-telemetry-power-monitoring-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-power-monitoring-combined-ca-bundle") pod "ae0e489e-6389-4ae0-947d-c75280d74e38" (UID: "ae0e489e-6389-4ae0-947d-c75280d74e38"). InnerVolumeSpecName "telemetry-power-monitoring-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 15:10:34 crc kubenswrapper[4897]: I1121 15:10:34.548455 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae0e489e-6389-4ae0-947d-c75280d74e38-kube-api-access-8d2hj" (OuterVolumeSpecName: "kube-api-access-8d2hj") pod "ae0e489e-6389-4ae0-947d-c75280d74e38" (UID: "ae0e489e-6389-4ae0-947d-c75280d74e38"). InnerVolumeSpecName "kube-api-access-8d2hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 15:10:34 crc kubenswrapper[4897]: I1121 15:10:34.580872 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae0e489e-6389-4ae0-947d-c75280d74e38-inventory" (OuterVolumeSpecName: "inventory") pod "ae0e489e-6389-4ae0-947d-c75280d74e38" (UID: "ae0e489e-6389-4ae0-947d-c75280d74e38"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 15:10:34 crc kubenswrapper[4897]: I1121 15:10:34.581380 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae0e489e-6389-4ae0-947d-c75280d74e38-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ae0e489e-6389-4ae0-947d-c75280d74e38" (UID: "ae0e489e-6389-4ae0-947d-c75280d74e38"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 15:10:34 crc kubenswrapper[4897]: I1121 15:10:34.586798 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae0e489e-6389-4ae0-947d-c75280d74e38-ceilometer-ipmi-config-data-2" (OuterVolumeSpecName: "ceilometer-ipmi-config-data-2") pod "ae0e489e-6389-4ae0-947d-c75280d74e38" (UID: "ae0e489e-6389-4ae0-947d-c75280d74e38"). InnerVolumeSpecName "ceilometer-ipmi-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 15:10:34 crc kubenswrapper[4897]: I1121 15:10:34.591416 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae0e489e-6389-4ae0-947d-c75280d74e38-ceilometer-ipmi-config-data-1" (OuterVolumeSpecName: "ceilometer-ipmi-config-data-1") pod "ae0e489e-6389-4ae0-947d-c75280d74e38" (UID: "ae0e489e-6389-4ae0-947d-c75280d74e38"). InnerVolumeSpecName "ceilometer-ipmi-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 15:10:34 crc kubenswrapper[4897]: I1121 15:10:34.601773 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae0e489e-6389-4ae0-947d-c75280d74e38-ceilometer-ipmi-config-data-0" (OuterVolumeSpecName: "ceilometer-ipmi-config-data-0") pod "ae0e489e-6389-4ae0-947d-c75280d74e38" (UID: "ae0e489e-6389-4ae0-947d-c75280d74e38"). InnerVolumeSpecName "ceilometer-ipmi-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 15:10:34 crc kubenswrapper[4897]: I1121 15:10:34.643471 4897 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ae0e489e-6389-4ae0-947d-c75280d74e38-inventory\") on node \"crc\" DevicePath \"\"" Nov 21 15:10:34 crc kubenswrapper[4897]: I1121 15:10:34.643585 4897 reconciler_common.go:293] "Volume detached for volume \"ceilometer-ipmi-config-data-2\" (UniqueName: \"kubernetes.io/secret/ae0e489e-6389-4ae0-947d-c75280d74e38-ceilometer-ipmi-config-data-2\") on node \"crc\" DevicePath \"\"" Nov 21 15:10:34 crc kubenswrapper[4897]: I1121 15:10:34.643600 4897 reconciler_common.go:293] "Volume detached for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae0e489e-6389-4ae0-947d-c75280d74e38-telemetry-power-monitoring-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 21 15:10:34 crc kubenswrapper[4897]: I1121 15:10:34.643610 4897 reconciler_common.go:293] "Volume detached for volume \"ceilometer-ipmi-config-data-0\" (UniqueName: \"kubernetes.io/secret/ae0e489e-6389-4ae0-947d-c75280d74e38-ceilometer-ipmi-config-data-0\") on node \"crc\" DevicePath \"\"" Nov 21 15:10:34 crc kubenswrapper[4897]: I1121 15:10:34.643621 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8d2hj\" (UniqueName: \"kubernetes.io/projected/ae0e489e-6389-4ae0-947d-c75280d74e38-kube-api-access-8d2hj\") on node \"crc\" DevicePath \"\"" Nov 21 15:10:34 crc kubenswrapper[4897]: I1121 15:10:34.643632 4897 reconciler_common.go:293] "Volume detached for volume \"ceilometer-ipmi-config-data-1\" (UniqueName: \"kubernetes.io/secret/ae0e489e-6389-4ae0-947d-c75280d74e38-ceilometer-ipmi-config-data-1\") on node \"crc\" DevicePath \"\"" Nov 21 15:10:34 crc kubenswrapper[4897]: I1121 15:10:34.643640 4897 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ae0e489e-6389-4ae0-947d-c75280d74e38-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 21 15:10:34 crc kubenswrapper[4897]: I1121 15:10:34.882233 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-n2lpc" event={"ID":"ae0e489e-6389-4ae0-947d-c75280d74e38","Type":"ContainerDied","Data":"fd5c3bf8cd40748e4b49be3be91910ef4569b0fabcdb4e5c31afa2f18decf3f2"} Nov 21 15:10:34 crc kubenswrapper[4897]: I1121 15:10:34.883055 4897 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fd5c3bf8cd40748e4b49be3be91910ef4569b0fabcdb4e5c31afa2f18decf3f2" Nov 21 15:10:34 crc kubenswrapper[4897]: I1121 15:10:34.882329 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-n2lpc" Nov 21 15:10:34 crc kubenswrapper[4897]: I1121 15:10:34.961455 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/logging-edpm-deployment-openstack-edpm-ipam-2d6cb"] Nov 21 15:10:34 crc kubenswrapper[4897]: E1121 15:10:34.962259 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61f3ca01-1e4b-4899-8ec3-25376d44695f" containerName="registry-server" Nov 21 15:10:34 crc kubenswrapper[4897]: I1121 15:10:34.962282 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="61f3ca01-1e4b-4899-8ec3-25376d44695f" containerName="registry-server" Nov 21 15:10:34 crc kubenswrapper[4897]: E1121 15:10:34.962327 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae0e489e-6389-4ae0-947d-c75280d74e38" containerName="telemetry-power-monitoring-edpm-deployment-openstack-edpm-ipam" Nov 21 15:10:34 crc kubenswrapper[4897]: I1121 15:10:34.962338 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae0e489e-6389-4ae0-947d-c75280d74e38" containerName="telemetry-power-monitoring-edpm-deployment-openstack-edpm-ipam" Nov 21 15:10:34 crc kubenswrapper[4897]: E1121 15:10:34.962383 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61f3ca01-1e4b-4899-8ec3-25376d44695f" containerName="extract-utilities" Nov 21 15:10:34 crc kubenswrapper[4897]: I1121 15:10:34.962392 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="61f3ca01-1e4b-4899-8ec3-25376d44695f" containerName="extract-utilities" Nov 21 15:10:34 crc kubenswrapper[4897]: E1121 15:10:34.962410 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61f3ca01-1e4b-4899-8ec3-25376d44695f" containerName="extract-content" Nov 21 15:10:34 crc kubenswrapper[4897]: I1121 15:10:34.962417 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="61f3ca01-1e4b-4899-8ec3-25376d44695f" containerName="extract-content" Nov 21 15:10:34 crc kubenswrapper[4897]: I1121 15:10:34.962702 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="61f3ca01-1e4b-4899-8ec3-25376d44695f" containerName="registry-server" Nov 21 15:10:34 crc kubenswrapper[4897]: I1121 15:10:34.962743 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae0e489e-6389-4ae0-947d-c75280d74e38" containerName="telemetry-power-monitoring-edpm-deployment-openstack-edpm-ipam" Nov 21 15:10:34 crc kubenswrapper[4897]: I1121 15:10:34.963887 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-2d6cb" Nov 21 15:10:34 crc kubenswrapper[4897]: I1121 15:10:34.967717 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 21 15:10:34 crc kubenswrapper[4897]: I1121 15:10:34.967953 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"logging-compute-config-data" Nov 21 15:10:34 crc kubenswrapper[4897]: I1121 15:10:34.968239 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 21 15:10:34 crc kubenswrapper[4897]: I1121 15:10:34.968401 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 21 15:10:34 crc kubenswrapper[4897]: I1121 15:10:34.968590 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-dj88m" Nov 21 15:10:34 crc kubenswrapper[4897]: I1121 15:10:34.978697 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/logging-edpm-deployment-openstack-edpm-ipam-2d6cb"] Nov 21 15:10:35 crc kubenswrapper[4897]: I1121 15:10:35.155294 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/0eadbb7f-062d-4d47-9498-f9bf05beae50-logging-compute-config-data-0\") pod \"logging-edpm-deployment-openstack-edpm-ipam-2d6cb\" (UID: \"0eadbb7f-062d-4d47-9498-f9bf05beae50\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-2d6cb" Nov 21 15:10:35 crc kubenswrapper[4897]: I1121 15:10:35.155740 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7xg9s\" (UniqueName: \"kubernetes.io/projected/0eadbb7f-062d-4d47-9498-f9bf05beae50-kube-api-access-7xg9s\") pod \"logging-edpm-deployment-openstack-edpm-ipam-2d6cb\" (UID: \"0eadbb7f-062d-4d47-9498-f9bf05beae50\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-2d6cb" Nov 21 15:10:35 crc kubenswrapper[4897]: I1121 15:10:35.156010 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/0eadbb7f-062d-4d47-9498-f9bf05beae50-logging-compute-config-data-1\") pod \"logging-edpm-deployment-openstack-edpm-ipam-2d6cb\" (UID: \"0eadbb7f-062d-4d47-9498-f9bf05beae50\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-2d6cb" Nov 21 15:10:35 crc kubenswrapper[4897]: I1121 15:10:35.156389 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0eadbb7f-062d-4d47-9498-f9bf05beae50-ssh-key\") pod \"logging-edpm-deployment-openstack-edpm-ipam-2d6cb\" (UID: \"0eadbb7f-062d-4d47-9498-f9bf05beae50\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-2d6cb" Nov 21 15:10:35 crc kubenswrapper[4897]: I1121 15:10:35.156712 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0eadbb7f-062d-4d47-9498-f9bf05beae50-inventory\") pod \"logging-edpm-deployment-openstack-edpm-ipam-2d6cb\" (UID: \"0eadbb7f-062d-4d47-9498-f9bf05beae50\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-2d6cb" Nov 21 15:10:35 crc kubenswrapper[4897]: I1121 15:10:35.259145 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0eadbb7f-062d-4d47-9498-f9bf05beae50-ssh-key\") pod \"logging-edpm-deployment-openstack-edpm-ipam-2d6cb\" (UID: \"0eadbb7f-062d-4d47-9498-f9bf05beae50\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-2d6cb" Nov 21 15:10:35 crc kubenswrapper[4897]: I1121 15:10:35.260156 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0eadbb7f-062d-4d47-9498-f9bf05beae50-inventory\") pod \"logging-edpm-deployment-openstack-edpm-ipam-2d6cb\" (UID: \"0eadbb7f-062d-4d47-9498-f9bf05beae50\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-2d6cb" Nov 21 15:10:35 crc kubenswrapper[4897]: I1121 15:10:35.260403 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/0eadbb7f-062d-4d47-9498-f9bf05beae50-logging-compute-config-data-0\") pod \"logging-edpm-deployment-openstack-edpm-ipam-2d6cb\" (UID: \"0eadbb7f-062d-4d47-9498-f9bf05beae50\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-2d6cb" Nov 21 15:10:35 crc kubenswrapper[4897]: I1121 15:10:35.260542 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7xg9s\" (UniqueName: \"kubernetes.io/projected/0eadbb7f-062d-4d47-9498-f9bf05beae50-kube-api-access-7xg9s\") pod \"logging-edpm-deployment-openstack-edpm-ipam-2d6cb\" (UID: \"0eadbb7f-062d-4d47-9498-f9bf05beae50\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-2d6cb" Nov 21 15:10:35 crc kubenswrapper[4897]: I1121 15:10:35.260966 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/0eadbb7f-062d-4d47-9498-f9bf05beae50-logging-compute-config-data-1\") pod \"logging-edpm-deployment-openstack-edpm-ipam-2d6cb\" (UID: \"0eadbb7f-062d-4d47-9498-f9bf05beae50\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-2d6cb" Nov 21 15:10:35 crc kubenswrapper[4897]: I1121 15:10:35.264473 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/0eadbb7f-062d-4d47-9498-f9bf05beae50-logging-compute-config-data-1\") pod \"logging-edpm-deployment-openstack-edpm-ipam-2d6cb\" (UID: \"0eadbb7f-062d-4d47-9498-f9bf05beae50\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-2d6cb" Nov 21 15:10:35 crc kubenswrapper[4897]: I1121 15:10:35.264667 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0eadbb7f-062d-4d47-9498-f9bf05beae50-ssh-key\") pod \"logging-edpm-deployment-openstack-edpm-ipam-2d6cb\" (UID: \"0eadbb7f-062d-4d47-9498-f9bf05beae50\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-2d6cb" Nov 21 15:10:35 crc kubenswrapper[4897]: I1121 15:10:35.265151 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0eadbb7f-062d-4d47-9498-f9bf05beae50-inventory\") pod \"logging-edpm-deployment-openstack-edpm-ipam-2d6cb\" (UID: \"0eadbb7f-062d-4d47-9498-f9bf05beae50\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-2d6cb" Nov 21 15:10:35 crc kubenswrapper[4897]: I1121 15:10:35.265284 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/0eadbb7f-062d-4d47-9498-f9bf05beae50-logging-compute-config-data-0\") pod \"logging-edpm-deployment-openstack-edpm-ipam-2d6cb\" (UID: \"0eadbb7f-062d-4d47-9498-f9bf05beae50\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-2d6cb" Nov 21 15:10:35 crc kubenswrapper[4897]: I1121 15:10:35.282415 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7xg9s\" (UniqueName: \"kubernetes.io/projected/0eadbb7f-062d-4d47-9498-f9bf05beae50-kube-api-access-7xg9s\") pod \"logging-edpm-deployment-openstack-edpm-ipam-2d6cb\" (UID: \"0eadbb7f-062d-4d47-9498-f9bf05beae50\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-2d6cb" Nov 21 15:10:35 crc kubenswrapper[4897]: I1121 15:10:35.581780 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-2d6cb" Nov 21 15:10:36 crc kubenswrapper[4897]: I1121 15:10:36.130812 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/logging-edpm-deployment-openstack-edpm-ipam-2d6cb"] Nov 21 15:10:36 crc kubenswrapper[4897]: I1121 15:10:36.903380 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-2d6cb" event={"ID":"0eadbb7f-062d-4d47-9498-f9bf05beae50","Type":"ContainerStarted","Data":"80b97a35dd26a8c9e6fb4f7f9dbbfa5184dc584ff682c71303898d70c2516675"} Nov 21 15:10:37 crc kubenswrapper[4897]: I1121 15:10:37.915981 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-2d6cb" event={"ID":"0eadbb7f-062d-4d47-9498-f9bf05beae50","Type":"ContainerStarted","Data":"0b7f66dc71cd7ac8232650f9552abd36f238c2919f4587d228230d05fd21ed2a"} Nov 21 15:10:37 crc kubenswrapper[4897]: I1121 15:10:37.937676 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-2d6cb" podStartSLOduration=3.324866705 podStartE2EDuration="3.937655022s" podCreationTimestamp="2025-11-21 15:10:34 +0000 UTC" firstStartedPulling="2025-11-21 15:10:36.136832968 +0000 UTC m=+3713.421426443" lastFinishedPulling="2025-11-21 15:10:36.749621285 +0000 UTC m=+3714.034214760" observedRunningTime="2025-11-21 15:10:37.931320461 +0000 UTC m=+3715.215913956" watchObservedRunningTime="2025-11-21 15:10:37.937655022 +0000 UTC m=+3715.222248497" Nov 21 15:10:42 crc kubenswrapper[4897]: I1121 15:10:42.089351 4897 scope.go:117] "RemoveContainer" containerID="31498b4cf1d12960e0bd3369fd9c8702f2e4dc6c2b6a5487ecab62d136f4bc3c" Nov 21 15:10:42 crc kubenswrapper[4897]: E1121 15:10:42.090217 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:10:52 crc kubenswrapper[4897]: I1121 15:10:52.080154 4897 generic.go:334] "Generic (PLEG): container finished" podID="0eadbb7f-062d-4d47-9498-f9bf05beae50" containerID="0b7f66dc71cd7ac8232650f9552abd36f238c2919f4587d228230d05fd21ed2a" exitCode=0 Nov 21 15:10:52 crc kubenswrapper[4897]: I1121 15:10:52.080246 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-2d6cb" event={"ID":"0eadbb7f-062d-4d47-9498-f9bf05beae50","Type":"ContainerDied","Data":"0b7f66dc71cd7ac8232650f9552abd36f238c2919f4587d228230d05fd21ed2a"} Nov 21 15:10:53 crc kubenswrapper[4897]: I1121 15:10:53.611534 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-2d6cb" Nov 21 15:10:53 crc kubenswrapper[4897]: I1121 15:10:53.648277 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7xg9s\" (UniqueName: \"kubernetes.io/projected/0eadbb7f-062d-4d47-9498-f9bf05beae50-kube-api-access-7xg9s\") pod \"0eadbb7f-062d-4d47-9498-f9bf05beae50\" (UID: \"0eadbb7f-062d-4d47-9498-f9bf05beae50\") " Nov 21 15:10:53 crc kubenswrapper[4897]: I1121 15:10:53.648424 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/0eadbb7f-062d-4d47-9498-f9bf05beae50-logging-compute-config-data-1\") pod \"0eadbb7f-062d-4d47-9498-f9bf05beae50\" (UID: \"0eadbb7f-062d-4d47-9498-f9bf05beae50\") " Nov 21 15:10:53 crc kubenswrapper[4897]: I1121 15:10:53.648619 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0eadbb7f-062d-4d47-9498-f9bf05beae50-inventory\") pod \"0eadbb7f-062d-4d47-9498-f9bf05beae50\" (UID: \"0eadbb7f-062d-4d47-9498-f9bf05beae50\") " Nov 21 15:10:53 crc kubenswrapper[4897]: I1121 15:10:53.648751 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/0eadbb7f-062d-4d47-9498-f9bf05beae50-logging-compute-config-data-0\") pod \"0eadbb7f-062d-4d47-9498-f9bf05beae50\" (UID: \"0eadbb7f-062d-4d47-9498-f9bf05beae50\") " Nov 21 15:10:53 crc kubenswrapper[4897]: I1121 15:10:53.648854 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0eadbb7f-062d-4d47-9498-f9bf05beae50-ssh-key\") pod \"0eadbb7f-062d-4d47-9498-f9bf05beae50\" (UID: \"0eadbb7f-062d-4d47-9498-f9bf05beae50\") " Nov 21 15:10:53 crc kubenswrapper[4897]: I1121 15:10:53.663269 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0eadbb7f-062d-4d47-9498-f9bf05beae50-kube-api-access-7xg9s" (OuterVolumeSpecName: "kube-api-access-7xg9s") pod "0eadbb7f-062d-4d47-9498-f9bf05beae50" (UID: "0eadbb7f-062d-4d47-9498-f9bf05beae50"). InnerVolumeSpecName "kube-api-access-7xg9s". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 15:10:53 crc kubenswrapper[4897]: I1121 15:10:53.690677 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0eadbb7f-062d-4d47-9498-f9bf05beae50-logging-compute-config-data-1" (OuterVolumeSpecName: "logging-compute-config-data-1") pod "0eadbb7f-062d-4d47-9498-f9bf05beae50" (UID: "0eadbb7f-062d-4d47-9498-f9bf05beae50"). InnerVolumeSpecName "logging-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 15:10:53 crc kubenswrapper[4897]: I1121 15:10:53.693673 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0eadbb7f-062d-4d47-9498-f9bf05beae50-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "0eadbb7f-062d-4d47-9498-f9bf05beae50" (UID: "0eadbb7f-062d-4d47-9498-f9bf05beae50"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 15:10:53 crc kubenswrapper[4897]: I1121 15:10:53.699524 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0eadbb7f-062d-4d47-9498-f9bf05beae50-logging-compute-config-data-0" (OuterVolumeSpecName: "logging-compute-config-data-0") pod "0eadbb7f-062d-4d47-9498-f9bf05beae50" (UID: "0eadbb7f-062d-4d47-9498-f9bf05beae50"). InnerVolumeSpecName "logging-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 15:10:53 crc kubenswrapper[4897]: I1121 15:10:53.701052 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0eadbb7f-062d-4d47-9498-f9bf05beae50-inventory" (OuterVolumeSpecName: "inventory") pod "0eadbb7f-062d-4d47-9498-f9bf05beae50" (UID: "0eadbb7f-062d-4d47-9498-f9bf05beae50"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 15:10:53 crc kubenswrapper[4897]: I1121 15:10:53.753909 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7xg9s\" (UniqueName: \"kubernetes.io/projected/0eadbb7f-062d-4d47-9498-f9bf05beae50-kube-api-access-7xg9s\") on node \"crc\" DevicePath \"\"" Nov 21 15:10:53 crc kubenswrapper[4897]: I1121 15:10:53.753953 4897 reconciler_common.go:293] "Volume detached for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/0eadbb7f-062d-4d47-9498-f9bf05beae50-logging-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Nov 21 15:10:53 crc kubenswrapper[4897]: I1121 15:10:53.753966 4897 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0eadbb7f-062d-4d47-9498-f9bf05beae50-inventory\") on node \"crc\" DevicePath \"\"" Nov 21 15:10:53 crc kubenswrapper[4897]: I1121 15:10:53.753978 4897 reconciler_common.go:293] "Volume detached for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/0eadbb7f-062d-4d47-9498-f9bf05beae50-logging-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Nov 21 15:10:53 crc kubenswrapper[4897]: I1121 15:10:53.753987 4897 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0eadbb7f-062d-4d47-9498-f9bf05beae50-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 21 15:10:54 crc kubenswrapper[4897]: I1121 15:10:54.110340 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-2d6cb" event={"ID":"0eadbb7f-062d-4d47-9498-f9bf05beae50","Type":"ContainerDied","Data":"80b97a35dd26a8c9e6fb4f7f9dbbfa5184dc584ff682c71303898d70c2516675"} Nov 21 15:10:54 crc kubenswrapper[4897]: I1121 15:10:54.110872 4897 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="80b97a35dd26a8c9e6fb4f7f9dbbfa5184dc584ff682c71303898d70c2516675" Nov 21 15:10:54 crc kubenswrapper[4897]: I1121 15:10:54.110754 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-2d6cb" Nov 21 15:10:56 crc kubenswrapper[4897]: I1121 15:10:56.089703 4897 scope.go:117] "RemoveContainer" containerID="31498b4cf1d12960e0bd3369fd9c8702f2e4dc6c2b6a5487ecab62d136f4bc3c" Nov 21 15:10:56 crc kubenswrapper[4897]: E1121 15:10:56.090361 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:11:09 crc kubenswrapper[4897]: I1121 15:11:09.088794 4897 scope.go:117] "RemoveContainer" containerID="31498b4cf1d12960e0bd3369fd9c8702f2e4dc6c2b6a5487ecab62d136f4bc3c" Nov 21 15:11:09 crc kubenswrapper[4897]: E1121 15:11:09.089503 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:11:21 crc kubenswrapper[4897]: I1121 15:11:21.090007 4897 scope.go:117] "RemoveContainer" containerID="31498b4cf1d12960e0bd3369fd9c8702f2e4dc6c2b6a5487ecab62d136f4bc3c" Nov 21 15:11:21 crc kubenswrapper[4897]: E1121 15:11:21.091181 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:11:36 crc kubenswrapper[4897]: I1121 15:11:36.089764 4897 scope.go:117] "RemoveContainer" containerID="31498b4cf1d12960e0bd3369fd9c8702f2e4dc6c2b6a5487ecab62d136f4bc3c" Nov 21 15:11:36 crc kubenswrapper[4897]: E1121 15:11:36.090741 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:11:47 crc kubenswrapper[4897]: I1121 15:11:47.089496 4897 scope.go:117] "RemoveContainer" containerID="31498b4cf1d12960e0bd3369fd9c8702f2e4dc6c2b6a5487ecab62d136f4bc3c" Nov 21 15:11:47 crc kubenswrapper[4897]: E1121 15:11:47.090139 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:11:58 crc kubenswrapper[4897]: I1121 15:11:58.089237 4897 scope.go:117] "RemoveContainer" containerID="31498b4cf1d12960e0bd3369fd9c8702f2e4dc6c2b6a5487ecab62d136f4bc3c" Nov 21 15:11:58 crc kubenswrapper[4897]: E1121 15:11:58.090028 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:12:13 crc kubenswrapper[4897]: I1121 15:12:13.089797 4897 scope.go:117] "RemoveContainer" containerID="31498b4cf1d12960e0bd3369fd9c8702f2e4dc6c2b6a5487ecab62d136f4bc3c" Nov 21 15:12:13 crc kubenswrapper[4897]: E1121 15:12:13.090700 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:12:27 crc kubenswrapper[4897]: I1121 15:12:27.089568 4897 scope.go:117] "RemoveContainer" containerID="31498b4cf1d12960e0bd3369fd9c8702f2e4dc6c2b6a5487ecab62d136f4bc3c" Nov 21 15:12:27 crc kubenswrapper[4897]: E1121 15:12:27.090438 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:12:40 crc kubenswrapper[4897]: I1121 15:12:40.089396 4897 scope.go:117] "RemoveContainer" containerID="31498b4cf1d12960e0bd3369fd9c8702f2e4dc6c2b6a5487ecab62d136f4bc3c" Nov 21 15:12:40 crc kubenswrapper[4897]: E1121 15:12:40.090115 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:12:55 crc kubenswrapper[4897]: I1121 15:12:55.090241 4897 scope.go:117] "RemoveContainer" containerID="31498b4cf1d12960e0bd3369fd9c8702f2e4dc6c2b6a5487ecab62d136f4bc3c" Nov 21 15:12:55 crc kubenswrapper[4897]: E1121 15:12:55.093346 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:12:58 crc kubenswrapper[4897]: I1121 15:12:58.737693 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-kb62c"] Nov 21 15:12:58 crc kubenswrapper[4897]: E1121 15:12:58.739030 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0eadbb7f-062d-4d47-9498-f9bf05beae50" containerName="logging-edpm-deployment-openstack-edpm-ipam" Nov 21 15:12:58 crc kubenswrapper[4897]: I1121 15:12:58.739049 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="0eadbb7f-062d-4d47-9498-f9bf05beae50" containerName="logging-edpm-deployment-openstack-edpm-ipam" Nov 21 15:12:58 crc kubenswrapper[4897]: I1121 15:12:58.739391 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="0eadbb7f-062d-4d47-9498-f9bf05beae50" containerName="logging-edpm-deployment-openstack-edpm-ipam" Nov 21 15:12:58 crc kubenswrapper[4897]: I1121 15:12:58.741364 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kb62c" Nov 21 15:12:58 crc kubenswrapper[4897]: I1121 15:12:58.750999 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kb62c"] Nov 21 15:12:58 crc kubenswrapper[4897]: I1121 15:12:58.885058 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tr7z5\" (UniqueName: \"kubernetes.io/projected/3a529e65-6ca1-4cf5-8894-a4f8eb7fa44b-kube-api-access-tr7z5\") pod \"community-operators-kb62c\" (UID: \"3a529e65-6ca1-4cf5-8894-a4f8eb7fa44b\") " pod="openshift-marketplace/community-operators-kb62c" Nov 21 15:12:58 crc kubenswrapper[4897]: I1121 15:12:58.885255 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a529e65-6ca1-4cf5-8894-a4f8eb7fa44b-utilities\") pod \"community-operators-kb62c\" (UID: \"3a529e65-6ca1-4cf5-8894-a4f8eb7fa44b\") " pod="openshift-marketplace/community-operators-kb62c" Nov 21 15:12:58 crc kubenswrapper[4897]: I1121 15:12:58.885700 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a529e65-6ca1-4cf5-8894-a4f8eb7fa44b-catalog-content\") pod \"community-operators-kb62c\" (UID: \"3a529e65-6ca1-4cf5-8894-a4f8eb7fa44b\") " pod="openshift-marketplace/community-operators-kb62c" Nov 21 15:12:58 crc kubenswrapper[4897]: I1121 15:12:58.988291 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a529e65-6ca1-4cf5-8894-a4f8eb7fa44b-catalog-content\") pod \"community-operators-kb62c\" (UID: \"3a529e65-6ca1-4cf5-8894-a4f8eb7fa44b\") " pod="openshift-marketplace/community-operators-kb62c" Nov 21 15:12:58 crc kubenswrapper[4897]: I1121 15:12:58.988423 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tr7z5\" (UniqueName: \"kubernetes.io/projected/3a529e65-6ca1-4cf5-8894-a4f8eb7fa44b-kube-api-access-tr7z5\") pod \"community-operators-kb62c\" (UID: \"3a529e65-6ca1-4cf5-8894-a4f8eb7fa44b\") " pod="openshift-marketplace/community-operators-kb62c" Nov 21 15:12:58 crc kubenswrapper[4897]: I1121 15:12:58.988473 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a529e65-6ca1-4cf5-8894-a4f8eb7fa44b-utilities\") pod \"community-operators-kb62c\" (UID: \"3a529e65-6ca1-4cf5-8894-a4f8eb7fa44b\") " pod="openshift-marketplace/community-operators-kb62c" Nov 21 15:12:58 crc kubenswrapper[4897]: I1121 15:12:58.988907 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a529e65-6ca1-4cf5-8894-a4f8eb7fa44b-catalog-content\") pod \"community-operators-kb62c\" (UID: \"3a529e65-6ca1-4cf5-8894-a4f8eb7fa44b\") " pod="openshift-marketplace/community-operators-kb62c" Nov 21 15:12:58 crc kubenswrapper[4897]: I1121 15:12:58.988947 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a529e65-6ca1-4cf5-8894-a4f8eb7fa44b-utilities\") pod \"community-operators-kb62c\" (UID: \"3a529e65-6ca1-4cf5-8894-a4f8eb7fa44b\") " pod="openshift-marketplace/community-operators-kb62c" Nov 21 15:12:59 crc kubenswrapper[4897]: I1121 15:12:59.010086 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tr7z5\" (UniqueName: \"kubernetes.io/projected/3a529e65-6ca1-4cf5-8894-a4f8eb7fa44b-kube-api-access-tr7z5\") pod \"community-operators-kb62c\" (UID: \"3a529e65-6ca1-4cf5-8894-a4f8eb7fa44b\") " pod="openshift-marketplace/community-operators-kb62c" Nov 21 15:12:59 crc kubenswrapper[4897]: I1121 15:12:59.097232 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kb62c" Nov 21 15:12:59 crc kubenswrapper[4897]: I1121 15:12:59.718349 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kb62c"] Nov 21 15:13:00 crc kubenswrapper[4897]: I1121 15:13:00.557993 4897 generic.go:334] "Generic (PLEG): container finished" podID="3a529e65-6ca1-4cf5-8894-a4f8eb7fa44b" containerID="b87958276e1252c0ee573d5bf2325da6f1fafd34a28402a6dc2cc72e0be57aab" exitCode=0 Nov 21 15:13:00 crc kubenswrapper[4897]: I1121 15:13:00.558077 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kb62c" event={"ID":"3a529e65-6ca1-4cf5-8894-a4f8eb7fa44b","Type":"ContainerDied","Data":"b87958276e1252c0ee573d5bf2325da6f1fafd34a28402a6dc2cc72e0be57aab"} Nov 21 15:13:00 crc kubenswrapper[4897]: I1121 15:13:00.558331 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kb62c" event={"ID":"3a529e65-6ca1-4cf5-8894-a4f8eb7fa44b","Type":"ContainerStarted","Data":"855cc5ddf7330f8b067795e968406426dcb59d9dc3d49df5d768022d4d49b693"} Nov 21 15:13:00 crc kubenswrapper[4897]: I1121 15:13:00.560617 4897 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 21 15:13:01 crc kubenswrapper[4897]: I1121 15:13:01.578304 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kb62c" event={"ID":"3a529e65-6ca1-4cf5-8894-a4f8eb7fa44b","Type":"ContainerStarted","Data":"94d8ed351be14a69bd786669ee9ca85b7dd87500bec4f3c9238659c6da24a1fa"} Nov 21 15:13:04 crc kubenswrapper[4897]: I1121 15:13:04.620976 4897 generic.go:334] "Generic (PLEG): container finished" podID="3a529e65-6ca1-4cf5-8894-a4f8eb7fa44b" containerID="94d8ed351be14a69bd786669ee9ca85b7dd87500bec4f3c9238659c6da24a1fa" exitCode=0 Nov 21 15:13:04 crc kubenswrapper[4897]: I1121 15:13:04.621047 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kb62c" event={"ID":"3a529e65-6ca1-4cf5-8894-a4f8eb7fa44b","Type":"ContainerDied","Data":"94d8ed351be14a69bd786669ee9ca85b7dd87500bec4f3c9238659c6da24a1fa"} Nov 21 15:13:05 crc kubenswrapper[4897]: I1121 15:13:05.635457 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kb62c" event={"ID":"3a529e65-6ca1-4cf5-8894-a4f8eb7fa44b","Type":"ContainerStarted","Data":"4a276318d772786886728f84f14720ac6e03d2dc38209b0a92efee741f742e5b"} Nov 21 15:13:05 crc kubenswrapper[4897]: I1121 15:13:05.666086 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-kb62c" podStartSLOduration=3.163091887 podStartE2EDuration="7.66606476s" podCreationTimestamp="2025-11-21 15:12:58 +0000 UTC" firstStartedPulling="2025-11-21 15:13:00.560303671 +0000 UTC m=+3857.844897146" lastFinishedPulling="2025-11-21 15:13:05.063276544 +0000 UTC m=+3862.347870019" observedRunningTime="2025-11-21 15:13:05.658199567 +0000 UTC m=+3862.942793052" watchObservedRunningTime="2025-11-21 15:13:05.66606476 +0000 UTC m=+3862.950658235" Nov 21 15:13:06 crc kubenswrapper[4897]: I1121 15:13:06.089358 4897 scope.go:117] "RemoveContainer" containerID="31498b4cf1d12960e0bd3369fd9c8702f2e4dc6c2b6a5487ecab62d136f4bc3c" Nov 21 15:13:06 crc kubenswrapper[4897]: E1121 15:13:06.089789 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:13:09 crc kubenswrapper[4897]: I1121 15:13:09.097589 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-kb62c" Nov 21 15:13:09 crc kubenswrapper[4897]: I1121 15:13:09.098189 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-kb62c" Nov 21 15:13:09 crc kubenswrapper[4897]: I1121 15:13:09.151204 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-kb62c" Nov 21 15:13:19 crc kubenswrapper[4897]: I1121 15:13:19.147530 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-kb62c" Nov 21 15:13:19 crc kubenswrapper[4897]: I1121 15:13:19.201209 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-kb62c"] Nov 21 15:13:19 crc kubenswrapper[4897]: I1121 15:13:19.841274 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-kb62c" podUID="3a529e65-6ca1-4cf5-8894-a4f8eb7fa44b" containerName="registry-server" containerID="cri-o://4a276318d772786886728f84f14720ac6e03d2dc38209b0a92efee741f742e5b" gracePeriod=2 Nov 21 15:13:20 crc kubenswrapper[4897]: I1121 15:13:20.089847 4897 scope.go:117] "RemoveContainer" containerID="31498b4cf1d12960e0bd3369fd9c8702f2e4dc6c2b6a5487ecab62d136f4bc3c" Nov 21 15:13:20 crc kubenswrapper[4897]: E1121 15:13:20.090443 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:13:20 crc kubenswrapper[4897]: I1121 15:13:20.425550 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kb62c" Nov 21 15:13:20 crc kubenswrapper[4897]: I1121 15:13:20.580915 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a529e65-6ca1-4cf5-8894-a4f8eb7fa44b-utilities\") pod \"3a529e65-6ca1-4cf5-8894-a4f8eb7fa44b\" (UID: \"3a529e65-6ca1-4cf5-8894-a4f8eb7fa44b\") " Nov 21 15:13:20 crc kubenswrapper[4897]: I1121 15:13:20.581090 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tr7z5\" (UniqueName: \"kubernetes.io/projected/3a529e65-6ca1-4cf5-8894-a4f8eb7fa44b-kube-api-access-tr7z5\") pod \"3a529e65-6ca1-4cf5-8894-a4f8eb7fa44b\" (UID: \"3a529e65-6ca1-4cf5-8894-a4f8eb7fa44b\") " Nov 21 15:13:20 crc kubenswrapper[4897]: I1121 15:13:20.581180 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a529e65-6ca1-4cf5-8894-a4f8eb7fa44b-catalog-content\") pod \"3a529e65-6ca1-4cf5-8894-a4f8eb7fa44b\" (UID: \"3a529e65-6ca1-4cf5-8894-a4f8eb7fa44b\") " Nov 21 15:13:20 crc kubenswrapper[4897]: I1121 15:13:20.583048 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3a529e65-6ca1-4cf5-8894-a4f8eb7fa44b-utilities" (OuterVolumeSpecName: "utilities") pod "3a529e65-6ca1-4cf5-8894-a4f8eb7fa44b" (UID: "3a529e65-6ca1-4cf5-8894-a4f8eb7fa44b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 15:13:20 crc kubenswrapper[4897]: I1121 15:13:20.592011 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a529e65-6ca1-4cf5-8894-a4f8eb7fa44b-kube-api-access-tr7z5" (OuterVolumeSpecName: "kube-api-access-tr7z5") pod "3a529e65-6ca1-4cf5-8894-a4f8eb7fa44b" (UID: "3a529e65-6ca1-4cf5-8894-a4f8eb7fa44b"). InnerVolumeSpecName "kube-api-access-tr7z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 15:13:20 crc kubenswrapper[4897]: I1121 15:13:20.673375 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3a529e65-6ca1-4cf5-8894-a4f8eb7fa44b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3a529e65-6ca1-4cf5-8894-a4f8eb7fa44b" (UID: "3a529e65-6ca1-4cf5-8894-a4f8eb7fa44b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 15:13:20 crc kubenswrapper[4897]: I1121 15:13:20.684963 4897 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a529e65-6ca1-4cf5-8894-a4f8eb7fa44b-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 21 15:13:20 crc kubenswrapper[4897]: I1121 15:13:20.685013 4897 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a529e65-6ca1-4cf5-8894-a4f8eb7fa44b-utilities\") on node \"crc\" DevicePath \"\"" Nov 21 15:13:20 crc kubenswrapper[4897]: I1121 15:13:20.685027 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tr7z5\" (UniqueName: \"kubernetes.io/projected/3a529e65-6ca1-4cf5-8894-a4f8eb7fa44b-kube-api-access-tr7z5\") on node \"crc\" DevicePath \"\"" Nov 21 15:13:20 crc kubenswrapper[4897]: I1121 15:13:20.857141 4897 generic.go:334] "Generic (PLEG): container finished" podID="3a529e65-6ca1-4cf5-8894-a4f8eb7fa44b" containerID="4a276318d772786886728f84f14720ac6e03d2dc38209b0a92efee741f742e5b" exitCode=0 Nov 21 15:13:20 crc kubenswrapper[4897]: I1121 15:13:20.857227 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kb62c" event={"ID":"3a529e65-6ca1-4cf5-8894-a4f8eb7fa44b","Type":"ContainerDied","Data":"4a276318d772786886728f84f14720ac6e03d2dc38209b0a92efee741f742e5b"} Nov 21 15:13:20 crc kubenswrapper[4897]: I1121 15:13:20.857672 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kb62c" event={"ID":"3a529e65-6ca1-4cf5-8894-a4f8eb7fa44b","Type":"ContainerDied","Data":"855cc5ddf7330f8b067795e968406426dcb59d9dc3d49df5d768022d4d49b693"} Nov 21 15:13:20 crc kubenswrapper[4897]: I1121 15:13:20.857699 4897 scope.go:117] "RemoveContainer" containerID="4a276318d772786886728f84f14720ac6e03d2dc38209b0a92efee741f742e5b" Nov 21 15:13:20 crc kubenswrapper[4897]: I1121 15:13:20.857262 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kb62c" Nov 21 15:13:20 crc kubenswrapper[4897]: I1121 15:13:20.899801 4897 scope.go:117] "RemoveContainer" containerID="94d8ed351be14a69bd786669ee9ca85b7dd87500bec4f3c9238659c6da24a1fa" Nov 21 15:13:20 crc kubenswrapper[4897]: I1121 15:13:20.921675 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-kb62c"] Nov 21 15:13:20 crc kubenswrapper[4897]: I1121 15:13:20.934579 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-kb62c"] Nov 21 15:13:20 crc kubenswrapper[4897]: I1121 15:13:20.939611 4897 scope.go:117] "RemoveContainer" containerID="b87958276e1252c0ee573d5bf2325da6f1fafd34a28402a6dc2cc72e0be57aab" Nov 21 15:13:20 crc kubenswrapper[4897]: I1121 15:13:20.988462 4897 scope.go:117] "RemoveContainer" containerID="4a276318d772786886728f84f14720ac6e03d2dc38209b0a92efee741f742e5b" Nov 21 15:13:20 crc kubenswrapper[4897]: E1121 15:13:20.988959 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a276318d772786886728f84f14720ac6e03d2dc38209b0a92efee741f742e5b\": container with ID starting with 4a276318d772786886728f84f14720ac6e03d2dc38209b0a92efee741f742e5b not found: ID does not exist" containerID="4a276318d772786886728f84f14720ac6e03d2dc38209b0a92efee741f742e5b" Nov 21 15:13:20 crc kubenswrapper[4897]: I1121 15:13:20.989031 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a276318d772786886728f84f14720ac6e03d2dc38209b0a92efee741f742e5b"} err="failed to get container status \"4a276318d772786886728f84f14720ac6e03d2dc38209b0a92efee741f742e5b\": rpc error: code = NotFound desc = could not find container \"4a276318d772786886728f84f14720ac6e03d2dc38209b0a92efee741f742e5b\": container with ID starting with 4a276318d772786886728f84f14720ac6e03d2dc38209b0a92efee741f742e5b not found: ID does not exist" Nov 21 15:13:20 crc kubenswrapper[4897]: I1121 15:13:20.989057 4897 scope.go:117] "RemoveContainer" containerID="94d8ed351be14a69bd786669ee9ca85b7dd87500bec4f3c9238659c6da24a1fa" Nov 21 15:13:20 crc kubenswrapper[4897]: E1121 15:13:20.989362 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"94d8ed351be14a69bd786669ee9ca85b7dd87500bec4f3c9238659c6da24a1fa\": container with ID starting with 94d8ed351be14a69bd786669ee9ca85b7dd87500bec4f3c9238659c6da24a1fa not found: ID does not exist" containerID="94d8ed351be14a69bd786669ee9ca85b7dd87500bec4f3c9238659c6da24a1fa" Nov 21 15:13:20 crc kubenswrapper[4897]: I1121 15:13:20.989416 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"94d8ed351be14a69bd786669ee9ca85b7dd87500bec4f3c9238659c6da24a1fa"} err="failed to get container status \"94d8ed351be14a69bd786669ee9ca85b7dd87500bec4f3c9238659c6da24a1fa\": rpc error: code = NotFound desc = could not find container \"94d8ed351be14a69bd786669ee9ca85b7dd87500bec4f3c9238659c6da24a1fa\": container with ID starting with 94d8ed351be14a69bd786669ee9ca85b7dd87500bec4f3c9238659c6da24a1fa not found: ID does not exist" Nov 21 15:13:20 crc kubenswrapper[4897]: I1121 15:13:20.989450 4897 scope.go:117] "RemoveContainer" containerID="b87958276e1252c0ee573d5bf2325da6f1fafd34a28402a6dc2cc72e0be57aab" Nov 21 15:13:20 crc kubenswrapper[4897]: E1121 15:13:20.989824 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b87958276e1252c0ee573d5bf2325da6f1fafd34a28402a6dc2cc72e0be57aab\": container with ID starting with b87958276e1252c0ee573d5bf2325da6f1fafd34a28402a6dc2cc72e0be57aab not found: ID does not exist" containerID="b87958276e1252c0ee573d5bf2325da6f1fafd34a28402a6dc2cc72e0be57aab" Nov 21 15:13:20 crc kubenswrapper[4897]: I1121 15:13:20.989864 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b87958276e1252c0ee573d5bf2325da6f1fafd34a28402a6dc2cc72e0be57aab"} err="failed to get container status \"b87958276e1252c0ee573d5bf2325da6f1fafd34a28402a6dc2cc72e0be57aab\": rpc error: code = NotFound desc = could not find container \"b87958276e1252c0ee573d5bf2325da6f1fafd34a28402a6dc2cc72e0be57aab\": container with ID starting with b87958276e1252c0ee573d5bf2325da6f1fafd34a28402a6dc2cc72e0be57aab not found: ID does not exist" Nov 21 15:13:22 crc kubenswrapper[4897]: I1121 15:13:22.101654 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3a529e65-6ca1-4cf5-8894-a4f8eb7fa44b" path="/var/lib/kubelet/pods/3a529e65-6ca1-4cf5-8894-a4f8eb7fa44b/volumes" Nov 21 15:13:32 crc kubenswrapper[4897]: I1121 15:13:32.089144 4897 scope.go:117] "RemoveContainer" containerID="31498b4cf1d12960e0bd3369fd9c8702f2e4dc6c2b6a5487ecab62d136f4bc3c" Nov 21 15:13:32 crc kubenswrapper[4897]: E1121 15:13:32.090080 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:13:46 crc kubenswrapper[4897]: I1121 15:13:46.088845 4897 scope.go:117] "RemoveContainer" containerID="31498b4cf1d12960e0bd3369fd9c8702f2e4dc6c2b6a5487ecab62d136f4bc3c" Nov 21 15:13:46 crc kubenswrapper[4897]: E1121 15:13:46.089584 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:13:58 crc kubenswrapper[4897]: I1121 15:13:58.089466 4897 scope.go:117] "RemoveContainer" containerID="31498b4cf1d12960e0bd3369fd9c8702f2e4dc6c2b6a5487ecab62d136f4bc3c" Nov 21 15:13:58 crc kubenswrapper[4897]: E1121 15:13:58.090350 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:14:12 crc kubenswrapper[4897]: I1121 15:14:12.089178 4897 scope.go:117] "RemoveContainer" containerID="31498b4cf1d12960e0bd3369fd9c8702f2e4dc6c2b6a5487ecab62d136f4bc3c" Nov 21 15:14:12 crc kubenswrapper[4897]: E1121 15:14:12.090099 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:14:23 crc kubenswrapper[4897]: I1121 15:14:23.089303 4897 scope.go:117] "RemoveContainer" containerID="31498b4cf1d12960e0bd3369fd9c8702f2e4dc6c2b6a5487ecab62d136f4bc3c" Nov 21 15:14:23 crc kubenswrapper[4897]: E1121 15:14:23.090183 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:14:35 crc kubenswrapper[4897]: I1121 15:14:35.089958 4897 scope.go:117] "RemoveContainer" containerID="31498b4cf1d12960e0bd3369fd9c8702f2e4dc6c2b6a5487ecab62d136f4bc3c" Nov 21 15:14:35 crc kubenswrapper[4897]: I1121 15:14:35.709680 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" event={"ID":"e7670227-d280-4847-b882-754429f56b0f","Type":"ContainerStarted","Data":"298866e849258a62c75c2157ebb87d240383aec553c6cd2bc7b23079d1c5505d"} Nov 21 15:15:00 crc kubenswrapper[4897]: I1121 15:15:00.150645 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29395635-svgmt"] Nov 21 15:15:00 crc kubenswrapper[4897]: E1121 15:15:00.151708 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a529e65-6ca1-4cf5-8894-a4f8eb7fa44b" containerName="registry-server" Nov 21 15:15:00 crc kubenswrapper[4897]: I1121 15:15:00.151727 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a529e65-6ca1-4cf5-8894-a4f8eb7fa44b" containerName="registry-server" Nov 21 15:15:00 crc kubenswrapper[4897]: E1121 15:15:00.151745 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a529e65-6ca1-4cf5-8894-a4f8eb7fa44b" containerName="extract-content" Nov 21 15:15:00 crc kubenswrapper[4897]: I1121 15:15:00.151751 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a529e65-6ca1-4cf5-8894-a4f8eb7fa44b" containerName="extract-content" Nov 21 15:15:00 crc kubenswrapper[4897]: E1121 15:15:00.151765 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a529e65-6ca1-4cf5-8894-a4f8eb7fa44b" containerName="extract-utilities" Nov 21 15:15:00 crc kubenswrapper[4897]: I1121 15:15:00.151771 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a529e65-6ca1-4cf5-8894-a4f8eb7fa44b" containerName="extract-utilities" Nov 21 15:15:00 crc kubenswrapper[4897]: I1121 15:15:00.152168 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a529e65-6ca1-4cf5-8894-a4f8eb7fa44b" containerName="registry-server" Nov 21 15:15:00 crc kubenswrapper[4897]: I1121 15:15:00.153826 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29395635-svgmt" Nov 21 15:15:00 crc kubenswrapper[4897]: I1121 15:15:00.155977 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 21 15:15:00 crc kubenswrapper[4897]: I1121 15:15:00.156402 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 21 15:15:00 crc kubenswrapper[4897]: I1121 15:15:00.163080 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29395635-svgmt"] Nov 21 15:15:00 crc kubenswrapper[4897]: I1121 15:15:00.298194 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a75d805d-9fcb-4bfd-8309-cf2f821e43d1-secret-volume\") pod \"collect-profiles-29395635-svgmt\" (UID: \"a75d805d-9fcb-4bfd-8309-cf2f821e43d1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29395635-svgmt" Nov 21 15:15:00 crc kubenswrapper[4897]: I1121 15:15:00.298537 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a75d805d-9fcb-4bfd-8309-cf2f821e43d1-config-volume\") pod \"collect-profiles-29395635-svgmt\" (UID: \"a75d805d-9fcb-4bfd-8309-cf2f821e43d1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29395635-svgmt" Nov 21 15:15:00 crc kubenswrapper[4897]: I1121 15:15:00.299093 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-25xv2\" (UniqueName: \"kubernetes.io/projected/a75d805d-9fcb-4bfd-8309-cf2f821e43d1-kube-api-access-25xv2\") pod \"collect-profiles-29395635-svgmt\" (UID: \"a75d805d-9fcb-4bfd-8309-cf2f821e43d1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29395635-svgmt" Nov 21 15:15:00 crc kubenswrapper[4897]: I1121 15:15:00.401336 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a75d805d-9fcb-4bfd-8309-cf2f821e43d1-secret-volume\") pod \"collect-profiles-29395635-svgmt\" (UID: \"a75d805d-9fcb-4bfd-8309-cf2f821e43d1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29395635-svgmt" Nov 21 15:15:00 crc kubenswrapper[4897]: I1121 15:15:00.401379 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a75d805d-9fcb-4bfd-8309-cf2f821e43d1-config-volume\") pod \"collect-profiles-29395635-svgmt\" (UID: \"a75d805d-9fcb-4bfd-8309-cf2f821e43d1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29395635-svgmt" Nov 21 15:15:00 crc kubenswrapper[4897]: I1121 15:15:00.401443 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-25xv2\" (UniqueName: \"kubernetes.io/projected/a75d805d-9fcb-4bfd-8309-cf2f821e43d1-kube-api-access-25xv2\") pod \"collect-profiles-29395635-svgmt\" (UID: \"a75d805d-9fcb-4bfd-8309-cf2f821e43d1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29395635-svgmt" Nov 21 15:15:00 crc kubenswrapper[4897]: I1121 15:15:00.402909 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a75d805d-9fcb-4bfd-8309-cf2f821e43d1-config-volume\") pod \"collect-profiles-29395635-svgmt\" (UID: \"a75d805d-9fcb-4bfd-8309-cf2f821e43d1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29395635-svgmt" Nov 21 15:15:00 crc kubenswrapper[4897]: I1121 15:15:00.408066 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a75d805d-9fcb-4bfd-8309-cf2f821e43d1-secret-volume\") pod \"collect-profiles-29395635-svgmt\" (UID: \"a75d805d-9fcb-4bfd-8309-cf2f821e43d1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29395635-svgmt" Nov 21 15:15:00 crc kubenswrapper[4897]: I1121 15:15:00.419447 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-25xv2\" (UniqueName: \"kubernetes.io/projected/a75d805d-9fcb-4bfd-8309-cf2f821e43d1-kube-api-access-25xv2\") pod \"collect-profiles-29395635-svgmt\" (UID: \"a75d805d-9fcb-4bfd-8309-cf2f821e43d1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29395635-svgmt" Nov 21 15:15:00 crc kubenswrapper[4897]: I1121 15:15:00.479071 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29395635-svgmt" Nov 21 15:15:00 crc kubenswrapper[4897]: I1121 15:15:00.997231 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29395635-svgmt"] Nov 21 15:15:01 crc kubenswrapper[4897]: I1121 15:15:01.980959 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29395635-svgmt" event={"ID":"a75d805d-9fcb-4bfd-8309-cf2f821e43d1","Type":"ContainerStarted","Data":"d997926891bef4c41e1bb4f60974f61ac45236ae28d6cf92a3825034418b0303"} Nov 21 15:15:01 crc kubenswrapper[4897]: I1121 15:15:01.981730 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29395635-svgmt" event={"ID":"a75d805d-9fcb-4bfd-8309-cf2f821e43d1","Type":"ContainerStarted","Data":"53c5d722657bfd156b04f619a4809c6bdde4394854be58962275f84fc9677571"} Nov 21 15:15:02 crc kubenswrapper[4897]: I1121 15:15:02.014287 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29395635-svgmt" podStartSLOduration=2.01427079 podStartE2EDuration="2.01427079s" podCreationTimestamp="2025-11-21 15:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 15:15:02.007296441 +0000 UTC m=+3979.291889916" watchObservedRunningTime="2025-11-21 15:15:02.01427079 +0000 UTC m=+3979.298864265" Nov 21 15:15:03 crc kubenswrapper[4897]: I1121 15:15:03.006467 4897 generic.go:334] "Generic (PLEG): container finished" podID="a75d805d-9fcb-4bfd-8309-cf2f821e43d1" containerID="d997926891bef4c41e1bb4f60974f61ac45236ae28d6cf92a3825034418b0303" exitCode=0 Nov 21 15:15:03 crc kubenswrapper[4897]: I1121 15:15:03.006608 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29395635-svgmt" event={"ID":"a75d805d-9fcb-4bfd-8309-cf2f821e43d1","Type":"ContainerDied","Data":"d997926891bef4c41e1bb4f60974f61ac45236ae28d6cf92a3825034418b0303"} Nov 21 15:15:04 crc kubenswrapper[4897]: I1121 15:15:04.548404 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29395635-svgmt" Nov 21 15:15:04 crc kubenswrapper[4897]: I1121 15:15:04.718030 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a75d805d-9fcb-4bfd-8309-cf2f821e43d1-config-volume\") pod \"a75d805d-9fcb-4bfd-8309-cf2f821e43d1\" (UID: \"a75d805d-9fcb-4bfd-8309-cf2f821e43d1\") " Nov 21 15:15:04 crc kubenswrapper[4897]: I1121 15:15:04.718287 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a75d805d-9fcb-4bfd-8309-cf2f821e43d1-secret-volume\") pod \"a75d805d-9fcb-4bfd-8309-cf2f821e43d1\" (UID: \"a75d805d-9fcb-4bfd-8309-cf2f821e43d1\") " Nov 21 15:15:04 crc kubenswrapper[4897]: I1121 15:15:04.718342 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-25xv2\" (UniqueName: \"kubernetes.io/projected/a75d805d-9fcb-4bfd-8309-cf2f821e43d1-kube-api-access-25xv2\") pod \"a75d805d-9fcb-4bfd-8309-cf2f821e43d1\" (UID: \"a75d805d-9fcb-4bfd-8309-cf2f821e43d1\") " Nov 21 15:15:04 crc kubenswrapper[4897]: I1121 15:15:04.719720 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a75d805d-9fcb-4bfd-8309-cf2f821e43d1-config-volume" (OuterVolumeSpecName: "config-volume") pod "a75d805d-9fcb-4bfd-8309-cf2f821e43d1" (UID: "a75d805d-9fcb-4bfd-8309-cf2f821e43d1"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 15:15:04 crc kubenswrapper[4897]: I1121 15:15:04.724597 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a75d805d-9fcb-4bfd-8309-cf2f821e43d1-kube-api-access-25xv2" (OuterVolumeSpecName: "kube-api-access-25xv2") pod "a75d805d-9fcb-4bfd-8309-cf2f821e43d1" (UID: "a75d805d-9fcb-4bfd-8309-cf2f821e43d1"). InnerVolumeSpecName "kube-api-access-25xv2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 15:15:04 crc kubenswrapper[4897]: I1121 15:15:04.724781 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a75d805d-9fcb-4bfd-8309-cf2f821e43d1-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "a75d805d-9fcb-4bfd-8309-cf2f821e43d1" (UID: "a75d805d-9fcb-4bfd-8309-cf2f821e43d1"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 15:15:04 crc kubenswrapper[4897]: I1121 15:15:04.820585 4897 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a75d805d-9fcb-4bfd-8309-cf2f821e43d1-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 21 15:15:04 crc kubenswrapper[4897]: I1121 15:15:04.820635 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-25xv2\" (UniqueName: \"kubernetes.io/projected/a75d805d-9fcb-4bfd-8309-cf2f821e43d1-kube-api-access-25xv2\") on node \"crc\" DevicePath \"\"" Nov 21 15:15:04 crc kubenswrapper[4897]: I1121 15:15:04.820646 4897 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a75d805d-9fcb-4bfd-8309-cf2f821e43d1-config-volume\") on node \"crc\" DevicePath \"\"" Nov 21 15:15:05 crc kubenswrapper[4897]: I1121 15:15:05.036498 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29395635-svgmt" event={"ID":"a75d805d-9fcb-4bfd-8309-cf2f821e43d1","Type":"ContainerDied","Data":"53c5d722657bfd156b04f619a4809c6bdde4394854be58962275f84fc9677571"} Nov 21 15:15:05 crc kubenswrapper[4897]: I1121 15:15:05.036563 4897 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="53c5d722657bfd156b04f619a4809c6bdde4394854be58962275f84fc9677571" Nov 21 15:15:05 crc kubenswrapper[4897]: I1121 15:15:05.036571 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29395635-svgmt" Nov 21 15:15:05 crc kubenswrapper[4897]: I1121 15:15:05.087561 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29395590-t8htq"] Nov 21 15:15:05 crc kubenswrapper[4897]: I1121 15:15:05.098105 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29395590-t8htq"] Nov 21 15:15:06 crc kubenswrapper[4897]: I1121 15:15:06.108125 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="99d4fc7b-7c46-47e7-8bf6-906dfd87af3a" path="/var/lib/kubelet/pods/99d4fc7b-7c46-47e7-8bf6-906dfd87af3a/volumes" Nov 21 15:15:22 crc kubenswrapper[4897]: I1121 15:15:22.872085 4897 scope.go:117] "RemoveContainer" containerID="695d365282e23d79ab96d852c2a92d6418c185f004a384bf60851c2e6f2d8673" Nov 21 15:15:56 crc kubenswrapper[4897]: I1121 15:15:56.170448 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-q4xm9"] Nov 21 15:15:56 crc kubenswrapper[4897]: E1121 15:15:56.172491 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a75d805d-9fcb-4bfd-8309-cf2f821e43d1" containerName="collect-profiles" Nov 21 15:15:56 crc kubenswrapper[4897]: I1121 15:15:56.172607 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="a75d805d-9fcb-4bfd-8309-cf2f821e43d1" containerName="collect-profiles" Nov 21 15:15:56 crc kubenswrapper[4897]: I1121 15:15:56.172997 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="a75d805d-9fcb-4bfd-8309-cf2f821e43d1" containerName="collect-profiles" Nov 21 15:15:56 crc kubenswrapper[4897]: I1121 15:15:56.174972 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q4xm9" Nov 21 15:15:56 crc kubenswrapper[4897]: I1121 15:15:56.182957 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-q4xm9"] Nov 21 15:15:56 crc kubenswrapper[4897]: I1121 15:15:56.344927 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dvpjt\" (UniqueName: \"kubernetes.io/projected/d33dc481-18ae-4b83-95ec-12d61284837c-kube-api-access-dvpjt\") pod \"redhat-marketplace-q4xm9\" (UID: \"d33dc481-18ae-4b83-95ec-12d61284837c\") " pod="openshift-marketplace/redhat-marketplace-q4xm9" Nov 21 15:15:56 crc kubenswrapper[4897]: I1121 15:15:56.345397 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d33dc481-18ae-4b83-95ec-12d61284837c-catalog-content\") pod \"redhat-marketplace-q4xm9\" (UID: \"d33dc481-18ae-4b83-95ec-12d61284837c\") " pod="openshift-marketplace/redhat-marketplace-q4xm9" Nov 21 15:15:56 crc kubenswrapper[4897]: I1121 15:15:56.345489 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d33dc481-18ae-4b83-95ec-12d61284837c-utilities\") pod \"redhat-marketplace-q4xm9\" (UID: \"d33dc481-18ae-4b83-95ec-12d61284837c\") " pod="openshift-marketplace/redhat-marketplace-q4xm9" Nov 21 15:15:56 crc kubenswrapper[4897]: I1121 15:15:56.447622 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d33dc481-18ae-4b83-95ec-12d61284837c-utilities\") pod \"redhat-marketplace-q4xm9\" (UID: \"d33dc481-18ae-4b83-95ec-12d61284837c\") " pod="openshift-marketplace/redhat-marketplace-q4xm9" Nov 21 15:15:56 crc kubenswrapper[4897]: I1121 15:15:56.447817 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dvpjt\" (UniqueName: \"kubernetes.io/projected/d33dc481-18ae-4b83-95ec-12d61284837c-kube-api-access-dvpjt\") pod \"redhat-marketplace-q4xm9\" (UID: \"d33dc481-18ae-4b83-95ec-12d61284837c\") " pod="openshift-marketplace/redhat-marketplace-q4xm9" Nov 21 15:15:56 crc kubenswrapper[4897]: I1121 15:15:56.447916 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d33dc481-18ae-4b83-95ec-12d61284837c-catalog-content\") pod \"redhat-marketplace-q4xm9\" (UID: \"d33dc481-18ae-4b83-95ec-12d61284837c\") " pod="openshift-marketplace/redhat-marketplace-q4xm9" Nov 21 15:15:56 crc kubenswrapper[4897]: I1121 15:15:56.448630 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d33dc481-18ae-4b83-95ec-12d61284837c-catalog-content\") pod \"redhat-marketplace-q4xm9\" (UID: \"d33dc481-18ae-4b83-95ec-12d61284837c\") " pod="openshift-marketplace/redhat-marketplace-q4xm9" Nov 21 15:15:56 crc kubenswrapper[4897]: I1121 15:15:56.448912 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d33dc481-18ae-4b83-95ec-12d61284837c-utilities\") pod \"redhat-marketplace-q4xm9\" (UID: \"d33dc481-18ae-4b83-95ec-12d61284837c\") " pod="openshift-marketplace/redhat-marketplace-q4xm9" Nov 21 15:15:56 crc kubenswrapper[4897]: I1121 15:15:56.473724 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dvpjt\" (UniqueName: \"kubernetes.io/projected/d33dc481-18ae-4b83-95ec-12d61284837c-kube-api-access-dvpjt\") pod \"redhat-marketplace-q4xm9\" (UID: \"d33dc481-18ae-4b83-95ec-12d61284837c\") " pod="openshift-marketplace/redhat-marketplace-q4xm9" Nov 21 15:15:56 crc kubenswrapper[4897]: I1121 15:15:56.512599 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q4xm9" Nov 21 15:15:57 crc kubenswrapper[4897]: I1121 15:15:57.089461 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-q4xm9"] Nov 21 15:15:57 crc kubenswrapper[4897]: W1121 15:15:57.106232 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd33dc481_18ae_4b83_95ec_12d61284837c.slice/crio-7644fa16c1294acb82ebd06b5d24d5e6bf5e6e738e36c0b468ef9cf7b1d9b559 WatchSource:0}: Error finding container 7644fa16c1294acb82ebd06b5d24d5e6bf5e6e738e36c0b468ef9cf7b1d9b559: Status 404 returned error can't find the container with id 7644fa16c1294acb82ebd06b5d24d5e6bf5e6e738e36c0b468ef9cf7b1d9b559 Nov 21 15:15:57 crc kubenswrapper[4897]: I1121 15:15:57.632970 4897 generic.go:334] "Generic (PLEG): container finished" podID="d33dc481-18ae-4b83-95ec-12d61284837c" containerID="eace95a075766632933c012de0af0143df8783bcf31e5caf898cb123dd823208" exitCode=0 Nov 21 15:15:57 crc kubenswrapper[4897]: I1121 15:15:57.633253 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q4xm9" event={"ID":"d33dc481-18ae-4b83-95ec-12d61284837c","Type":"ContainerDied","Data":"eace95a075766632933c012de0af0143df8783bcf31e5caf898cb123dd823208"} Nov 21 15:15:57 crc kubenswrapper[4897]: I1121 15:15:57.633698 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q4xm9" event={"ID":"d33dc481-18ae-4b83-95ec-12d61284837c","Type":"ContainerStarted","Data":"7644fa16c1294acb82ebd06b5d24d5e6bf5e6e738e36c0b468ef9cf7b1d9b559"} Nov 21 15:15:58 crc kubenswrapper[4897]: I1121 15:15:58.656368 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q4xm9" event={"ID":"d33dc481-18ae-4b83-95ec-12d61284837c","Type":"ContainerStarted","Data":"49848e9c06d63cbcb7cdad3740d67238190026578a3bc5dbb0b6552c3121d63f"} Nov 21 15:15:59 crc kubenswrapper[4897]: I1121 15:15:59.670205 4897 generic.go:334] "Generic (PLEG): container finished" podID="d33dc481-18ae-4b83-95ec-12d61284837c" containerID="49848e9c06d63cbcb7cdad3740d67238190026578a3bc5dbb0b6552c3121d63f" exitCode=0 Nov 21 15:15:59 crc kubenswrapper[4897]: I1121 15:15:59.670556 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q4xm9" event={"ID":"d33dc481-18ae-4b83-95ec-12d61284837c","Type":"ContainerDied","Data":"49848e9c06d63cbcb7cdad3740d67238190026578a3bc5dbb0b6552c3121d63f"} Nov 21 15:16:00 crc kubenswrapper[4897]: I1121 15:16:00.686105 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q4xm9" event={"ID":"d33dc481-18ae-4b83-95ec-12d61284837c","Type":"ContainerStarted","Data":"937bd2899cdc2a03ab8ecb61b8263e1ad1cd81979428567d08e6e05f862a55c2"} Nov 21 15:16:00 crc kubenswrapper[4897]: I1121 15:16:00.711841 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-q4xm9" podStartSLOduration=2.267070196 podStartE2EDuration="4.711812951s" podCreationTimestamp="2025-11-21 15:15:56 +0000 UTC" firstStartedPulling="2025-11-21 15:15:57.637030402 +0000 UTC m=+4034.921623877" lastFinishedPulling="2025-11-21 15:16:00.081773157 +0000 UTC m=+4037.366366632" observedRunningTime="2025-11-21 15:16:00.709313583 +0000 UTC m=+4037.993907068" watchObservedRunningTime="2025-11-21 15:16:00.711812951 +0000 UTC m=+4037.996406426" Nov 21 15:16:06 crc kubenswrapper[4897]: I1121 15:16:06.513839 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-q4xm9" Nov 21 15:16:06 crc kubenswrapper[4897]: I1121 15:16:06.514426 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-q4xm9" Nov 21 15:16:06 crc kubenswrapper[4897]: I1121 15:16:06.581930 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-q4xm9" Nov 21 15:16:06 crc kubenswrapper[4897]: I1121 15:16:06.801834 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-q4xm9" Nov 21 15:16:06 crc kubenswrapper[4897]: I1121 15:16:06.850707 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-q4xm9"] Nov 21 15:16:08 crc kubenswrapper[4897]: I1121 15:16:08.770982 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-q4xm9" podUID="d33dc481-18ae-4b83-95ec-12d61284837c" containerName="registry-server" containerID="cri-o://937bd2899cdc2a03ab8ecb61b8263e1ad1cd81979428567d08e6e05f862a55c2" gracePeriod=2 Nov 21 15:16:09 crc kubenswrapper[4897]: I1121 15:16:09.356929 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q4xm9" Nov 21 15:16:09 crc kubenswrapper[4897]: I1121 15:16:09.520413 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d33dc481-18ae-4b83-95ec-12d61284837c-catalog-content\") pod \"d33dc481-18ae-4b83-95ec-12d61284837c\" (UID: \"d33dc481-18ae-4b83-95ec-12d61284837c\") " Nov 21 15:16:09 crc kubenswrapper[4897]: I1121 15:16:09.520674 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dvpjt\" (UniqueName: \"kubernetes.io/projected/d33dc481-18ae-4b83-95ec-12d61284837c-kube-api-access-dvpjt\") pod \"d33dc481-18ae-4b83-95ec-12d61284837c\" (UID: \"d33dc481-18ae-4b83-95ec-12d61284837c\") " Nov 21 15:16:09 crc kubenswrapper[4897]: I1121 15:16:09.520715 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d33dc481-18ae-4b83-95ec-12d61284837c-utilities\") pod \"d33dc481-18ae-4b83-95ec-12d61284837c\" (UID: \"d33dc481-18ae-4b83-95ec-12d61284837c\") " Nov 21 15:16:09 crc kubenswrapper[4897]: I1121 15:16:09.521696 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d33dc481-18ae-4b83-95ec-12d61284837c-utilities" (OuterVolumeSpecName: "utilities") pod "d33dc481-18ae-4b83-95ec-12d61284837c" (UID: "d33dc481-18ae-4b83-95ec-12d61284837c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 15:16:09 crc kubenswrapper[4897]: I1121 15:16:09.526739 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d33dc481-18ae-4b83-95ec-12d61284837c-kube-api-access-dvpjt" (OuterVolumeSpecName: "kube-api-access-dvpjt") pod "d33dc481-18ae-4b83-95ec-12d61284837c" (UID: "d33dc481-18ae-4b83-95ec-12d61284837c"). InnerVolumeSpecName "kube-api-access-dvpjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 15:16:09 crc kubenswrapper[4897]: I1121 15:16:09.539521 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d33dc481-18ae-4b83-95ec-12d61284837c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d33dc481-18ae-4b83-95ec-12d61284837c" (UID: "d33dc481-18ae-4b83-95ec-12d61284837c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 15:16:09 crc kubenswrapper[4897]: I1121 15:16:09.623918 4897 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d33dc481-18ae-4b83-95ec-12d61284837c-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 21 15:16:09 crc kubenswrapper[4897]: I1121 15:16:09.623969 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dvpjt\" (UniqueName: \"kubernetes.io/projected/d33dc481-18ae-4b83-95ec-12d61284837c-kube-api-access-dvpjt\") on node \"crc\" DevicePath \"\"" Nov 21 15:16:09 crc kubenswrapper[4897]: I1121 15:16:09.623982 4897 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d33dc481-18ae-4b83-95ec-12d61284837c-utilities\") on node \"crc\" DevicePath \"\"" Nov 21 15:16:09 crc kubenswrapper[4897]: I1121 15:16:09.783183 4897 generic.go:334] "Generic (PLEG): container finished" podID="d33dc481-18ae-4b83-95ec-12d61284837c" containerID="937bd2899cdc2a03ab8ecb61b8263e1ad1cd81979428567d08e6e05f862a55c2" exitCode=0 Nov 21 15:16:09 crc kubenswrapper[4897]: I1121 15:16:09.783227 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q4xm9" event={"ID":"d33dc481-18ae-4b83-95ec-12d61284837c","Type":"ContainerDied","Data":"937bd2899cdc2a03ab8ecb61b8263e1ad1cd81979428567d08e6e05f862a55c2"} Nov 21 15:16:09 crc kubenswrapper[4897]: I1121 15:16:09.783243 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q4xm9" Nov 21 15:16:09 crc kubenswrapper[4897]: I1121 15:16:09.783258 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q4xm9" event={"ID":"d33dc481-18ae-4b83-95ec-12d61284837c","Type":"ContainerDied","Data":"7644fa16c1294acb82ebd06b5d24d5e6bf5e6e738e36c0b468ef9cf7b1d9b559"} Nov 21 15:16:09 crc kubenswrapper[4897]: I1121 15:16:09.783276 4897 scope.go:117] "RemoveContainer" containerID="937bd2899cdc2a03ab8ecb61b8263e1ad1cd81979428567d08e6e05f862a55c2" Nov 21 15:16:09 crc kubenswrapper[4897]: I1121 15:16:09.805131 4897 scope.go:117] "RemoveContainer" containerID="49848e9c06d63cbcb7cdad3740d67238190026578a3bc5dbb0b6552c3121d63f" Nov 21 15:16:09 crc kubenswrapper[4897]: I1121 15:16:09.821377 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-q4xm9"] Nov 21 15:16:09 crc kubenswrapper[4897]: I1121 15:16:09.835786 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-q4xm9"] Nov 21 15:16:09 crc kubenswrapper[4897]: I1121 15:16:09.848089 4897 scope.go:117] "RemoveContainer" containerID="eace95a075766632933c012de0af0143df8783bcf31e5caf898cb123dd823208" Nov 21 15:16:09 crc kubenswrapper[4897]: I1121 15:16:09.906827 4897 scope.go:117] "RemoveContainer" containerID="937bd2899cdc2a03ab8ecb61b8263e1ad1cd81979428567d08e6e05f862a55c2" Nov 21 15:16:09 crc kubenswrapper[4897]: E1121 15:16:09.907626 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"937bd2899cdc2a03ab8ecb61b8263e1ad1cd81979428567d08e6e05f862a55c2\": container with ID starting with 937bd2899cdc2a03ab8ecb61b8263e1ad1cd81979428567d08e6e05f862a55c2 not found: ID does not exist" containerID="937bd2899cdc2a03ab8ecb61b8263e1ad1cd81979428567d08e6e05f862a55c2" Nov 21 15:16:09 crc kubenswrapper[4897]: I1121 15:16:09.907665 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"937bd2899cdc2a03ab8ecb61b8263e1ad1cd81979428567d08e6e05f862a55c2"} err="failed to get container status \"937bd2899cdc2a03ab8ecb61b8263e1ad1cd81979428567d08e6e05f862a55c2\": rpc error: code = NotFound desc = could not find container \"937bd2899cdc2a03ab8ecb61b8263e1ad1cd81979428567d08e6e05f862a55c2\": container with ID starting with 937bd2899cdc2a03ab8ecb61b8263e1ad1cd81979428567d08e6e05f862a55c2 not found: ID does not exist" Nov 21 15:16:09 crc kubenswrapper[4897]: I1121 15:16:09.907686 4897 scope.go:117] "RemoveContainer" containerID="49848e9c06d63cbcb7cdad3740d67238190026578a3bc5dbb0b6552c3121d63f" Nov 21 15:16:09 crc kubenswrapper[4897]: E1121 15:16:09.908432 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"49848e9c06d63cbcb7cdad3740d67238190026578a3bc5dbb0b6552c3121d63f\": container with ID starting with 49848e9c06d63cbcb7cdad3740d67238190026578a3bc5dbb0b6552c3121d63f not found: ID does not exist" containerID="49848e9c06d63cbcb7cdad3740d67238190026578a3bc5dbb0b6552c3121d63f" Nov 21 15:16:09 crc kubenswrapper[4897]: I1121 15:16:09.908466 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"49848e9c06d63cbcb7cdad3740d67238190026578a3bc5dbb0b6552c3121d63f"} err="failed to get container status \"49848e9c06d63cbcb7cdad3740d67238190026578a3bc5dbb0b6552c3121d63f\": rpc error: code = NotFound desc = could not find container \"49848e9c06d63cbcb7cdad3740d67238190026578a3bc5dbb0b6552c3121d63f\": container with ID starting with 49848e9c06d63cbcb7cdad3740d67238190026578a3bc5dbb0b6552c3121d63f not found: ID does not exist" Nov 21 15:16:09 crc kubenswrapper[4897]: I1121 15:16:09.908487 4897 scope.go:117] "RemoveContainer" containerID="eace95a075766632933c012de0af0143df8783bcf31e5caf898cb123dd823208" Nov 21 15:16:09 crc kubenswrapper[4897]: E1121 15:16:09.909585 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eace95a075766632933c012de0af0143df8783bcf31e5caf898cb123dd823208\": container with ID starting with eace95a075766632933c012de0af0143df8783bcf31e5caf898cb123dd823208 not found: ID does not exist" containerID="eace95a075766632933c012de0af0143df8783bcf31e5caf898cb123dd823208" Nov 21 15:16:09 crc kubenswrapper[4897]: I1121 15:16:09.909614 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eace95a075766632933c012de0af0143df8783bcf31e5caf898cb123dd823208"} err="failed to get container status \"eace95a075766632933c012de0af0143df8783bcf31e5caf898cb123dd823208\": rpc error: code = NotFound desc = could not find container \"eace95a075766632933c012de0af0143df8783bcf31e5caf898cb123dd823208\": container with ID starting with eace95a075766632933c012de0af0143df8783bcf31e5caf898cb123dd823208 not found: ID does not exist" Nov 21 15:16:10 crc kubenswrapper[4897]: I1121 15:16:10.103996 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d33dc481-18ae-4b83-95ec-12d61284837c" path="/var/lib/kubelet/pods/d33dc481-18ae-4b83-95ec-12d61284837c/volumes" Nov 21 15:17:04 crc kubenswrapper[4897]: I1121 15:17:04.370813 4897 patch_prober.go:28] interesting pod/machine-config-daemon-krv5b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 21 15:17:04 crc kubenswrapper[4897]: I1121 15:17:04.371458 4897 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 21 15:17:05 crc kubenswrapper[4897]: E1121 15:17:05.931462 4897 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.233:60144->38.102.83.233:43735: write tcp 38.102.83.233:60144->38.102.83.233:43735: write: broken pipe Nov 21 15:17:34 crc kubenswrapper[4897]: I1121 15:17:34.370658 4897 patch_prober.go:28] interesting pod/machine-config-daemon-krv5b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 21 15:17:34 crc kubenswrapper[4897]: I1121 15:17:34.371486 4897 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 21 15:18:04 crc kubenswrapper[4897]: I1121 15:18:04.370788 4897 patch_prober.go:28] interesting pod/machine-config-daemon-krv5b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 21 15:18:04 crc kubenswrapper[4897]: I1121 15:18:04.371356 4897 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 21 15:18:04 crc kubenswrapper[4897]: I1121 15:18:04.371401 4897 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" Nov 21 15:18:04 crc kubenswrapper[4897]: I1121 15:18:04.372007 4897 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"298866e849258a62c75c2157ebb87d240383aec553c6cd2bc7b23079d1c5505d"} pod="openshift-machine-config-operator/machine-config-daemon-krv5b" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 21 15:18:04 crc kubenswrapper[4897]: I1121 15:18:04.372109 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" containerName="machine-config-daemon" containerID="cri-o://298866e849258a62c75c2157ebb87d240383aec553c6cd2bc7b23079d1c5505d" gracePeriod=600 Nov 21 15:18:05 crc kubenswrapper[4897]: I1121 15:18:05.060718 4897 generic.go:334] "Generic (PLEG): container finished" podID="e7670227-d280-4847-b882-754429f56b0f" containerID="298866e849258a62c75c2157ebb87d240383aec553c6cd2bc7b23079d1c5505d" exitCode=0 Nov 21 15:18:05 crc kubenswrapper[4897]: I1121 15:18:05.060778 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" event={"ID":"e7670227-d280-4847-b882-754429f56b0f","Type":"ContainerDied","Data":"298866e849258a62c75c2157ebb87d240383aec553c6cd2bc7b23079d1c5505d"} Nov 21 15:18:05 crc kubenswrapper[4897]: I1121 15:18:05.061073 4897 scope.go:117] "RemoveContainer" containerID="31498b4cf1d12960e0bd3369fd9c8702f2e4dc6c2b6a5487ecab62d136f4bc3c" Nov 21 15:18:06 crc kubenswrapper[4897]: I1121 15:18:06.074192 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" event={"ID":"e7670227-d280-4847-b882-754429f56b0f","Type":"ContainerStarted","Data":"95e70a2cadfac38d1f3f3ba5e655bb58f1dc4e2d41e3d3c5a54f91a8584b921e"} Nov 21 15:19:02 crc kubenswrapper[4897]: I1121 15:19:02.693096 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-gcpmv"] Nov 21 15:19:02 crc kubenswrapper[4897]: E1121 15:19:02.694285 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d33dc481-18ae-4b83-95ec-12d61284837c" containerName="extract-content" Nov 21 15:19:02 crc kubenswrapper[4897]: I1121 15:19:02.694303 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="d33dc481-18ae-4b83-95ec-12d61284837c" containerName="extract-content" Nov 21 15:19:02 crc kubenswrapper[4897]: E1121 15:19:02.694345 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d33dc481-18ae-4b83-95ec-12d61284837c" containerName="extract-utilities" Nov 21 15:19:02 crc kubenswrapper[4897]: I1121 15:19:02.694351 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="d33dc481-18ae-4b83-95ec-12d61284837c" containerName="extract-utilities" Nov 21 15:19:02 crc kubenswrapper[4897]: E1121 15:19:02.694370 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d33dc481-18ae-4b83-95ec-12d61284837c" containerName="registry-server" Nov 21 15:19:02 crc kubenswrapper[4897]: I1121 15:19:02.694379 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="d33dc481-18ae-4b83-95ec-12d61284837c" containerName="registry-server" Nov 21 15:19:02 crc kubenswrapper[4897]: I1121 15:19:02.694628 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="d33dc481-18ae-4b83-95ec-12d61284837c" containerName="registry-server" Nov 21 15:19:02 crc kubenswrapper[4897]: I1121 15:19:02.696703 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gcpmv" Nov 21 15:19:02 crc kubenswrapper[4897]: I1121 15:19:02.706635 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gcpmv"] Nov 21 15:19:02 crc kubenswrapper[4897]: I1121 15:19:02.812849 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qtd2d\" (UniqueName: \"kubernetes.io/projected/dcbb3655-443e-4ef1-9789-2f4d956120fc-kube-api-access-qtd2d\") pod \"redhat-operators-gcpmv\" (UID: \"dcbb3655-443e-4ef1-9789-2f4d956120fc\") " pod="openshift-marketplace/redhat-operators-gcpmv" Nov 21 15:19:02 crc kubenswrapper[4897]: I1121 15:19:02.813102 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dcbb3655-443e-4ef1-9789-2f4d956120fc-catalog-content\") pod \"redhat-operators-gcpmv\" (UID: \"dcbb3655-443e-4ef1-9789-2f4d956120fc\") " pod="openshift-marketplace/redhat-operators-gcpmv" Nov 21 15:19:02 crc kubenswrapper[4897]: I1121 15:19:02.813334 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dcbb3655-443e-4ef1-9789-2f4d956120fc-utilities\") pod \"redhat-operators-gcpmv\" (UID: \"dcbb3655-443e-4ef1-9789-2f4d956120fc\") " pod="openshift-marketplace/redhat-operators-gcpmv" Nov 21 15:19:02 crc kubenswrapper[4897]: I1121 15:19:02.915551 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dcbb3655-443e-4ef1-9789-2f4d956120fc-utilities\") pod \"redhat-operators-gcpmv\" (UID: \"dcbb3655-443e-4ef1-9789-2f4d956120fc\") " pod="openshift-marketplace/redhat-operators-gcpmv" Nov 21 15:19:02 crc kubenswrapper[4897]: I1121 15:19:02.915690 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qtd2d\" (UniqueName: \"kubernetes.io/projected/dcbb3655-443e-4ef1-9789-2f4d956120fc-kube-api-access-qtd2d\") pod \"redhat-operators-gcpmv\" (UID: \"dcbb3655-443e-4ef1-9789-2f4d956120fc\") " pod="openshift-marketplace/redhat-operators-gcpmv" Nov 21 15:19:02 crc kubenswrapper[4897]: I1121 15:19:02.915831 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dcbb3655-443e-4ef1-9789-2f4d956120fc-catalog-content\") pod \"redhat-operators-gcpmv\" (UID: \"dcbb3655-443e-4ef1-9789-2f4d956120fc\") " pod="openshift-marketplace/redhat-operators-gcpmv" Nov 21 15:19:02 crc kubenswrapper[4897]: I1121 15:19:02.916025 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dcbb3655-443e-4ef1-9789-2f4d956120fc-utilities\") pod \"redhat-operators-gcpmv\" (UID: \"dcbb3655-443e-4ef1-9789-2f4d956120fc\") " pod="openshift-marketplace/redhat-operators-gcpmv" Nov 21 15:19:02 crc kubenswrapper[4897]: I1121 15:19:02.916254 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dcbb3655-443e-4ef1-9789-2f4d956120fc-catalog-content\") pod \"redhat-operators-gcpmv\" (UID: \"dcbb3655-443e-4ef1-9789-2f4d956120fc\") " pod="openshift-marketplace/redhat-operators-gcpmv" Nov 21 15:19:03 crc kubenswrapper[4897]: I1121 15:19:03.019140 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qtd2d\" (UniqueName: \"kubernetes.io/projected/dcbb3655-443e-4ef1-9789-2f4d956120fc-kube-api-access-qtd2d\") pod \"redhat-operators-gcpmv\" (UID: \"dcbb3655-443e-4ef1-9789-2f4d956120fc\") " pod="openshift-marketplace/redhat-operators-gcpmv" Nov 21 15:19:03 crc kubenswrapper[4897]: I1121 15:19:03.030984 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gcpmv" Nov 21 15:19:03 crc kubenswrapper[4897]: I1121 15:19:03.645919 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gcpmv"] Nov 21 15:19:04 crc kubenswrapper[4897]: I1121 15:19:04.002792 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gcpmv" event={"ID":"dcbb3655-443e-4ef1-9789-2f4d956120fc","Type":"ContainerStarted","Data":"c04c98a7a77dbf92c22214f2218026e25274de0b80208cbdde2ce9e69f4a62aa"} Nov 21 15:19:05 crc kubenswrapper[4897]: I1121 15:19:05.030548 4897 generic.go:334] "Generic (PLEG): container finished" podID="dcbb3655-443e-4ef1-9789-2f4d956120fc" containerID="01503ee47e781daf912326d771fef8ec10c15cbb44937a63e4fd9c1d82e53b44" exitCode=0 Nov 21 15:19:05 crc kubenswrapper[4897]: I1121 15:19:05.030765 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gcpmv" event={"ID":"dcbb3655-443e-4ef1-9789-2f4d956120fc","Type":"ContainerDied","Data":"01503ee47e781daf912326d771fef8ec10c15cbb44937a63e4fd9c1d82e53b44"} Nov 21 15:19:05 crc kubenswrapper[4897]: I1121 15:19:05.033386 4897 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 21 15:19:05 crc kubenswrapper[4897]: I1121 15:19:05.089536 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-hgmvr"] Nov 21 15:19:05 crc kubenswrapper[4897]: I1121 15:19:05.092188 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hgmvr" Nov 21 15:19:05 crc kubenswrapper[4897]: I1121 15:19:05.112062 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hgmvr"] Nov 21 15:19:05 crc kubenswrapper[4897]: I1121 15:19:05.225069 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/65094879-bfe5-49e3-8906-f97edd0a5d8c-utilities\") pod \"certified-operators-hgmvr\" (UID: \"65094879-bfe5-49e3-8906-f97edd0a5d8c\") " pod="openshift-marketplace/certified-operators-hgmvr" Nov 21 15:19:05 crc kubenswrapper[4897]: I1121 15:19:05.225141 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/65094879-bfe5-49e3-8906-f97edd0a5d8c-catalog-content\") pod \"certified-operators-hgmvr\" (UID: \"65094879-bfe5-49e3-8906-f97edd0a5d8c\") " pod="openshift-marketplace/certified-operators-hgmvr" Nov 21 15:19:05 crc kubenswrapper[4897]: I1121 15:19:05.225365 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sv7fd\" (UniqueName: \"kubernetes.io/projected/65094879-bfe5-49e3-8906-f97edd0a5d8c-kube-api-access-sv7fd\") pod \"certified-operators-hgmvr\" (UID: \"65094879-bfe5-49e3-8906-f97edd0a5d8c\") " pod="openshift-marketplace/certified-operators-hgmvr" Nov 21 15:19:05 crc kubenswrapper[4897]: I1121 15:19:05.328677 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/65094879-bfe5-49e3-8906-f97edd0a5d8c-utilities\") pod \"certified-operators-hgmvr\" (UID: \"65094879-bfe5-49e3-8906-f97edd0a5d8c\") " pod="openshift-marketplace/certified-operators-hgmvr" Nov 21 15:19:05 crc kubenswrapper[4897]: I1121 15:19:05.330038 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/65094879-bfe5-49e3-8906-f97edd0a5d8c-catalog-content\") pod \"certified-operators-hgmvr\" (UID: \"65094879-bfe5-49e3-8906-f97edd0a5d8c\") " pod="openshift-marketplace/certified-operators-hgmvr" Nov 21 15:19:05 crc kubenswrapper[4897]: I1121 15:19:05.329199 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/65094879-bfe5-49e3-8906-f97edd0a5d8c-utilities\") pod \"certified-operators-hgmvr\" (UID: \"65094879-bfe5-49e3-8906-f97edd0a5d8c\") " pod="openshift-marketplace/certified-operators-hgmvr" Nov 21 15:19:05 crc kubenswrapper[4897]: I1121 15:19:05.330290 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sv7fd\" (UniqueName: \"kubernetes.io/projected/65094879-bfe5-49e3-8906-f97edd0a5d8c-kube-api-access-sv7fd\") pod \"certified-operators-hgmvr\" (UID: \"65094879-bfe5-49e3-8906-f97edd0a5d8c\") " pod="openshift-marketplace/certified-operators-hgmvr" Nov 21 15:19:05 crc kubenswrapper[4897]: I1121 15:19:05.330631 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/65094879-bfe5-49e3-8906-f97edd0a5d8c-catalog-content\") pod \"certified-operators-hgmvr\" (UID: \"65094879-bfe5-49e3-8906-f97edd0a5d8c\") " pod="openshift-marketplace/certified-operators-hgmvr" Nov 21 15:19:05 crc kubenswrapper[4897]: I1121 15:19:05.363059 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sv7fd\" (UniqueName: \"kubernetes.io/projected/65094879-bfe5-49e3-8906-f97edd0a5d8c-kube-api-access-sv7fd\") pod \"certified-operators-hgmvr\" (UID: \"65094879-bfe5-49e3-8906-f97edd0a5d8c\") " pod="openshift-marketplace/certified-operators-hgmvr" Nov 21 15:19:05 crc kubenswrapper[4897]: I1121 15:19:05.425060 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hgmvr" Nov 21 15:19:05 crc kubenswrapper[4897]: I1121 15:19:05.986732 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hgmvr"] Nov 21 15:19:05 crc kubenswrapper[4897]: W1121 15:19:05.987589 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod65094879_bfe5_49e3_8906_f97edd0a5d8c.slice/crio-5f9b0a187873eee52e0ee439181cafa634f3c48c37e28b7ade38fe53e5441a9d WatchSource:0}: Error finding container 5f9b0a187873eee52e0ee439181cafa634f3c48c37e28b7ade38fe53e5441a9d: Status 404 returned error can't find the container with id 5f9b0a187873eee52e0ee439181cafa634f3c48c37e28b7ade38fe53e5441a9d Nov 21 15:19:06 crc kubenswrapper[4897]: I1121 15:19:06.043184 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hgmvr" event={"ID":"65094879-bfe5-49e3-8906-f97edd0a5d8c","Type":"ContainerStarted","Data":"5f9b0a187873eee52e0ee439181cafa634f3c48c37e28b7ade38fe53e5441a9d"} Nov 21 15:19:07 crc kubenswrapper[4897]: I1121 15:19:07.055794 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gcpmv" event={"ID":"dcbb3655-443e-4ef1-9789-2f4d956120fc","Type":"ContainerStarted","Data":"a4b2f3b349ae5868caf73e4cec6dda7d8a4f9ab10ea334c18407e3bc9aa2d072"} Nov 21 15:19:07 crc kubenswrapper[4897]: I1121 15:19:07.058452 4897 generic.go:334] "Generic (PLEG): container finished" podID="65094879-bfe5-49e3-8906-f97edd0a5d8c" containerID="2f63de24e11156c3723afd15daef46d6695e7838c6137c6ab56cbc1015a9c0ae" exitCode=0 Nov 21 15:19:07 crc kubenswrapper[4897]: I1121 15:19:07.058483 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hgmvr" event={"ID":"65094879-bfe5-49e3-8906-f97edd0a5d8c","Type":"ContainerDied","Data":"2f63de24e11156c3723afd15daef46d6695e7838c6137c6ab56cbc1015a9c0ae"} Nov 21 15:19:09 crc kubenswrapper[4897]: I1121 15:19:09.081393 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hgmvr" event={"ID":"65094879-bfe5-49e3-8906-f97edd0a5d8c","Type":"ContainerStarted","Data":"7bf011426c9b3975667bac37c46316d30968086db9a4a3f88022eb9cf75503f7"} Nov 21 15:19:14 crc kubenswrapper[4897]: I1121 15:19:14.130151 4897 generic.go:334] "Generic (PLEG): container finished" podID="65094879-bfe5-49e3-8906-f97edd0a5d8c" containerID="7bf011426c9b3975667bac37c46316d30968086db9a4a3f88022eb9cf75503f7" exitCode=0 Nov 21 15:19:14 crc kubenswrapper[4897]: I1121 15:19:14.130238 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hgmvr" event={"ID":"65094879-bfe5-49e3-8906-f97edd0a5d8c","Type":"ContainerDied","Data":"7bf011426c9b3975667bac37c46316d30968086db9a4a3f88022eb9cf75503f7"} Nov 21 15:19:16 crc kubenswrapper[4897]: I1121 15:19:16.155999 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hgmvr" event={"ID":"65094879-bfe5-49e3-8906-f97edd0a5d8c","Type":"ContainerStarted","Data":"de805b3572f8c897092ec3d1a7a854e9eefb965530e1729c994bfde2ce5d48cc"} Nov 21 15:19:17 crc kubenswrapper[4897]: I1121 15:19:17.169805 4897 generic.go:334] "Generic (PLEG): container finished" podID="dcbb3655-443e-4ef1-9789-2f4d956120fc" containerID="a4b2f3b349ae5868caf73e4cec6dda7d8a4f9ab10ea334c18407e3bc9aa2d072" exitCode=0 Nov 21 15:19:17 crc kubenswrapper[4897]: I1121 15:19:17.170128 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gcpmv" event={"ID":"dcbb3655-443e-4ef1-9789-2f4d956120fc","Type":"ContainerDied","Data":"a4b2f3b349ae5868caf73e4cec6dda7d8a4f9ab10ea334c18407e3bc9aa2d072"} Nov 21 15:19:17 crc kubenswrapper[4897]: I1121 15:19:17.193878 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-hgmvr" podStartSLOduration=4.530696474 podStartE2EDuration="12.193855307s" podCreationTimestamp="2025-11-21 15:19:05 +0000 UTC" firstStartedPulling="2025-11-21 15:19:07.061722133 +0000 UTC m=+4224.346315608" lastFinishedPulling="2025-11-21 15:19:14.724880966 +0000 UTC m=+4232.009474441" observedRunningTime="2025-11-21 15:19:16.182215903 +0000 UTC m=+4233.466809378" watchObservedRunningTime="2025-11-21 15:19:17.193855307 +0000 UTC m=+4234.478448782" Nov 21 15:19:19 crc kubenswrapper[4897]: I1121 15:19:19.193499 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gcpmv" event={"ID":"dcbb3655-443e-4ef1-9789-2f4d956120fc","Type":"ContainerStarted","Data":"ffd186191ab231cc9d4fca8b01153f99931d5a31724c71887dc1e04c3dce1158"} Nov 21 15:19:19 crc kubenswrapper[4897]: I1121 15:19:19.222227 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-gcpmv" podStartSLOduration=3.919468466 podStartE2EDuration="17.222200934s" podCreationTimestamp="2025-11-21 15:19:02 +0000 UTC" firstStartedPulling="2025-11-21 15:19:05.03316831 +0000 UTC m=+4222.317761785" lastFinishedPulling="2025-11-21 15:19:18.335900778 +0000 UTC m=+4235.620494253" observedRunningTime="2025-11-21 15:19:19.21178973 +0000 UTC m=+4236.496383225" watchObservedRunningTime="2025-11-21 15:19:19.222200934 +0000 UTC m=+4236.506794409" Nov 21 15:19:23 crc kubenswrapper[4897]: I1121 15:19:23.032550 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-gcpmv" Nov 21 15:19:23 crc kubenswrapper[4897]: I1121 15:19:23.034370 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-gcpmv" Nov 21 15:19:24 crc kubenswrapper[4897]: I1121 15:19:24.095827 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-gcpmv" podUID="dcbb3655-443e-4ef1-9789-2f4d956120fc" containerName="registry-server" probeResult="failure" output=< Nov 21 15:19:24 crc kubenswrapper[4897]: timeout: failed to connect service ":50051" within 1s Nov 21 15:19:24 crc kubenswrapper[4897]: > Nov 21 15:19:25 crc kubenswrapper[4897]: I1121 15:19:25.425795 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-hgmvr" Nov 21 15:19:25 crc kubenswrapper[4897]: I1121 15:19:25.426866 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-hgmvr" Nov 21 15:19:25 crc kubenswrapper[4897]: I1121 15:19:25.483960 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-hgmvr" Nov 21 15:19:26 crc kubenswrapper[4897]: I1121 15:19:26.307669 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-hgmvr" Nov 21 15:19:26 crc kubenswrapper[4897]: I1121 15:19:26.356800 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hgmvr"] Nov 21 15:19:28 crc kubenswrapper[4897]: I1121 15:19:28.283877 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-hgmvr" podUID="65094879-bfe5-49e3-8906-f97edd0a5d8c" containerName="registry-server" containerID="cri-o://de805b3572f8c897092ec3d1a7a854e9eefb965530e1729c994bfde2ce5d48cc" gracePeriod=2 Nov 21 15:19:28 crc kubenswrapper[4897]: I1121 15:19:28.881656 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hgmvr" Nov 21 15:19:28 crc kubenswrapper[4897]: I1121 15:19:28.929422 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/65094879-bfe5-49e3-8906-f97edd0a5d8c-catalog-content\") pod \"65094879-bfe5-49e3-8906-f97edd0a5d8c\" (UID: \"65094879-bfe5-49e3-8906-f97edd0a5d8c\") " Nov 21 15:19:28 crc kubenswrapper[4897]: I1121 15:19:28.929554 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/65094879-bfe5-49e3-8906-f97edd0a5d8c-utilities\") pod \"65094879-bfe5-49e3-8906-f97edd0a5d8c\" (UID: \"65094879-bfe5-49e3-8906-f97edd0a5d8c\") " Nov 21 15:19:28 crc kubenswrapper[4897]: I1121 15:19:28.929790 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sv7fd\" (UniqueName: \"kubernetes.io/projected/65094879-bfe5-49e3-8906-f97edd0a5d8c-kube-api-access-sv7fd\") pod \"65094879-bfe5-49e3-8906-f97edd0a5d8c\" (UID: \"65094879-bfe5-49e3-8906-f97edd0a5d8c\") " Nov 21 15:19:28 crc kubenswrapper[4897]: I1121 15:19:28.930458 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/65094879-bfe5-49e3-8906-f97edd0a5d8c-utilities" (OuterVolumeSpecName: "utilities") pod "65094879-bfe5-49e3-8906-f97edd0a5d8c" (UID: "65094879-bfe5-49e3-8906-f97edd0a5d8c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 15:19:28 crc kubenswrapper[4897]: I1121 15:19:28.935473 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/65094879-bfe5-49e3-8906-f97edd0a5d8c-kube-api-access-sv7fd" (OuterVolumeSpecName: "kube-api-access-sv7fd") pod "65094879-bfe5-49e3-8906-f97edd0a5d8c" (UID: "65094879-bfe5-49e3-8906-f97edd0a5d8c"). InnerVolumeSpecName "kube-api-access-sv7fd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 15:19:28 crc kubenswrapper[4897]: I1121 15:19:28.977352 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/65094879-bfe5-49e3-8906-f97edd0a5d8c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "65094879-bfe5-49e3-8906-f97edd0a5d8c" (UID: "65094879-bfe5-49e3-8906-f97edd0a5d8c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 15:19:29 crc kubenswrapper[4897]: I1121 15:19:29.033085 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sv7fd\" (UniqueName: \"kubernetes.io/projected/65094879-bfe5-49e3-8906-f97edd0a5d8c-kube-api-access-sv7fd\") on node \"crc\" DevicePath \"\"" Nov 21 15:19:29 crc kubenswrapper[4897]: I1121 15:19:29.033129 4897 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/65094879-bfe5-49e3-8906-f97edd0a5d8c-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 21 15:19:29 crc kubenswrapper[4897]: I1121 15:19:29.033141 4897 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/65094879-bfe5-49e3-8906-f97edd0a5d8c-utilities\") on node \"crc\" DevicePath \"\"" Nov 21 15:19:29 crc kubenswrapper[4897]: I1121 15:19:29.297770 4897 generic.go:334] "Generic (PLEG): container finished" podID="65094879-bfe5-49e3-8906-f97edd0a5d8c" containerID="de805b3572f8c897092ec3d1a7a854e9eefb965530e1729c994bfde2ce5d48cc" exitCode=0 Nov 21 15:19:29 crc kubenswrapper[4897]: I1121 15:19:29.298005 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hgmvr" event={"ID":"65094879-bfe5-49e3-8906-f97edd0a5d8c","Type":"ContainerDied","Data":"de805b3572f8c897092ec3d1a7a854e9eefb965530e1729c994bfde2ce5d48cc"} Nov 21 15:19:29 crc kubenswrapper[4897]: I1121 15:19:29.298148 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hgmvr" Nov 21 15:19:29 crc kubenswrapper[4897]: I1121 15:19:29.298149 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hgmvr" event={"ID":"65094879-bfe5-49e3-8906-f97edd0a5d8c","Type":"ContainerDied","Data":"5f9b0a187873eee52e0ee439181cafa634f3c48c37e28b7ade38fe53e5441a9d"} Nov 21 15:19:29 crc kubenswrapper[4897]: I1121 15:19:29.298165 4897 scope.go:117] "RemoveContainer" containerID="de805b3572f8c897092ec3d1a7a854e9eefb965530e1729c994bfde2ce5d48cc" Nov 21 15:19:29 crc kubenswrapper[4897]: I1121 15:19:29.320892 4897 scope.go:117] "RemoveContainer" containerID="7bf011426c9b3975667bac37c46316d30968086db9a4a3f88022eb9cf75503f7" Nov 21 15:19:29 crc kubenswrapper[4897]: I1121 15:19:29.338603 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hgmvr"] Nov 21 15:19:29 crc kubenswrapper[4897]: I1121 15:19:29.348655 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-hgmvr"] Nov 21 15:19:29 crc kubenswrapper[4897]: I1121 15:19:29.363867 4897 scope.go:117] "RemoveContainer" containerID="2f63de24e11156c3723afd15daef46d6695e7838c6137c6ab56cbc1015a9c0ae" Nov 21 15:19:29 crc kubenswrapper[4897]: I1121 15:19:29.422772 4897 scope.go:117] "RemoveContainer" containerID="de805b3572f8c897092ec3d1a7a854e9eefb965530e1729c994bfde2ce5d48cc" Nov 21 15:19:29 crc kubenswrapper[4897]: E1121 15:19:29.426730 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de805b3572f8c897092ec3d1a7a854e9eefb965530e1729c994bfde2ce5d48cc\": container with ID starting with de805b3572f8c897092ec3d1a7a854e9eefb965530e1729c994bfde2ce5d48cc not found: ID does not exist" containerID="de805b3572f8c897092ec3d1a7a854e9eefb965530e1729c994bfde2ce5d48cc" Nov 21 15:19:29 crc kubenswrapper[4897]: I1121 15:19:29.426781 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de805b3572f8c897092ec3d1a7a854e9eefb965530e1729c994bfde2ce5d48cc"} err="failed to get container status \"de805b3572f8c897092ec3d1a7a854e9eefb965530e1729c994bfde2ce5d48cc\": rpc error: code = NotFound desc = could not find container \"de805b3572f8c897092ec3d1a7a854e9eefb965530e1729c994bfde2ce5d48cc\": container with ID starting with de805b3572f8c897092ec3d1a7a854e9eefb965530e1729c994bfde2ce5d48cc not found: ID does not exist" Nov 21 15:19:29 crc kubenswrapper[4897]: I1121 15:19:29.426809 4897 scope.go:117] "RemoveContainer" containerID="7bf011426c9b3975667bac37c46316d30968086db9a4a3f88022eb9cf75503f7" Nov 21 15:19:29 crc kubenswrapper[4897]: E1121 15:19:29.427432 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7bf011426c9b3975667bac37c46316d30968086db9a4a3f88022eb9cf75503f7\": container with ID starting with 7bf011426c9b3975667bac37c46316d30968086db9a4a3f88022eb9cf75503f7 not found: ID does not exist" containerID="7bf011426c9b3975667bac37c46316d30968086db9a4a3f88022eb9cf75503f7" Nov 21 15:19:29 crc kubenswrapper[4897]: I1121 15:19:29.427531 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7bf011426c9b3975667bac37c46316d30968086db9a4a3f88022eb9cf75503f7"} err="failed to get container status \"7bf011426c9b3975667bac37c46316d30968086db9a4a3f88022eb9cf75503f7\": rpc error: code = NotFound desc = could not find container \"7bf011426c9b3975667bac37c46316d30968086db9a4a3f88022eb9cf75503f7\": container with ID starting with 7bf011426c9b3975667bac37c46316d30968086db9a4a3f88022eb9cf75503f7 not found: ID does not exist" Nov 21 15:19:29 crc kubenswrapper[4897]: I1121 15:19:29.427593 4897 scope.go:117] "RemoveContainer" containerID="2f63de24e11156c3723afd15daef46d6695e7838c6137c6ab56cbc1015a9c0ae" Nov 21 15:19:29 crc kubenswrapper[4897]: E1121 15:19:29.428220 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2f63de24e11156c3723afd15daef46d6695e7838c6137c6ab56cbc1015a9c0ae\": container with ID starting with 2f63de24e11156c3723afd15daef46d6695e7838c6137c6ab56cbc1015a9c0ae not found: ID does not exist" containerID="2f63de24e11156c3723afd15daef46d6695e7838c6137c6ab56cbc1015a9c0ae" Nov 21 15:19:29 crc kubenswrapper[4897]: I1121 15:19:29.428278 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f63de24e11156c3723afd15daef46d6695e7838c6137c6ab56cbc1015a9c0ae"} err="failed to get container status \"2f63de24e11156c3723afd15daef46d6695e7838c6137c6ab56cbc1015a9c0ae\": rpc error: code = NotFound desc = could not find container \"2f63de24e11156c3723afd15daef46d6695e7838c6137c6ab56cbc1015a9c0ae\": container with ID starting with 2f63de24e11156c3723afd15daef46d6695e7838c6137c6ab56cbc1015a9c0ae not found: ID does not exist" Nov 21 15:19:30 crc kubenswrapper[4897]: I1121 15:19:30.100897 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="65094879-bfe5-49e3-8906-f97edd0a5d8c" path="/var/lib/kubelet/pods/65094879-bfe5-49e3-8906-f97edd0a5d8c/volumes" Nov 21 15:19:34 crc kubenswrapper[4897]: I1121 15:19:34.087760 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-gcpmv" podUID="dcbb3655-443e-4ef1-9789-2f4d956120fc" containerName="registry-server" probeResult="failure" output=< Nov 21 15:19:34 crc kubenswrapper[4897]: timeout: failed to connect service ":50051" within 1s Nov 21 15:19:34 crc kubenswrapper[4897]: > Nov 21 15:19:43 crc kubenswrapper[4897]: I1121 15:19:43.099143 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-gcpmv" Nov 21 15:19:43 crc kubenswrapper[4897]: I1121 15:19:43.175149 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-gcpmv" Nov 21 15:19:43 crc kubenswrapper[4897]: I1121 15:19:43.350334 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-gcpmv"] Nov 21 15:19:44 crc kubenswrapper[4897]: I1121 15:19:44.460502 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-gcpmv" podUID="dcbb3655-443e-4ef1-9789-2f4d956120fc" containerName="registry-server" containerID="cri-o://ffd186191ab231cc9d4fca8b01153f99931d5a31724c71887dc1e04c3dce1158" gracePeriod=2 Nov 21 15:19:45 crc kubenswrapper[4897]: I1121 15:19:45.120462 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gcpmv" Nov 21 15:19:45 crc kubenswrapper[4897]: I1121 15:19:45.158035 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qtd2d\" (UniqueName: \"kubernetes.io/projected/dcbb3655-443e-4ef1-9789-2f4d956120fc-kube-api-access-qtd2d\") pod \"dcbb3655-443e-4ef1-9789-2f4d956120fc\" (UID: \"dcbb3655-443e-4ef1-9789-2f4d956120fc\") " Nov 21 15:19:45 crc kubenswrapper[4897]: I1121 15:19:45.158141 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dcbb3655-443e-4ef1-9789-2f4d956120fc-catalog-content\") pod \"dcbb3655-443e-4ef1-9789-2f4d956120fc\" (UID: \"dcbb3655-443e-4ef1-9789-2f4d956120fc\") " Nov 21 15:19:45 crc kubenswrapper[4897]: I1121 15:19:45.158276 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dcbb3655-443e-4ef1-9789-2f4d956120fc-utilities\") pod \"dcbb3655-443e-4ef1-9789-2f4d956120fc\" (UID: \"dcbb3655-443e-4ef1-9789-2f4d956120fc\") " Nov 21 15:19:45 crc kubenswrapper[4897]: I1121 15:19:45.161328 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dcbb3655-443e-4ef1-9789-2f4d956120fc-utilities" (OuterVolumeSpecName: "utilities") pod "dcbb3655-443e-4ef1-9789-2f4d956120fc" (UID: "dcbb3655-443e-4ef1-9789-2f4d956120fc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 15:19:45 crc kubenswrapper[4897]: I1121 15:19:45.166723 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dcbb3655-443e-4ef1-9789-2f4d956120fc-kube-api-access-qtd2d" (OuterVolumeSpecName: "kube-api-access-qtd2d") pod "dcbb3655-443e-4ef1-9789-2f4d956120fc" (UID: "dcbb3655-443e-4ef1-9789-2f4d956120fc"). InnerVolumeSpecName "kube-api-access-qtd2d". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 15:19:45 crc kubenswrapper[4897]: I1121 15:19:45.262364 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qtd2d\" (UniqueName: \"kubernetes.io/projected/dcbb3655-443e-4ef1-9789-2f4d956120fc-kube-api-access-qtd2d\") on node \"crc\" DevicePath \"\"" Nov 21 15:19:45 crc kubenswrapper[4897]: I1121 15:19:45.262400 4897 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dcbb3655-443e-4ef1-9789-2f4d956120fc-utilities\") on node \"crc\" DevicePath \"\"" Nov 21 15:19:45 crc kubenswrapper[4897]: I1121 15:19:45.267270 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dcbb3655-443e-4ef1-9789-2f4d956120fc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "dcbb3655-443e-4ef1-9789-2f4d956120fc" (UID: "dcbb3655-443e-4ef1-9789-2f4d956120fc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 15:19:45 crc kubenswrapper[4897]: I1121 15:19:45.364681 4897 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dcbb3655-443e-4ef1-9789-2f4d956120fc-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 21 15:19:45 crc kubenswrapper[4897]: I1121 15:19:45.476179 4897 generic.go:334] "Generic (PLEG): container finished" podID="dcbb3655-443e-4ef1-9789-2f4d956120fc" containerID="ffd186191ab231cc9d4fca8b01153f99931d5a31724c71887dc1e04c3dce1158" exitCode=0 Nov 21 15:19:45 crc kubenswrapper[4897]: I1121 15:19:45.476230 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gcpmv" event={"ID":"dcbb3655-443e-4ef1-9789-2f4d956120fc","Type":"ContainerDied","Data":"ffd186191ab231cc9d4fca8b01153f99931d5a31724c71887dc1e04c3dce1158"} Nov 21 15:19:45 crc kubenswrapper[4897]: I1121 15:19:45.476270 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gcpmv" event={"ID":"dcbb3655-443e-4ef1-9789-2f4d956120fc","Type":"ContainerDied","Data":"c04c98a7a77dbf92c22214f2218026e25274de0b80208cbdde2ce9e69f4a62aa"} Nov 21 15:19:45 crc kubenswrapper[4897]: I1121 15:19:45.476290 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gcpmv" Nov 21 15:19:45 crc kubenswrapper[4897]: I1121 15:19:45.476324 4897 scope.go:117] "RemoveContainer" containerID="ffd186191ab231cc9d4fca8b01153f99931d5a31724c71887dc1e04c3dce1158" Nov 21 15:19:45 crc kubenswrapper[4897]: I1121 15:19:45.515781 4897 scope.go:117] "RemoveContainer" containerID="a4b2f3b349ae5868caf73e4cec6dda7d8a4f9ab10ea334c18407e3bc9aa2d072" Nov 21 15:19:45 crc kubenswrapper[4897]: I1121 15:19:45.524687 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-gcpmv"] Nov 21 15:19:45 crc kubenswrapper[4897]: I1121 15:19:45.536403 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-gcpmv"] Nov 21 15:19:45 crc kubenswrapper[4897]: I1121 15:19:45.567825 4897 scope.go:117] "RemoveContainer" containerID="01503ee47e781daf912326d771fef8ec10c15cbb44937a63e4fd9c1d82e53b44" Nov 21 15:19:45 crc kubenswrapper[4897]: I1121 15:19:45.622784 4897 scope.go:117] "RemoveContainer" containerID="ffd186191ab231cc9d4fca8b01153f99931d5a31724c71887dc1e04c3dce1158" Nov 21 15:19:45 crc kubenswrapper[4897]: E1121 15:19:45.627734 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ffd186191ab231cc9d4fca8b01153f99931d5a31724c71887dc1e04c3dce1158\": container with ID starting with ffd186191ab231cc9d4fca8b01153f99931d5a31724c71887dc1e04c3dce1158 not found: ID does not exist" containerID="ffd186191ab231cc9d4fca8b01153f99931d5a31724c71887dc1e04c3dce1158" Nov 21 15:19:45 crc kubenswrapper[4897]: I1121 15:19:45.627786 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ffd186191ab231cc9d4fca8b01153f99931d5a31724c71887dc1e04c3dce1158"} err="failed to get container status \"ffd186191ab231cc9d4fca8b01153f99931d5a31724c71887dc1e04c3dce1158\": rpc error: code = NotFound desc = could not find container \"ffd186191ab231cc9d4fca8b01153f99931d5a31724c71887dc1e04c3dce1158\": container with ID starting with ffd186191ab231cc9d4fca8b01153f99931d5a31724c71887dc1e04c3dce1158 not found: ID does not exist" Nov 21 15:19:45 crc kubenswrapper[4897]: I1121 15:19:45.627828 4897 scope.go:117] "RemoveContainer" containerID="a4b2f3b349ae5868caf73e4cec6dda7d8a4f9ab10ea334c18407e3bc9aa2d072" Nov 21 15:19:45 crc kubenswrapper[4897]: E1121 15:19:45.628506 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a4b2f3b349ae5868caf73e4cec6dda7d8a4f9ab10ea334c18407e3bc9aa2d072\": container with ID starting with a4b2f3b349ae5868caf73e4cec6dda7d8a4f9ab10ea334c18407e3bc9aa2d072 not found: ID does not exist" containerID="a4b2f3b349ae5868caf73e4cec6dda7d8a4f9ab10ea334c18407e3bc9aa2d072" Nov 21 15:19:45 crc kubenswrapper[4897]: I1121 15:19:45.628582 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4b2f3b349ae5868caf73e4cec6dda7d8a4f9ab10ea334c18407e3bc9aa2d072"} err="failed to get container status \"a4b2f3b349ae5868caf73e4cec6dda7d8a4f9ab10ea334c18407e3bc9aa2d072\": rpc error: code = NotFound desc = could not find container \"a4b2f3b349ae5868caf73e4cec6dda7d8a4f9ab10ea334c18407e3bc9aa2d072\": container with ID starting with a4b2f3b349ae5868caf73e4cec6dda7d8a4f9ab10ea334c18407e3bc9aa2d072 not found: ID does not exist" Nov 21 15:19:45 crc kubenswrapper[4897]: I1121 15:19:45.628614 4897 scope.go:117] "RemoveContainer" containerID="01503ee47e781daf912326d771fef8ec10c15cbb44937a63e4fd9c1d82e53b44" Nov 21 15:19:45 crc kubenswrapper[4897]: E1121 15:19:45.630641 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"01503ee47e781daf912326d771fef8ec10c15cbb44937a63e4fd9c1d82e53b44\": container with ID starting with 01503ee47e781daf912326d771fef8ec10c15cbb44937a63e4fd9c1d82e53b44 not found: ID does not exist" containerID="01503ee47e781daf912326d771fef8ec10c15cbb44937a63e4fd9c1d82e53b44" Nov 21 15:19:45 crc kubenswrapper[4897]: I1121 15:19:45.630686 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01503ee47e781daf912326d771fef8ec10c15cbb44937a63e4fd9c1d82e53b44"} err="failed to get container status \"01503ee47e781daf912326d771fef8ec10c15cbb44937a63e4fd9c1d82e53b44\": rpc error: code = NotFound desc = could not find container \"01503ee47e781daf912326d771fef8ec10c15cbb44937a63e4fd9c1d82e53b44\": container with ID starting with 01503ee47e781daf912326d771fef8ec10c15cbb44937a63e4fd9c1d82e53b44 not found: ID does not exist" Nov 21 15:19:46 crc kubenswrapper[4897]: I1121 15:19:46.107665 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dcbb3655-443e-4ef1-9789-2f4d956120fc" path="/var/lib/kubelet/pods/dcbb3655-443e-4ef1-9789-2f4d956120fc/volumes" Nov 21 15:20:34 crc kubenswrapper[4897]: I1121 15:20:34.370996 4897 patch_prober.go:28] interesting pod/machine-config-daemon-krv5b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 21 15:20:34 crc kubenswrapper[4897]: I1121 15:20:34.371545 4897 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 21 15:21:04 crc kubenswrapper[4897]: I1121 15:21:04.370905 4897 patch_prober.go:28] interesting pod/machine-config-daemon-krv5b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 21 15:21:04 crc kubenswrapper[4897]: I1121 15:21:04.372154 4897 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 21 15:21:19 crc kubenswrapper[4897]: I1121 15:21:19.717923 4897 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="c628be92-b88f-4e02-a981-4b8fe6fd7c13" containerName="ceilometer-central-agent" probeResult="failure" output="command timed out" Nov 21 15:21:20 crc kubenswrapper[4897]: I1121 15:21:20.667830 4897 patch_prober.go:28] interesting pod/console-operator-58897d9998-dfltg container/console-operator namespace/openshift-console-operator: Liveness probe status=failure output="Get \"https://10.217.0.10:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 21 15:21:20 crc kubenswrapper[4897]: I1121 15:21:20.667894 4897 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console-operator/console-operator-58897d9998-dfltg" podUID="18d8d58b-3dab-4949-b9d1-313870a96dd5" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.10:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 21 15:21:21 crc kubenswrapper[4897]: I1121 15:21:21.295620 4897 patch_prober.go:28] interesting pod/authentication-operator-69f744f599-8kddm container/authentication-operator namespace/openshift-authentication-operator: Liveness probe status=failure output="Get \"https://10.217.0.17:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 21 15:21:21 crc kubenswrapper[4897]: I1121 15:21:21.295987 4897 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-authentication-operator/authentication-operator-69f744f599-8kddm" podUID="a59737d6-7609-4077-8fb8-d925f25a65f8" containerName="authentication-operator" probeResult="failure" output="Get \"https://10.217.0.17:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Nov 21 15:21:23 crc kubenswrapper[4897]: I1121 15:21:23.766676 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-cr58g" podUID="e3891c15-c0b9-4965-84fe-05262fe71853" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.103:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 21 15:21:34 crc kubenswrapper[4897]: I1121 15:21:34.370650 4897 patch_prober.go:28] interesting pod/machine-config-daemon-krv5b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 21 15:21:34 crc kubenswrapper[4897]: I1121 15:21:34.371289 4897 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 21 15:21:34 crc kubenswrapper[4897]: I1121 15:21:34.371343 4897 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" Nov 21 15:21:34 crc kubenswrapper[4897]: I1121 15:21:34.372408 4897 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"95e70a2cadfac38d1f3f3ba5e655bb58f1dc4e2d41e3d3c5a54f91a8584b921e"} pod="openshift-machine-config-operator/machine-config-daemon-krv5b" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 21 15:21:34 crc kubenswrapper[4897]: I1121 15:21:34.372471 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" containerName="machine-config-daemon" containerID="cri-o://95e70a2cadfac38d1f3f3ba5e655bb58f1dc4e2d41e3d3c5a54f91a8584b921e" gracePeriod=600 Nov 21 15:21:35 crc kubenswrapper[4897]: E1121 15:21:35.063249 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:21:35 crc kubenswrapper[4897]: I1121 15:21:35.688264 4897 generic.go:334] "Generic (PLEG): container finished" podID="e7670227-d280-4847-b882-754429f56b0f" containerID="95e70a2cadfac38d1f3f3ba5e655bb58f1dc4e2d41e3d3c5a54f91a8584b921e" exitCode=0 Nov 21 15:21:35 crc kubenswrapper[4897]: I1121 15:21:35.688333 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" event={"ID":"e7670227-d280-4847-b882-754429f56b0f","Type":"ContainerDied","Data":"95e70a2cadfac38d1f3f3ba5e655bb58f1dc4e2d41e3d3c5a54f91a8584b921e"} Nov 21 15:21:35 crc kubenswrapper[4897]: I1121 15:21:35.688382 4897 scope.go:117] "RemoveContainer" containerID="298866e849258a62c75c2157ebb87d240383aec553c6cd2bc7b23079d1c5505d" Nov 21 15:21:35 crc kubenswrapper[4897]: I1121 15:21:35.690225 4897 scope.go:117] "RemoveContainer" containerID="95e70a2cadfac38d1f3f3ba5e655bb58f1dc4e2d41e3d3c5a54f91a8584b921e" Nov 21 15:21:35 crc kubenswrapper[4897]: E1121 15:21:35.690892 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:21:48 crc kubenswrapper[4897]: I1121 15:21:48.089779 4897 scope.go:117] "RemoveContainer" containerID="95e70a2cadfac38d1f3f3ba5e655bb58f1dc4e2d41e3d3c5a54f91a8584b921e" Nov 21 15:21:48 crc kubenswrapper[4897]: E1121 15:21:48.090473 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:22:01 crc kubenswrapper[4897]: I1121 15:22:01.089884 4897 scope.go:117] "RemoveContainer" containerID="95e70a2cadfac38d1f3f3ba5e655bb58f1dc4e2d41e3d3c5a54f91a8584b921e" Nov 21 15:22:01 crc kubenswrapper[4897]: E1121 15:22:01.090671 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:22:15 crc kubenswrapper[4897]: I1121 15:22:15.090581 4897 scope.go:117] "RemoveContainer" containerID="95e70a2cadfac38d1f3f3ba5e655bb58f1dc4e2d41e3d3c5a54f91a8584b921e" Nov 21 15:22:15 crc kubenswrapper[4897]: E1121 15:22:15.091467 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:22:30 crc kubenswrapper[4897]: I1121 15:22:30.089145 4897 scope.go:117] "RemoveContainer" containerID="95e70a2cadfac38d1f3f3ba5e655bb58f1dc4e2d41e3d3c5a54f91a8584b921e" Nov 21 15:22:30 crc kubenswrapper[4897]: E1121 15:22:30.090133 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:22:42 crc kubenswrapper[4897]: I1121 15:22:42.089351 4897 scope.go:117] "RemoveContainer" containerID="95e70a2cadfac38d1f3f3ba5e655bb58f1dc4e2d41e3d3c5a54f91a8584b921e" Nov 21 15:22:42 crc kubenswrapper[4897]: E1121 15:22:42.090287 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:22:53 crc kubenswrapper[4897]: I1121 15:22:53.090261 4897 scope.go:117] "RemoveContainer" containerID="95e70a2cadfac38d1f3f3ba5e655bb58f1dc4e2d41e3d3c5a54f91a8584b921e" Nov 21 15:22:53 crc kubenswrapper[4897]: E1121 15:22:53.091395 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:23:06 crc kubenswrapper[4897]: I1121 15:23:06.089387 4897 scope.go:117] "RemoveContainer" containerID="95e70a2cadfac38d1f3f3ba5e655bb58f1dc4e2d41e3d3c5a54f91a8584b921e" Nov 21 15:23:06 crc kubenswrapper[4897]: E1121 15:23:06.090127 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:23:20 crc kubenswrapper[4897]: I1121 15:23:20.088777 4897 scope.go:117] "RemoveContainer" containerID="95e70a2cadfac38d1f3f3ba5e655bb58f1dc4e2d41e3d3c5a54f91a8584b921e" Nov 21 15:23:20 crc kubenswrapper[4897]: E1121 15:23:20.089486 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:23:31 crc kubenswrapper[4897]: I1121 15:23:31.089615 4897 scope.go:117] "RemoveContainer" containerID="95e70a2cadfac38d1f3f3ba5e655bb58f1dc4e2d41e3d3c5a54f91a8584b921e" Nov 21 15:23:31 crc kubenswrapper[4897]: E1121 15:23:31.090479 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:23:38 crc kubenswrapper[4897]: I1121 15:23:38.695690 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-td294"] Nov 21 15:23:38 crc kubenswrapper[4897]: E1121 15:23:38.697544 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65094879-bfe5-49e3-8906-f97edd0a5d8c" containerName="extract-content" Nov 21 15:23:38 crc kubenswrapper[4897]: I1121 15:23:38.697569 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="65094879-bfe5-49e3-8906-f97edd0a5d8c" containerName="extract-content" Nov 21 15:23:38 crc kubenswrapper[4897]: E1121 15:23:38.697592 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65094879-bfe5-49e3-8906-f97edd0a5d8c" containerName="extract-utilities" Nov 21 15:23:38 crc kubenswrapper[4897]: I1121 15:23:38.697599 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="65094879-bfe5-49e3-8906-f97edd0a5d8c" containerName="extract-utilities" Nov 21 15:23:38 crc kubenswrapper[4897]: E1121 15:23:38.697650 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dcbb3655-443e-4ef1-9789-2f4d956120fc" containerName="extract-content" Nov 21 15:23:38 crc kubenswrapper[4897]: I1121 15:23:38.697660 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="dcbb3655-443e-4ef1-9789-2f4d956120fc" containerName="extract-content" Nov 21 15:23:38 crc kubenswrapper[4897]: E1121 15:23:38.697672 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65094879-bfe5-49e3-8906-f97edd0a5d8c" containerName="registry-server" Nov 21 15:23:38 crc kubenswrapper[4897]: I1121 15:23:38.697679 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="65094879-bfe5-49e3-8906-f97edd0a5d8c" containerName="registry-server" Nov 21 15:23:38 crc kubenswrapper[4897]: E1121 15:23:38.697702 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dcbb3655-443e-4ef1-9789-2f4d956120fc" containerName="registry-server" Nov 21 15:23:38 crc kubenswrapper[4897]: I1121 15:23:38.697709 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="dcbb3655-443e-4ef1-9789-2f4d956120fc" containerName="registry-server" Nov 21 15:23:38 crc kubenswrapper[4897]: E1121 15:23:38.697731 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dcbb3655-443e-4ef1-9789-2f4d956120fc" containerName="extract-utilities" Nov 21 15:23:38 crc kubenswrapper[4897]: I1121 15:23:38.697739 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="dcbb3655-443e-4ef1-9789-2f4d956120fc" containerName="extract-utilities" Nov 21 15:23:38 crc kubenswrapper[4897]: I1121 15:23:38.698094 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="65094879-bfe5-49e3-8906-f97edd0a5d8c" containerName="registry-server" Nov 21 15:23:38 crc kubenswrapper[4897]: I1121 15:23:38.698138 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="dcbb3655-443e-4ef1-9789-2f4d956120fc" containerName="registry-server" Nov 21 15:23:38 crc kubenswrapper[4897]: I1121 15:23:38.706486 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-td294" Nov 21 15:23:38 crc kubenswrapper[4897]: I1121 15:23:38.713826 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-td294"] Nov 21 15:23:38 crc kubenswrapper[4897]: I1121 15:23:38.755303 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6ztsf\" (UniqueName: \"kubernetes.io/projected/9cd224da-05fd-4393-8d11-9ae21f18978e-kube-api-access-6ztsf\") pod \"community-operators-td294\" (UID: \"9cd224da-05fd-4393-8d11-9ae21f18978e\") " pod="openshift-marketplace/community-operators-td294" Nov 21 15:23:38 crc kubenswrapper[4897]: I1121 15:23:38.755404 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9cd224da-05fd-4393-8d11-9ae21f18978e-catalog-content\") pod \"community-operators-td294\" (UID: \"9cd224da-05fd-4393-8d11-9ae21f18978e\") " pod="openshift-marketplace/community-operators-td294" Nov 21 15:23:38 crc kubenswrapper[4897]: I1121 15:23:38.756127 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9cd224da-05fd-4393-8d11-9ae21f18978e-utilities\") pod \"community-operators-td294\" (UID: \"9cd224da-05fd-4393-8d11-9ae21f18978e\") " pod="openshift-marketplace/community-operators-td294" Nov 21 15:23:38 crc kubenswrapper[4897]: I1121 15:23:38.859021 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9cd224da-05fd-4393-8d11-9ae21f18978e-utilities\") pod \"community-operators-td294\" (UID: \"9cd224da-05fd-4393-8d11-9ae21f18978e\") " pod="openshift-marketplace/community-operators-td294" Nov 21 15:23:38 crc kubenswrapper[4897]: I1121 15:23:38.859176 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6ztsf\" (UniqueName: \"kubernetes.io/projected/9cd224da-05fd-4393-8d11-9ae21f18978e-kube-api-access-6ztsf\") pod \"community-operators-td294\" (UID: \"9cd224da-05fd-4393-8d11-9ae21f18978e\") " pod="openshift-marketplace/community-operators-td294" Nov 21 15:23:38 crc kubenswrapper[4897]: I1121 15:23:38.859268 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9cd224da-05fd-4393-8d11-9ae21f18978e-catalog-content\") pod \"community-operators-td294\" (UID: \"9cd224da-05fd-4393-8d11-9ae21f18978e\") " pod="openshift-marketplace/community-operators-td294" Nov 21 15:23:38 crc kubenswrapper[4897]: I1121 15:23:38.859798 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9cd224da-05fd-4393-8d11-9ae21f18978e-catalog-content\") pod \"community-operators-td294\" (UID: \"9cd224da-05fd-4393-8d11-9ae21f18978e\") " pod="openshift-marketplace/community-operators-td294" Nov 21 15:23:38 crc kubenswrapper[4897]: I1121 15:23:38.860084 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9cd224da-05fd-4393-8d11-9ae21f18978e-utilities\") pod \"community-operators-td294\" (UID: \"9cd224da-05fd-4393-8d11-9ae21f18978e\") " pod="openshift-marketplace/community-operators-td294" Nov 21 15:23:38 crc kubenswrapper[4897]: I1121 15:23:38.888427 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6ztsf\" (UniqueName: \"kubernetes.io/projected/9cd224da-05fd-4393-8d11-9ae21f18978e-kube-api-access-6ztsf\") pod \"community-operators-td294\" (UID: \"9cd224da-05fd-4393-8d11-9ae21f18978e\") " pod="openshift-marketplace/community-operators-td294" Nov 21 15:23:39 crc kubenswrapper[4897]: I1121 15:23:39.076287 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-td294" Nov 21 15:23:39 crc kubenswrapper[4897]: W1121 15:23:39.642281 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9cd224da_05fd_4393_8d11_9ae21f18978e.slice/crio-9d71943a01f1119f01ebae54af638075d27a9db37c34a5a42569ecb7e9e4f982 WatchSource:0}: Error finding container 9d71943a01f1119f01ebae54af638075d27a9db37c34a5a42569ecb7e9e4f982: Status 404 returned error can't find the container with id 9d71943a01f1119f01ebae54af638075d27a9db37c34a5a42569ecb7e9e4f982 Nov 21 15:23:39 crc kubenswrapper[4897]: I1121 15:23:39.644204 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-td294"] Nov 21 15:23:40 crc kubenswrapper[4897]: I1121 15:23:40.023572 4897 generic.go:334] "Generic (PLEG): container finished" podID="9cd224da-05fd-4393-8d11-9ae21f18978e" containerID="671a583db8e7044f867ba2aabc1a8b7ae898a7dcbaf2e30fdfbc0401167acbc6" exitCode=0 Nov 21 15:23:40 crc kubenswrapper[4897]: I1121 15:23:40.023819 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-td294" event={"ID":"9cd224da-05fd-4393-8d11-9ae21f18978e","Type":"ContainerDied","Data":"671a583db8e7044f867ba2aabc1a8b7ae898a7dcbaf2e30fdfbc0401167acbc6"} Nov 21 15:23:40 crc kubenswrapper[4897]: I1121 15:23:40.024084 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-td294" event={"ID":"9cd224da-05fd-4393-8d11-9ae21f18978e","Type":"ContainerStarted","Data":"9d71943a01f1119f01ebae54af638075d27a9db37c34a5a42569ecb7e9e4f982"} Nov 21 15:23:42 crc kubenswrapper[4897]: I1121 15:23:42.045616 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-td294" event={"ID":"9cd224da-05fd-4393-8d11-9ae21f18978e","Type":"ContainerStarted","Data":"3b9a75cd650540323110f1d7109699e3faec20a36881fc945638b91a5207df79"} Nov 21 15:23:42 crc kubenswrapper[4897]: I1121 15:23:42.089200 4897 scope.go:117] "RemoveContainer" containerID="95e70a2cadfac38d1f3f3ba5e655bb58f1dc4e2d41e3d3c5a54f91a8584b921e" Nov 21 15:23:42 crc kubenswrapper[4897]: E1121 15:23:42.089482 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:23:43 crc kubenswrapper[4897]: I1121 15:23:43.058701 4897 generic.go:334] "Generic (PLEG): container finished" podID="9cd224da-05fd-4393-8d11-9ae21f18978e" containerID="3b9a75cd650540323110f1d7109699e3faec20a36881fc945638b91a5207df79" exitCode=0 Nov 21 15:23:43 crc kubenswrapper[4897]: I1121 15:23:43.058980 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-td294" event={"ID":"9cd224da-05fd-4393-8d11-9ae21f18978e","Type":"ContainerDied","Data":"3b9a75cd650540323110f1d7109699e3faec20a36881fc945638b91a5207df79"} Nov 21 15:23:46 crc kubenswrapper[4897]: I1121 15:23:46.111547 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-td294" event={"ID":"9cd224da-05fd-4393-8d11-9ae21f18978e","Type":"ContainerStarted","Data":"3ca1fc328dfe06a24eacab1f14957a81f200e8b052f8c966fc70fbef3c63a5b8"} Nov 21 15:23:46 crc kubenswrapper[4897]: I1121 15:23:46.143277 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-td294" podStartSLOduration=3.256164118 podStartE2EDuration="8.143257158s" podCreationTimestamp="2025-11-21 15:23:38 +0000 UTC" firstStartedPulling="2025-11-21 15:23:40.027440829 +0000 UTC m=+4497.312034314" lastFinishedPulling="2025-11-21 15:23:44.914533879 +0000 UTC m=+4502.199127354" observedRunningTime="2025-11-21 15:23:46.130281816 +0000 UTC m=+4503.414875291" watchObservedRunningTime="2025-11-21 15:23:46.143257158 +0000 UTC m=+4503.427850633" Nov 21 15:23:49 crc kubenswrapper[4897]: I1121 15:23:49.077270 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-td294" Nov 21 15:23:49 crc kubenswrapper[4897]: I1121 15:23:49.078749 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-td294" Nov 21 15:23:49 crc kubenswrapper[4897]: I1121 15:23:49.164396 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-td294" Nov 21 15:23:50 crc kubenswrapper[4897]: I1121 15:23:50.206032 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-td294" Nov 21 15:23:50 crc kubenswrapper[4897]: I1121 15:23:50.265383 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-td294"] Nov 21 15:23:52 crc kubenswrapper[4897]: I1121 15:23:52.178248 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-td294" podUID="9cd224da-05fd-4393-8d11-9ae21f18978e" containerName="registry-server" containerID="cri-o://3ca1fc328dfe06a24eacab1f14957a81f200e8b052f8c966fc70fbef3c63a5b8" gracePeriod=2 Nov 21 15:23:52 crc kubenswrapper[4897]: I1121 15:23:52.702365 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-td294" Nov 21 15:23:52 crc kubenswrapper[4897]: I1121 15:23:52.891446 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ztsf\" (UniqueName: \"kubernetes.io/projected/9cd224da-05fd-4393-8d11-9ae21f18978e-kube-api-access-6ztsf\") pod \"9cd224da-05fd-4393-8d11-9ae21f18978e\" (UID: \"9cd224da-05fd-4393-8d11-9ae21f18978e\") " Nov 21 15:23:52 crc kubenswrapper[4897]: I1121 15:23:52.891669 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9cd224da-05fd-4393-8d11-9ae21f18978e-utilities\") pod \"9cd224da-05fd-4393-8d11-9ae21f18978e\" (UID: \"9cd224da-05fd-4393-8d11-9ae21f18978e\") " Nov 21 15:23:52 crc kubenswrapper[4897]: I1121 15:23:52.891795 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9cd224da-05fd-4393-8d11-9ae21f18978e-catalog-content\") pod \"9cd224da-05fd-4393-8d11-9ae21f18978e\" (UID: \"9cd224da-05fd-4393-8d11-9ae21f18978e\") " Nov 21 15:23:52 crc kubenswrapper[4897]: I1121 15:23:52.892377 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9cd224da-05fd-4393-8d11-9ae21f18978e-utilities" (OuterVolumeSpecName: "utilities") pod "9cd224da-05fd-4393-8d11-9ae21f18978e" (UID: "9cd224da-05fd-4393-8d11-9ae21f18978e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 15:23:52 crc kubenswrapper[4897]: I1121 15:23:52.892762 4897 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9cd224da-05fd-4393-8d11-9ae21f18978e-utilities\") on node \"crc\" DevicePath \"\"" Nov 21 15:23:52 crc kubenswrapper[4897]: I1121 15:23:52.901785 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9cd224da-05fd-4393-8d11-9ae21f18978e-kube-api-access-6ztsf" (OuterVolumeSpecName: "kube-api-access-6ztsf") pod "9cd224da-05fd-4393-8d11-9ae21f18978e" (UID: "9cd224da-05fd-4393-8d11-9ae21f18978e"). InnerVolumeSpecName "kube-api-access-6ztsf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 15:23:52 crc kubenswrapper[4897]: I1121 15:23:52.995809 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ztsf\" (UniqueName: \"kubernetes.io/projected/9cd224da-05fd-4393-8d11-9ae21f18978e-kube-api-access-6ztsf\") on node \"crc\" DevicePath \"\"" Nov 21 15:23:53 crc kubenswrapper[4897]: I1121 15:23:53.097070 4897 scope.go:117] "RemoveContainer" containerID="95e70a2cadfac38d1f3f3ba5e655bb58f1dc4e2d41e3d3c5a54f91a8584b921e" Nov 21 15:23:53 crc kubenswrapper[4897]: E1121 15:23:53.098059 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:23:53 crc kubenswrapper[4897]: I1121 15:23:53.213088 4897 generic.go:334] "Generic (PLEG): container finished" podID="9cd224da-05fd-4393-8d11-9ae21f18978e" containerID="3ca1fc328dfe06a24eacab1f14957a81f200e8b052f8c966fc70fbef3c63a5b8" exitCode=0 Nov 21 15:23:53 crc kubenswrapper[4897]: I1121 15:23:53.213145 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-td294" event={"ID":"9cd224da-05fd-4393-8d11-9ae21f18978e","Type":"ContainerDied","Data":"3ca1fc328dfe06a24eacab1f14957a81f200e8b052f8c966fc70fbef3c63a5b8"} Nov 21 15:23:53 crc kubenswrapper[4897]: I1121 15:23:53.213170 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-td294" Nov 21 15:23:53 crc kubenswrapper[4897]: I1121 15:23:53.213190 4897 scope.go:117] "RemoveContainer" containerID="3ca1fc328dfe06a24eacab1f14957a81f200e8b052f8c966fc70fbef3c63a5b8" Nov 21 15:23:53 crc kubenswrapper[4897]: I1121 15:23:53.213178 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-td294" event={"ID":"9cd224da-05fd-4393-8d11-9ae21f18978e","Type":"ContainerDied","Data":"9d71943a01f1119f01ebae54af638075d27a9db37c34a5a42569ecb7e9e4f982"} Nov 21 15:23:53 crc kubenswrapper[4897]: I1121 15:23:53.245470 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9cd224da-05fd-4393-8d11-9ae21f18978e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9cd224da-05fd-4393-8d11-9ae21f18978e" (UID: "9cd224da-05fd-4393-8d11-9ae21f18978e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 15:23:53 crc kubenswrapper[4897]: I1121 15:23:53.261270 4897 scope.go:117] "RemoveContainer" containerID="3b9a75cd650540323110f1d7109699e3faec20a36881fc945638b91a5207df79" Nov 21 15:23:53 crc kubenswrapper[4897]: I1121 15:23:53.309629 4897 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9cd224da-05fd-4393-8d11-9ae21f18978e-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 21 15:23:53 crc kubenswrapper[4897]: I1121 15:23:53.554787 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-td294"] Nov 21 15:23:53 crc kubenswrapper[4897]: I1121 15:23:53.566932 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-td294"] Nov 21 15:23:53 crc kubenswrapper[4897]: I1121 15:23:53.853024 4897 scope.go:117] "RemoveContainer" containerID="671a583db8e7044f867ba2aabc1a8b7ae898a7dcbaf2e30fdfbc0401167acbc6" Nov 21 15:23:53 crc kubenswrapper[4897]: I1121 15:23:53.902490 4897 scope.go:117] "RemoveContainer" containerID="3ca1fc328dfe06a24eacab1f14957a81f200e8b052f8c966fc70fbef3c63a5b8" Nov 21 15:23:53 crc kubenswrapper[4897]: E1121 15:23:53.902947 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3ca1fc328dfe06a24eacab1f14957a81f200e8b052f8c966fc70fbef3c63a5b8\": container with ID starting with 3ca1fc328dfe06a24eacab1f14957a81f200e8b052f8c966fc70fbef3c63a5b8 not found: ID does not exist" containerID="3ca1fc328dfe06a24eacab1f14957a81f200e8b052f8c966fc70fbef3c63a5b8" Nov 21 15:23:53 crc kubenswrapper[4897]: I1121 15:23:53.902997 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3ca1fc328dfe06a24eacab1f14957a81f200e8b052f8c966fc70fbef3c63a5b8"} err="failed to get container status \"3ca1fc328dfe06a24eacab1f14957a81f200e8b052f8c966fc70fbef3c63a5b8\": rpc error: code = NotFound desc = could not find container \"3ca1fc328dfe06a24eacab1f14957a81f200e8b052f8c966fc70fbef3c63a5b8\": container with ID starting with 3ca1fc328dfe06a24eacab1f14957a81f200e8b052f8c966fc70fbef3c63a5b8 not found: ID does not exist" Nov 21 15:23:53 crc kubenswrapper[4897]: I1121 15:23:53.903027 4897 scope.go:117] "RemoveContainer" containerID="3b9a75cd650540323110f1d7109699e3faec20a36881fc945638b91a5207df79" Nov 21 15:23:53 crc kubenswrapper[4897]: E1121 15:23:53.903418 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3b9a75cd650540323110f1d7109699e3faec20a36881fc945638b91a5207df79\": container with ID starting with 3b9a75cd650540323110f1d7109699e3faec20a36881fc945638b91a5207df79 not found: ID does not exist" containerID="3b9a75cd650540323110f1d7109699e3faec20a36881fc945638b91a5207df79" Nov 21 15:23:53 crc kubenswrapper[4897]: I1121 15:23:53.903457 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b9a75cd650540323110f1d7109699e3faec20a36881fc945638b91a5207df79"} err="failed to get container status \"3b9a75cd650540323110f1d7109699e3faec20a36881fc945638b91a5207df79\": rpc error: code = NotFound desc = could not find container \"3b9a75cd650540323110f1d7109699e3faec20a36881fc945638b91a5207df79\": container with ID starting with 3b9a75cd650540323110f1d7109699e3faec20a36881fc945638b91a5207df79 not found: ID does not exist" Nov 21 15:23:53 crc kubenswrapper[4897]: I1121 15:23:53.903485 4897 scope.go:117] "RemoveContainer" containerID="671a583db8e7044f867ba2aabc1a8b7ae898a7dcbaf2e30fdfbc0401167acbc6" Nov 21 15:23:53 crc kubenswrapper[4897]: E1121 15:23:53.903770 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"671a583db8e7044f867ba2aabc1a8b7ae898a7dcbaf2e30fdfbc0401167acbc6\": container with ID starting with 671a583db8e7044f867ba2aabc1a8b7ae898a7dcbaf2e30fdfbc0401167acbc6 not found: ID does not exist" containerID="671a583db8e7044f867ba2aabc1a8b7ae898a7dcbaf2e30fdfbc0401167acbc6" Nov 21 15:23:53 crc kubenswrapper[4897]: I1121 15:23:53.903800 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"671a583db8e7044f867ba2aabc1a8b7ae898a7dcbaf2e30fdfbc0401167acbc6"} err="failed to get container status \"671a583db8e7044f867ba2aabc1a8b7ae898a7dcbaf2e30fdfbc0401167acbc6\": rpc error: code = NotFound desc = could not find container \"671a583db8e7044f867ba2aabc1a8b7ae898a7dcbaf2e30fdfbc0401167acbc6\": container with ID starting with 671a583db8e7044f867ba2aabc1a8b7ae898a7dcbaf2e30fdfbc0401167acbc6 not found: ID does not exist" Nov 21 15:23:54 crc kubenswrapper[4897]: I1121 15:23:54.104441 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9cd224da-05fd-4393-8d11-9ae21f18978e" path="/var/lib/kubelet/pods/9cd224da-05fd-4393-8d11-9ae21f18978e/volumes" Nov 21 15:24:05 crc kubenswrapper[4897]: I1121 15:24:05.089469 4897 scope.go:117] "RemoveContainer" containerID="95e70a2cadfac38d1f3f3ba5e655bb58f1dc4e2d41e3d3c5a54f91a8584b921e" Nov 21 15:24:05 crc kubenswrapper[4897]: E1121 15:24:05.090443 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:24:16 crc kubenswrapper[4897]: I1121 15:24:16.089294 4897 scope.go:117] "RemoveContainer" containerID="95e70a2cadfac38d1f3f3ba5e655bb58f1dc4e2d41e3d3c5a54f91a8584b921e" Nov 21 15:24:16 crc kubenswrapper[4897]: E1121 15:24:16.091631 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:24:31 crc kubenswrapper[4897]: I1121 15:24:31.089055 4897 scope.go:117] "RemoveContainer" containerID="95e70a2cadfac38d1f3f3ba5e655bb58f1dc4e2d41e3d3c5a54f91a8584b921e" Nov 21 15:24:31 crc kubenswrapper[4897]: E1121 15:24:31.091012 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:24:45 crc kubenswrapper[4897]: I1121 15:24:45.089849 4897 scope.go:117] "RemoveContainer" containerID="95e70a2cadfac38d1f3f3ba5e655bb58f1dc4e2d41e3d3c5a54f91a8584b921e" Nov 21 15:24:45 crc kubenswrapper[4897]: E1121 15:24:45.090836 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:24:58 crc kubenswrapper[4897]: I1121 15:24:58.089762 4897 scope.go:117] "RemoveContainer" containerID="95e70a2cadfac38d1f3f3ba5e655bb58f1dc4e2d41e3d3c5a54f91a8584b921e" Nov 21 15:24:58 crc kubenswrapper[4897]: E1121 15:24:58.090591 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:25:11 crc kubenswrapper[4897]: I1121 15:25:11.090050 4897 scope.go:117] "RemoveContainer" containerID="95e70a2cadfac38d1f3f3ba5e655bb58f1dc4e2d41e3d3c5a54f91a8584b921e" Nov 21 15:25:11 crc kubenswrapper[4897]: E1121 15:25:11.091774 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:25:22 crc kubenswrapper[4897]: I1121 15:25:22.090066 4897 scope.go:117] "RemoveContainer" containerID="95e70a2cadfac38d1f3f3ba5e655bb58f1dc4e2d41e3d3c5a54f91a8584b921e" Nov 21 15:25:22 crc kubenswrapper[4897]: E1121 15:25:22.091778 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:25:34 crc kubenswrapper[4897]: I1121 15:25:34.116100 4897 scope.go:117] "RemoveContainer" containerID="95e70a2cadfac38d1f3f3ba5e655bb58f1dc4e2d41e3d3c5a54f91a8584b921e" Nov 21 15:25:34 crc kubenswrapper[4897]: E1121 15:25:34.117492 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:25:46 crc kubenswrapper[4897]: I1121 15:25:46.089965 4897 scope.go:117] "RemoveContainer" containerID="95e70a2cadfac38d1f3f3ba5e655bb58f1dc4e2d41e3d3c5a54f91a8584b921e" Nov 21 15:25:46 crc kubenswrapper[4897]: E1121 15:25:46.090696 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:26:00 crc kubenswrapper[4897]: I1121 15:26:00.089888 4897 scope.go:117] "RemoveContainer" containerID="95e70a2cadfac38d1f3f3ba5e655bb58f1dc4e2d41e3d3c5a54f91a8584b921e" Nov 21 15:26:00 crc kubenswrapper[4897]: E1121 15:26:00.090581 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:26:11 crc kubenswrapper[4897]: I1121 15:26:11.089908 4897 scope.go:117] "RemoveContainer" containerID="95e70a2cadfac38d1f3f3ba5e655bb58f1dc4e2d41e3d3c5a54f91a8584b921e" Nov 21 15:26:11 crc kubenswrapper[4897]: E1121 15:26:11.090706 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:26:22 crc kubenswrapper[4897]: I1121 15:26:22.089283 4897 scope.go:117] "RemoveContainer" containerID="95e70a2cadfac38d1f3f3ba5e655bb58f1dc4e2d41e3d3c5a54f91a8584b921e" Nov 21 15:26:22 crc kubenswrapper[4897]: E1121 15:26:22.090088 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:26:34 crc kubenswrapper[4897]: I1121 15:26:34.097663 4897 scope.go:117] "RemoveContainer" containerID="95e70a2cadfac38d1f3f3ba5e655bb58f1dc4e2d41e3d3c5a54f91a8584b921e" Nov 21 15:26:34 crc kubenswrapper[4897]: E1121 15:26:34.098609 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:26:48 crc kubenswrapper[4897]: I1121 15:26:48.088959 4897 scope.go:117] "RemoveContainer" containerID="95e70a2cadfac38d1f3f3ba5e655bb58f1dc4e2d41e3d3c5a54f91a8584b921e" Nov 21 15:26:49 crc kubenswrapper[4897]: I1121 15:26:49.190559 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" event={"ID":"e7670227-d280-4847-b882-754429f56b0f","Type":"ContainerStarted","Data":"2a7c1a6f8fe489b93e0267f838232cae90d796b8213c4aa28e9ba6aeb840a22d"} Nov 21 15:29:04 crc kubenswrapper[4897]: I1121 15:29:04.370537 4897 patch_prober.go:28] interesting pod/machine-config-daemon-krv5b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 21 15:29:04 crc kubenswrapper[4897]: I1121 15:29:04.371032 4897 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 21 15:29:34 crc kubenswrapper[4897]: I1121 15:29:34.371327 4897 patch_prober.go:28] interesting pod/machine-config-daemon-krv5b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 21 15:29:34 crc kubenswrapper[4897]: I1121 15:29:34.371819 4897 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 21 15:29:49 crc kubenswrapper[4897]: I1121 15:29:49.827153 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-zdspd"] Nov 21 15:29:49 crc kubenswrapper[4897]: E1121 15:29:49.828491 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cd224da-05fd-4393-8d11-9ae21f18978e" containerName="registry-server" Nov 21 15:29:49 crc kubenswrapper[4897]: I1121 15:29:49.828518 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cd224da-05fd-4393-8d11-9ae21f18978e" containerName="registry-server" Nov 21 15:29:49 crc kubenswrapper[4897]: E1121 15:29:49.828568 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cd224da-05fd-4393-8d11-9ae21f18978e" containerName="extract-utilities" Nov 21 15:29:49 crc kubenswrapper[4897]: I1121 15:29:49.828575 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cd224da-05fd-4393-8d11-9ae21f18978e" containerName="extract-utilities" Nov 21 15:29:49 crc kubenswrapper[4897]: E1121 15:29:49.828605 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cd224da-05fd-4393-8d11-9ae21f18978e" containerName="extract-content" Nov 21 15:29:49 crc kubenswrapper[4897]: I1121 15:29:49.828613 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cd224da-05fd-4393-8d11-9ae21f18978e" containerName="extract-content" Nov 21 15:29:49 crc kubenswrapper[4897]: I1121 15:29:49.828898 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="9cd224da-05fd-4393-8d11-9ae21f18978e" containerName="registry-server" Nov 21 15:29:49 crc kubenswrapper[4897]: I1121 15:29:49.831389 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zdspd" Nov 21 15:29:49 crc kubenswrapper[4897]: I1121 15:29:49.843719 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zdspd"] Nov 21 15:29:49 crc kubenswrapper[4897]: I1121 15:29:49.964577 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5c7kn\" (UniqueName: \"kubernetes.io/projected/8f40194d-4a22-45a2-8892-d66931e738c5-kube-api-access-5c7kn\") pod \"redhat-operators-zdspd\" (UID: \"8f40194d-4a22-45a2-8892-d66931e738c5\") " pod="openshift-marketplace/redhat-operators-zdspd" Nov 21 15:29:49 crc kubenswrapper[4897]: I1121 15:29:49.964752 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f40194d-4a22-45a2-8892-d66931e738c5-utilities\") pod \"redhat-operators-zdspd\" (UID: \"8f40194d-4a22-45a2-8892-d66931e738c5\") " pod="openshift-marketplace/redhat-operators-zdspd" Nov 21 15:29:49 crc kubenswrapper[4897]: I1121 15:29:49.964803 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f40194d-4a22-45a2-8892-d66931e738c5-catalog-content\") pod \"redhat-operators-zdspd\" (UID: \"8f40194d-4a22-45a2-8892-d66931e738c5\") " pod="openshift-marketplace/redhat-operators-zdspd" Nov 21 15:29:50 crc kubenswrapper[4897]: I1121 15:29:50.066188 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f40194d-4a22-45a2-8892-d66931e738c5-catalog-content\") pod \"redhat-operators-zdspd\" (UID: \"8f40194d-4a22-45a2-8892-d66931e738c5\") " pod="openshift-marketplace/redhat-operators-zdspd" Nov 21 15:29:50 crc kubenswrapper[4897]: I1121 15:29:50.066752 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f40194d-4a22-45a2-8892-d66931e738c5-catalog-content\") pod \"redhat-operators-zdspd\" (UID: \"8f40194d-4a22-45a2-8892-d66931e738c5\") " pod="openshift-marketplace/redhat-operators-zdspd" Nov 21 15:29:50 crc kubenswrapper[4897]: I1121 15:29:50.067013 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5c7kn\" (UniqueName: \"kubernetes.io/projected/8f40194d-4a22-45a2-8892-d66931e738c5-kube-api-access-5c7kn\") pod \"redhat-operators-zdspd\" (UID: \"8f40194d-4a22-45a2-8892-d66931e738c5\") " pod="openshift-marketplace/redhat-operators-zdspd" Nov 21 15:29:50 crc kubenswrapper[4897]: I1121 15:29:50.067177 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f40194d-4a22-45a2-8892-d66931e738c5-utilities\") pod \"redhat-operators-zdspd\" (UID: \"8f40194d-4a22-45a2-8892-d66931e738c5\") " pod="openshift-marketplace/redhat-operators-zdspd" Nov 21 15:29:50 crc kubenswrapper[4897]: I1121 15:29:50.067543 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f40194d-4a22-45a2-8892-d66931e738c5-utilities\") pod \"redhat-operators-zdspd\" (UID: \"8f40194d-4a22-45a2-8892-d66931e738c5\") " pod="openshift-marketplace/redhat-operators-zdspd" Nov 21 15:29:50 crc kubenswrapper[4897]: I1121 15:29:50.088965 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5c7kn\" (UniqueName: \"kubernetes.io/projected/8f40194d-4a22-45a2-8892-d66931e738c5-kube-api-access-5c7kn\") pod \"redhat-operators-zdspd\" (UID: \"8f40194d-4a22-45a2-8892-d66931e738c5\") " pod="openshift-marketplace/redhat-operators-zdspd" Nov 21 15:29:50 crc kubenswrapper[4897]: I1121 15:29:50.171663 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zdspd" Nov 21 15:29:50 crc kubenswrapper[4897]: I1121 15:29:50.737406 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zdspd"] Nov 21 15:29:51 crc kubenswrapper[4897]: I1121 15:29:51.402370 4897 generic.go:334] "Generic (PLEG): container finished" podID="8f40194d-4a22-45a2-8892-d66931e738c5" containerID="9f42ad5bb82bb271bbb11b79ecf595b05c81deef1e6a991a86b36dc7782b67d5" exitCode=0 Nov 21 15:29:51 crc kubenswrapper[4897]: I1121 15:29:51.402702 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zdspd" event={"ID":"8f40194d-4a22-45a2-8892-d66931e738c5","Type":"ContainerDied","Data":"9f42ad5bb82bb271bbb11b79ecf595b05c81deef1e6a991a86b36dc7782b67d5"} Nov 21 15:29:51 crc kubenswrapper[4897]: I1121 15:29:51.402731 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zdspd" event={"ID":"8f40194d-4a22-45a2-8892-d66931e738c5","Type":"ContainerStarted","Data":"e163e2ff7dc58b335051a7c2cf6b96d0a60c64f065b1c084ff7f9b3ee214c6f4"} Nov 21 15:29:51 crc kubenswrapper[4897]: I1121 15:29:51.405830 4897 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 21 15:29:52 crc kubenswrapper[4897]: I1121 15:29:52.416944 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zdspd" event={"ID":"8f40194d-4a22-45a2-8892-d66931e738c5","Type":"ContainerStarted","Data":"61d1447cb9750806e7956941105921d5f366a6ebbc6c96540330f603fca858fe"} Nov 21 15:29:58 crc kubenswrapper[4897]: I1121 15:29:58.482955 4897 generic.go:334] "Generic (PLEG): container finished" podID="8f40194d-4a22-45a2-8892-d66931e738c5" containerID="61d1447cb9750806e7956941105921d5f366a6ebbc6c96540330f603fca858fe" exitCode=0 Nov 21 15:29:58 crc kubenswrapper[4897]: I1121 15:29:58.483072 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zdspd" event={"ID":"8f40194d-4a22-45a2-8892-d66931e738c5","Type":"ContainerDied","Data":"61d1447cb9750806e7956941105921d5f366a6ebbc6c96540330f603fca858fe"} Nov 21 15:30:00 crc kubenswrapper[4897]: I1121 15:30:00.150134 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29395650-rcqwk"] Nov 21 15:30:00 crc kubenswrapper[4897]: I1121 15:30:00.157699 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29395650-rcqwk" Nov 21 15:30:00 crc kubenswrapper[4897]: I1121 15:30:00.161722 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 21 15:30:00 crc kubenswrapper[4897]: I1121 15:30:00.166076 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 21 15:30:00 crc kubenswrapper[4897]: I1121 15:30:00.179637 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29395650-rcqwk"] Nov 21 15:30:00 crc kubenswrapper[4897]: I1121 15:30:00.335682 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gsg4d\" (UniqueName: \"kubernetes.io/projected/4fa61b28-cda9-47e7-8fb4-3b8cd8cfcc11-kube-api-access-gsg4d\") pod \"collect-profiles-29395650-rcqwk\" (UID: \"4fa61b28-cda9-47e7-8fb4-3b8cd8cfcc11\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29395650-rcqwk" Nov 21 15:30:00 crc kubenswrapper[4897]: I1121 15:30:00.335823 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4fa61b28-cda9-47e7-8fb4-3b8cd8cfcc11-secret-volume\") pod \"collect-profiles-29395650-rcqwk\" (UID: \"4fa61b28-cda9-47e7-8fb4-3b8cd8cfcc11\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29395650-rcqwk" Nov 21 15:30:00 crc kubenswrapper[4897]: I1121 15:30:00.336312 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4fa61b28-cda9-47e7-8fb4-3b8cd8cfcc11-config-volume\") pod \"collect-profiles-29395650-rcqwk\" (UID: \"4fa61b28-cda9-47e7-8fb4-3b8cd8cfcc11\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29395650-rcqwk" Nov 21 15:30:00 crc kubenswrapper[4897]: I1121 15:30:00.438314 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gsg4d\" (UniqueName: \"kubernetes.io/projected/4fa61b28-cda9-47e7-8fb4-3b8cd8cfcc11-kube-api-access-gsg4d\") pod \"collect-profiles-29395650-rcqwk\" (UID: \"4fa61b28-cda9-47e7-8fb4-3b8cd8cfcc11\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29395650-rcqwk" Nov 21 15:30:00 crc kubenswrapper[4897]: I1121 15:30:00.438687 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4fa61b28-cda9-47e7-8fb4-3b8cd8cfcc11-secret-volume\") pod \"collect-profiles-29395650-rcqwk\" (UID: \"4fa61b28-cda9-47e7-8fb4-3b8cd8cfcc11\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29395650-rcqwk" Nov 21 15:30:00 crc kubenswrapper[4897]: I1121 15:30:00.438932 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4fa61b28-cda9-47e7-8fb4-3b8cd8cfcc11-config-volume\") pod \"collect-profiles-29395650-rcqwk\" (UID: \"4fa61b28-cda9-47e7-8fb4-3b8cd8cfcc11\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29395650-rcqwk" Nov 21 15:30:00 crc kubenswrapper[4897]: I1121 15:30:00.439884 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4fa61b28-cda9-47e7-8fb4-3b8cd8cfcc11-config-volume\") pod \"collect-profiles-29395650-rcqwk\" (UID: \"4fa61b28-cda9-47e7-8fb4-3b8cd8cfcc11\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29395650-rcqwk" Nov 21 15:30:00 crc kubenswrapper[4897]: I1121 15:30:00.446975 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4fa61b28-cda9-47e7-8fb4-3b8cd8cfcc11-secret-volume\") pod \"collect-profiles-29395650-rcqwk\" (UID: \"4fa61b28-cda9-47e7-8fb4-3b8cd8cfcc11\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29395650-rcqwk" Nov 21 15:30:00 crc kubenswrapper[4897]: I1121 15:30:00.460994 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gsg4d\" (UniqueName: \"kubernetes.io/projected/4fa61b28-cda9-47e7-8fb4-3b8cd8cfcc11-kube-api-access-gsg4d\") pod \"collect-profiles-29395650-rcqwk\" (UID: \"4fa61b28-cda9-47e7-8fb4-3b8cd8cfcc11\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29395650-rcqwk" Nov 21 15:30:00 crc kubenswrapper[4897]: I1121 15:30:00.482642 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29395650-rcqwk" Nov 21 15:30:00 crc kubenswrapper[4897]: I1121 15:30:00.506732 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zdspd" event={"ID":"8f40194d-4a22-45a2-8892-d66931e738c5","Type":"ContainerStarted","Data":"478b9ee4cfff944b273e3eea0c75cf9302ae644e035bc8d7d2e4b7e2987ba6ae"} Nov 21 15:30:00 crc kubenswrapper[4897]: I1121 15:30:00.530271 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-zdspd" podStartSLOduration=3.850475872 podStartE2EDuration="11.530245047s" podCreationTimestamp="2025-11-21 15:29:49 +0000 UTC" firstStartedPulling="2025-11-21 15:29:51.405475 +0000 UTC m=+4868.690068475" lastFinishedPulling="2025-11-21 15:29:59.085244175 +0000 UTC m=+4876.369837650" observedRunningTime="2025-11-21 15:30:00.526101474 +0000 UTC m=+4877.810694949" watchObservedRunningTime="2025-11-21 15:30:00.530245047 +0000 UTC m=+4877.814838512" Nov 21 15:30:01 crc kubenswrapper[4897]: I1121 15:30:01.002724 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29395650-rcqwk"] Nov 21 15:30:01 crc kubenswrapper[4897]: I1121 15:30:01.521389 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29395650-rcqwk" event={"ID":"4fa61b28-cda9-47e7-8fb4-3b8cd8cfcc11","Type":"ContainerStarted","Data":"3239c255fa70ede567b6711a2c773ccfa1e6aced8c460e2d3933483a5af9176a"} Nov 21 15:30:01 crc kubenswrapper[4897]: I1121 15:30:01.521443 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29395650-rcqwk" event={"ID":"4fa61b28-cda9-47e7-8fb4-3b8cd8cfcc11","Type":"ContainerStarted","Data":"f1d656a7bc99771a5afa7dfbd42fca16e43225c298159b4bbdcc32ee9399f416"} Nov 21 15:30:01 crc kubenswrapper[4897]: I1121 15:30:01.544427 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29395650-rcqwk" podStartSLOduration=1.544408759 podStartE2EDuration="1.544408759s" podCreationTimestamp="2025-11-21 15:30:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 15:30:01.53412345 +0000 UTC m=+4878.818716935" watchObservedRunningTime="2025-11-21 15:30:01.544408759 +0000 UTC m=+4878.829002234" Nov 21 15:30:02 crc kubenswrapper[4897]: I1121 15:30:02.542129 4897 generic.go:334] "Generic (PLEG): container finished" podID="4fa61b28-cda9-47e7-8fb4-3b8cd8cfcc11" containerID="3239c255fa70ede567b6711a2c773ccfa1e6aced8c460e2d3933483a5af9176a" exitCode=0 Nov 21 15:30:02 crc kubenswrapper[4897]: I1121 15:30:02.542202 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29395650-rcqwk" event={"ID":"4fa61b28-cda9-47e7-8fb4-3b8cd8cfcc11","Type":"ContainerDied","Data":"3239c255fa70ede567b6711a2c773ccfa1e6aced8c460e2d3933483a5af9176a"} Nov 21 15:30:03 crc kubenswrapper[4897]: I1121 15:30:03.959606 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29395650-rcqwk" Nov 21 15:30:04 crc kubenswrapper[4897]: I1121 15:30:04.132062 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4fa61b28-cda9-47e7-8fb4-3b8cd8cfcc11-config-volume\") pod \"4fa61b28-cda9-47e7-8fb4-3b8cd8cfcc11\" (UID: \"4fa61b28-cda9-47e7-8fb4-3b8cd8cfcc11\") " Nov 21 15:30:04 crc kubenswrapper[4897]: I1121 15:30:04.132115 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4fa61b28-cda9-47e7-8fb4-3b8cd8cfcc11-secret-volume\") pod \"4fa61b28-cda9-47e7-8fb4-3b8cd8cfcc11\" (UID: \"4fa61b28-cda9-47e7-8fb4-3b8cd8cfcc11\") " Nov 21 15:30:04 crc kubenswrapper[4897]: I1121 15:30:04.132166 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gsg4d\" (UniqueName: \"kubernetes.io/projected/4fa61b28-cda9-47e7-8fb4-3b8cd8cfcc11-kube-api-access-gsg4d\") pod \"4fa61b28-cda9-47e7-8fb4-3b8cd8cfcc11\" (UID: \"4fa61b28-cda9-47e7-8fb4-3b8cd8cfcc11\") " Nov 21 15:30:04 crc kubenswrapper[4897]: I1121 15:30:04.132857 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4fa61b28-cda9-47e7-8fb4-3b8cd8cfcc11-config-volume" (OuterVolumeSpecName: "config-volume") pod "4fa61b28-cda9-47e7-8fb4-3b8cd8cfcc11" (UID: "4fa61b28-cda9-47e7-8fb4-3b8cd8cfcc11"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 15:30:04 crc kubenswrapper[4897]: I1121 15:30:04.143014 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4fa61b28-cda9-47e7-8fb4-3b8cd8cfcc11-kube-api-access-gsg4d" (OuterVolumeSpecName: "kube-api-access-gsg4d") pod "4fa61b28-cda9-47e7-8fb4-3b8cd8cfcc11" (UID: "4fa61b28-cda9-47e7-8fb4-3b8cd8cfcc11"). InnerVolumeSpecName "kube-api-access-gsg4d". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 15:30:04 crc kubenswrapper[4897]: I1121 15:30:04.144379 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4fa61b28-cda9-47e7-8fb4-3b8cd8cfcc11-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "4fa61b28-cda9-47e7-8fb4-3b8cd8cfcc11" (UID: "4fa61b28-cda9-47e7-8fb4-3b8cd8cfcc11"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 15:30:04 crc kubenswrapper[4897]: I1121 15:30:04.238615 4897 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4fa61b28-cda9-47e7-8fb4-3b8cd8cfcc11-config-volume\") on node \"crc\" DevicePath \"\"" Nov 21 15:30:04 crc kubenswrapper[4897]: I1121 15:30:04.239030 4897 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4fa61b28-cda9-47e7-8fb4-3b8cd8cfcc11-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 21 15:30:04 crc kubenswrapper[4897]: I1121 15:30:04.239040 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gsg4d\" (UniqueName: \"kubernetes.io/projected/4fa61b28-cda9-47e7-8fb4-3b8cd8cfcc11-kube-api-access-gsg4d\") on node \"crc\" DevicePath \"\"" Nov 21 15:30:04 crc kubenswrapper[4897]: I1121 15:30:04.370640 4897 patch_prober.go:28] interesting pod/machine-config-daemon-krv5b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 21 15:30:04 crc kubenswrapper[4897]: I1121 15:30:04.370700 4897 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 21 15:30:04 crc kubenswrapper[4897]: I1121 15:30:04.370748 4897 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" Nov 21 15:30:04 crc kubenswrapper[4897]: I1121 15:30:04.371643 4897 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2a7c1a6f8fe489b93e0267f838232cae90d796b8213c4aa28e9ba6aeb840a22d"} pod="openshift-machine-config-operator/machine-config-daemon-krv5b" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 21 15:30:04 crc kubenswrapper[4897]: I1121 15:30:04.371695 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" containerName="machine-config-daemon" containerID="cri-o://2a7c1a6f8fe489b93e0267f838232cae90d796b8213c4aa28e9ba6aeb840a22d" gracePeriod=600 Nov 21 15:30:04 crc kubenswrapper[4897]: I1121 15:30:04.570276 4897 generic.go:334] "Generic (PLEG): container finished" podID="e7670227-d280-4847-b882-754429f56b0f" containerID="2a7c1a6f8fe489b93e0267f838232cae90d796b8213c4aa28e9ba6aeb840a22d" exitCode=0 Nov 21 15:30:04 crc kubenswrapper[4897]: I1121 15:30:04.570328 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" event={"ID":"e7670227-d280-4847-b882-754429f56b0f","Type":"ContainerDied","Data":"2a7c1a6f8fe489b93e0267f838232cae90d796b8213c4aa28e9ba6aeb840a22d"} Nov 21 15:30:04 crc kubenswrapper[4897]: I1121 15:30:04.570395 4897 scope.go:117] "RemoveContainer" containerID="95e70a2cadfac38d1f3f3ba5e655bb58f1dc4e2d41e3d3c5a54f91a8584b921e" Nov 21 15:30:04 crc kubenswrapper[4897]: I1121 15:30:04.575692 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29395650-rcqwk" event={"ID":"4fa61b28-cda9-47e7-8fb4-3b8cd8cfcc11","Type":"ContainerDied","Data":"f1d656a7bc99771a5afa7dfbd42fca16e43225c298159b4bbdcc32ee9399f416"} Nov 21 15:30:04 crc kubenswrapper[4897]: I1121 15:30:04.575738 4897 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f1d656a7bc99771a5afa7dfbd42fca16e43225c298159b4bbdcc32ee9399f416" Nov 21 15:30:04 crc kubenswrapper[4897]: I1121 15:30:04.575804 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29395650-rcqwk" Nov 21 15:30:04 crc kubenswrapper[4897]: I1121 15:30:04.618860 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29395605-gqtnp"] Nov 21 15:30:04 crc kubenswrapper[4897]: I1121 15:30:04.629085 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29395605-gqtnp"] Nov 21 15:30:05 crc kubenswrapper[4897]: I1121 15:30:05.586654 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" event={"ID":"e7670227-d280-4847-b882-754429f56b0f","Type":"ContainerStarted","Data":"abf30c46658df7e750da25f3feeb4a134367a6c61c5b5221d6703974a1dd4d2e"} Nov 21 15:30:06 crc kubenswrapper[4897]: I1121 15:30:06.110502 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="52dcab2f-b253-4bb2-a479-518edcb7801a" path="/var/lib/kubelet/pods/52dcab2f-b253-4bb2-a479-518edcb7801a/volumes" Nov 21 15:30:10 crc kubenswrapper[4897]: I1121 15:30:10.171944 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-zdspd" Nov 21 15:30:10 crc kubenswrapper[4897]: I1121 15:30:10.172588 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-zdspd" Nov 21 15:30:11 crc kubenswrapper[4897]: I1121 15:30:11.222394 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-zdspd" podUID="8f40194d-4a22-45a2-8892-d66931e738c5" containerName="registry-server" probeResult="failure" output=< Nov 21 15:30:11 crc kubenswrapper[4897]: timeout: failed to connect service ":50051" within 1s Nov 21 15:30:11 crc kubenswrapper[4897]: > Nov 21 15:30:16 crc kubenswrapper[4897]: I1121 15:30:16.555447 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-frgrd"] Nov 21 15:30:16 crc kubenswrapper[4897]: E1121 15:30:16.558534 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4fa61b28-cda9-47e7-8fb4-3b8cd8cfcc11" containerName="collect-profiles" Nov 21 15:30:16 crc kubenswrapper[4897]: I1121 15:30:16.558554 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="4fa61b28-cda9-47e7-8fb4-3b8cd8cfcc11" containerName="collect-profiles" Nov 21 15:30:16 crc kubenswrapper[4897]: I1121 15:30:16.559407 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="4fa61b28-cda9-47e7-8fb4-3b8cd8cfcc11" containerName="collect-profiles" Nov 21 15:30:16 crc kubenswrapper[4897]: I1121 15:30:16.562802 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-frgrd" Nov 21 15:30:16 crc kubenswrapper[4897]: I1121 15:30:16.572714 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-frgrd"] Nov 21 15:30:16 crc kubenswrapper[4897]: I1121 15:30:16.639292 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-92chg\" (UniqueName: \"kubernetes.io/projected/d360d107-e163-45c9-8fa6-d5f47c78a9cc-kube-api-access-92chg\") pod \"certified-operators-frgrd\" (UID: \"d360d107-e163-45c9-8fa6-d5f47c78a9cc\") " pod="openshift-marketplace/certified-operators-frgrd" Nov 21 15:30:16 crc kubenswrapper[4897]: I1121 15:30:16.639425 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d360d107-e163-45c9-8fa6-d5f47c78a9cc-utilities\") pod \"certified-operators-frgrd\" (UID: \"d360d107-e163-45c9-8fa6-d5f47c78a9cc\") " pod="openshift-marketplace/certified-operators-frgrd" Nov 21 15:30:16 crc kubenswrapper[4897]: I1121 15:30:16.639473 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d360d107-e163-45c9-8fa6-d5f47c78a9cc-catalog-content\") pod \"certified-operators-frgrd\" (UID: \"d360d107-e163-45c9-8fa6-d5f47c78a9cc\") " pod="openshift-marketplace/certified-operators-frgrd" Nov 21 15:30:16 crc kubenswrapper[4897]: I1121 15:30:16.741784 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-92chg\" (UniqueName: \"kubernetes.io/projected/d360d107-e163-45c9-8fa6-d5f47c78a9cc-kube-api-access-92chg\") pod \"certified-operators-frgrd\" (UID: \"d360d107-e163-45c9-8fa6-d5f47c78a9cc\") " pod="openshift-marketplace/certified-operators-frgrd" Nov 21 15:30:16 crc kubenswrapper[4897]: I1121 15:30:16.741891 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d360d107-e163-45c9-8fa6-d5f47c78a9cc-utilities\") pod \"certified-operators-frgrd\" (UID: \"d360d107-e163-45c9-8fa6-d5f47c78a9cc\") " pod="openshift-marketplace/certified-operators-frgrd" Nov 21 15:30:16 crc kubenswrapper[4897]: I1121 15:30:16.741934 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d360d107-e163-45c9-8fa6-d5f47c78a9cc-catalog-content\") pod \"certified-operators-frgrd\" (UID: \"d360d107-e163-45c9-8fa6-d5f47c78a9cc\") " pod="openshift-marketplace/certified-operators-frgrd" Nov 21 15:30:16 crc kubenswrapper[4897]: I1121 15:30:16.742868 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d360d107-e163-45c9-8fa6-d5f47c78a9cc-catalog-content\") pod \"certified-operators-frgrd\" (UID: \"d360d107-e163-45c9-8fa6-d5f47c78a9cc\") " pod="openshift-marketplace/certified-operators-frgrd" Nov 21 15:30:16 crc kubenswrapper[4897]: I1121 15:30:16.742982 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d360d107-e163-45c9-8fa6-d5f47c78a9cc-utilities\") pod \"certified-operators-frgrd\" (UID: \"d360d107-e163-45c9-8fa6-d5f47c78a9cc\") " pod="openshift-marketplace/certified-operators-frgrd" Nov 21 15:30:16 crc kubenswrapper[4897]: I1121 15:30:16.765576 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-92chg\" (UniqueName: \"kubernetes.io/projected/d360d107-e163-45c9-8fa6-d5f47c78a9cc-kube-api-access-92chg\") pod \"certified-operators-frgrd\" (UID: \"d360d107-e163-45c9-8fa6-d5f47c78a9cc\") " pod="openshift-marketplace/certified-operators-frgrd" Nov 21 15:30:16 crc kubenswrapper[4897]: I1121 15:30:16.930704 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-frgrd" Nov 21 15:30:17 crc kubenswrapper[4897]: W1121 15:30:17.489442 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd360d107_e163_45c9_8fa6_d5f47c78a9cc.slice/crio-6ad11fb2779ce57171120eb8a2064304a553b0b2751cbb34bb1d7b926728cb39 WatchSource:0}: Error finding container 6ad11fb2779ce57171120eb8a2064304a553b0b2751cbb34bb1d7b926728cb39: Status 404 returned error can't find the container with id 6ad11fb2779ce57171120eb8a2064304a553b0b2751cbb34bb1d7b926728cb39 Nov 21 15:30:17 crc kubenswrapper[4897]: I1121 15:30:17.491068 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-frgrd"] Nov 21 15:30:17 crc kubenswrapper[4897]: I1121 15:30:17.724807 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-frgrd" event={"ID":"d360d107-e163-45c9-8fa6-d5f47c78a9cc","Type":"ContainerStarted","Data":"6ad11fb2779ce57171120eb8a2064304a553b0b2751cbb34bb1d7b926728cb39"} Nov 21 15:30:18 crc kubenswrapper[4897]: I1121 15:30:18.737825 4897 generic.go:334] "Generic (PLEG): container finished" podID="d360d107-e163-45c9-8fa6-d5f47c78a9cc" containerID="0d1a541ac7151fc5c88b28da7b0f36d9efb37de9100a43f8719fbfe359263a66" exitCode=0 Nov 21 15:30:18 crc kubenswrapper[4897]: I1121 15:30:18.737927 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-frgrd" event={"ID":"d360d107-e163-45c9-8fa6-d5f47c78a9cc","Type":"ContainerDied","Data":"0d1a541ac7151fc5c88b28da7b0f36d9efb37de9100a43f8719fbfe359263a66"} Nov 21 15:30:18 crc kubenswrapper[4897]: I1121 15:30:18.946892 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-bnzxb"] Nov 21 15:30:18 crc kubenswrapper[4897]: I1121 15:30:18.950000 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bnzxb" Nov 21 15:30:18 crc kubenswrapper[4897]: I1121 15:30:18.962458 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bnzxb"] Nov 21 15:30:19 crc kubenswrapper[4897]: I1121 15:30:19.118955 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/60077aeb-0865-4eee-b44e-60be41cf4326-utilities\") pod \"redhat-marketplace-bnzxb\" (UID: \"60077aeb-0865-4eee-b44e-60be41cf4326\") " pod="openshift-marketplace/redhat-marketplace-bnzxb" Nov 21 15:30:19 crc kubenswrapper[4897]: I1121 15:30:19.119412 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ktcn8\" (UniqueName: \"kubernetes.io/projected/60077aeb-0865-4eee-b44e-60be41cf4326-kube-api-access-ktcn8\") pod \"redhat-marketplace-bnzxb\" (UID: \"60077aeb-0865-4eee-b44e-60be41cf4326\") " pod="openshift-marketplace/redhat-marketplace-bnzxb" Nov 21 15:30:19 crc kubenswrapper[4897]: I1121 15:30:19.119690 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/60077aeb-0865-4eee-b44e-60be41cf4326-catalog-content\") pod \"redhat-marketplace-bnzxb\" (UID: \"60077aeb-0865-4eee-b44e-60be41cf4326\") " pod="openshift-marketplace/redhat-marketplace-bnzxb" Nov 21 15:30:19 crc kubenswrapper[4897]: I1121 15:30:19.222970 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/60077aeb-0865-4eee-b44e-60be41cf4326-catalog-content\") pod \"redhat-marketplace-bnzxb\" (UID: \"60077aeb-0865-4eee-b44e-60be41cf4326\") " pod="openshift-marketplace/redhat-marketplace-bnzxb" Nov 21 15:30:19 crc kubenswrapper[4897]: I1121 15:30:19.223250 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/60077aeb-0865-4eee-b44e-60be41cf4326-utilities\") pod \"redhat-marketplace-bnzxb\" (UID: \"60077aeb-0865-4eee-b44e-60be41cf4326\") " pod="openshift-marketplace/redhat-marketplace-bnzxb" Nov 21 15:30:19 crc kubenswrapper[4897]: I1121 15:30:19.223562 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ktcn8\" (UniqueName: \"kubernetes.io/projected/60077aeb-0865-4eee-b44e-60be41cf4326-kube-api-access-ktcn8\") pod \"redhat-marketplace-bnzxb\" (UID: \"60077aeb-0865-4eee-b44e-60be41cf4326\") " pod="openshift-marketplace/redhat-marketplace-bnzxb" Nov 21 15:30:19 crc kubenswrapper[4897]: I1121 15:30:19.223845 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/60077aeb-0865-4eee-b44e-60be41cf4326-utilities\") pod \"redhat-marketplace-bnzxb\" (UID: \"60077aeb-0865-4eee-b44e-60be41cf4326\") " pod="openshift-marketplace/redhat-marketplace-bnzxb" Nov 21 15:30:19 crc kubenswrapper[4897]: I1121 15:30:19.224837 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/60077aeb-0865-4eee-b44e-60be41cf4326-catalog-content\") pod \"redhat-marketplace-bnzxb\" (UID: \"60077aeb-0865-4eee-b44e-60be41cf4326\") " pod="openshift-marketplace/redhat-marketplace-bnzxb" Nov 21 15:30:19 crc kubenswrapper[4897]: I1121 15:30:19.248485 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ktcn8\" (UniqueName: \"kubernetes.io/projected/60077aeb-0865-4eee-b44e-60be41cf4326-kube-api-access-ktcn8\") pod \"redhat-marketplace-bnzxb\" (UID: \"60077aeb-0865-4eee-b44e-60be41cf4326\") " pod="openshift-marketplace/redhat-marketplace-bnzxb" Nov 21 15:30:19 crc kubenswrapper[4897]: I1121 15:30:19.286484 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bnzxb" Nov 21 15:30:19 crc kubenswrapper[4897]: I1121 15:30:19.850261 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bnzxb"] Nov 21 15:30:20 crc kubenswrapper[4897]: I1121 15:30:20.763916 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-frgrd" event={"ID":"d360d107-e163-45c9-8fa6-d5f47c78a9cc","Type":"ContainerStarted","Data":"2986347fe6fceca60228697b3b4f7cf09c6ccee53de91a6b0e75d1a75313b43b"} Nov 21 15:30:20 crc kubenswrapper[4897]: I1121 15:30:20.767097 4897 generic.go:334] "Generic (PLEG): container finished" podID="60077aeb-0865-4eee-b44e-60be41cf4326" containerID="75ccd25815f0b39c2b26a41228c4e231b7c3345ed2d38c4157327c4e0eb1f0f2" exitCode=0 Nov 21 15:30:20 crc kubenswrapper[4897]: I1121 15:30:20.767134 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bnzxb" event={"ID":"60077aeb-0865-4eee-b44e-60be41cf4326","Type":"ContainerDied","Data":"75ccd25815f0b39c2b26a41228c4e231b7c3345ed2d38c4157327c4e0eb1f0f2"} Nov 21 15:30:20 crc kubenswrapper[4897]: I1121 15:30:20.767160 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bnzxb" event={"ID":"60077aeb-0865-4eee-b44e-60be41cf4326","Type":"ContainerStarted","Data":"0ff91d0b041ab50bf32d36a9ea1f384479355021186f5676bb619fc69cb28a87"} Nov 21 15:30:21 crc kubenswrapper[4897]: I1121 15:30:21.220762 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-zdspd" podUID="8f40194d-4a22-45a2-8892-d66931e738c5" containerName="registry-server" probeResult="failure" output=< Nov 21 15:30:21 crc kubenswrapper[4897]: timeout: failed to connect service ":50051" within 1s Nov 21 15:30:21 crc kubenswrapper[4897]: > Nov 21 15:30:22 crc kubenswrapper[4897]: I1121 15:30:22.790267 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bnzxb" event={"ID":"60077aeb-0865-4eee-b44e-60be41cf4326","Type":"ContainerStarted","Data":"0d91752aeaed8f1597a4e93218c0e815e67f9d1f3e450156fd9a2a5fdc3079f9"} Nov 21 15:30:22 crc kubenswrapper[4897]: I1121 15:30:22.792782 4897 generic.go:334] "Generic (PLEG): container finished" podID="d360d107-e163-45c9-8fa6-d5f47c78a9cc" containerID="2986347fe6fceca60228697b3b4f7cf09c6ccee53de91a6b0e75d1a75313b43b" exitCode=0 Nov 21 15:30:22 crc kubenswrapper[4897]: I1121 15:30:22.792830 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-frgrd" event={"ID":"d360d107-e163-45c9-8fa6-d5f47c78a9cc","Type":"ContainerDied","Data":"2986347fe6fceca60228697b3b4f7cf09c6ccee53de91a6b0e75d1a75313b43b"} Nov 21 15:30:23 crc kubenswrapper[4897]: I1121 15:30:23.340441 4897 scope.go:117] "RemoveContainer" containerID="1098b67859440bd4a11d413ae31321bc4824e29a052fc8467fe213efa49a6d66" Nov 21 15:30:23 crc kubenswrapper[4897]: I1121 15:30:23.812535 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-frgrd" event={"ID":"d360d107-e163-45c9-8fa6-d5f47c78a9cc","Type":"ContainerStarted","Data":"e48d9667959f644ef2ed5f8fca5c3a8a67489ac32bfe162586f8b39633975614"} Nov 21 15:30:23 crc kubenswrapper[4897]: I1121 15:30:23.814901 4897 generic.go:334] "Generic (PLEG): container finished" podID="60077aeb-0865-4eee-b44e-60be41cf4326" containerID="0d91752aeaed8f1597a4e93218c0e815e67f9d1f3e450156fd9a2a5fdc3079f9" exitCode=0 Nov 21 15:30:23 crc kubenswrapper[4897]: I1121 15:30:23.814984 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bnzxb" event={"ID":"60077aeb-0865-4eee-b44e-60be41cf4326","Type":"ContainerDied","Data":"0d91752aeaed8f1597a4e93218c0e815e67f9d1f3e450156fd9a2a5fdc3079f9"} Nov 21 15:30:23 crc kubenswrapper[4897]: I1121 15:30:23.837043 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-frgrd" podStartSLOduration=3.253634608 podStartE2EDuration="7.837017507s" podCreationTimestamp="2025-11-21 15:30:16 +0000 UTC" firstStartedPulling="2025-11-21 15:30:18.74009744 +0000 UTC m=+4896.024690915" lastFinishedPulling="2025-11-21 15:30:23.323480339 +0000 UTC m=+4900.608073814" observedRunningTime="2025-11-21 15:30:23.83455026 +0000 UTC m=+4901.119143735" watchObservedRunningTime="2025-11-21 15:30:23.837017507 +0000 UTC m=+4901.121610982" Nov 21 15:30:24 crc kubenswrapper[4897]: I1121 15:30:24.827715 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bnzxb" event={"ID":"60077aeb-0865-4eee-b44e-60be41cf4326","Type":"ContainerStarted","Data":"dd37f145a4fd7c1393bab3fc534103bebc9ee7b44dad417e1f638d1f0a79455b"} Nov 21 15:30:24 crc kubenswrapper[4897]: I1121 15:30:24.852073 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-bnzxb" podStartSLOduration=3.36348907 podStartE2EDuration="6.852053723s" podCreationTimestamp="2025-11-21 15:30:18 +0000 UTC" firstStartedPulling="2025-11-21 15:30:20.768517719 +0000 UTC m=+4898.053111194" lastFinishedPulling="2025-11-21 15:30:24.257082372 +0000 UTC m=+4901.541675847" observedRunningTime="2025-11-21 15:30:24.850644045 +0000 UTC m=+4902.135237520" watchObservedRunningTime="2025-11-21 15:30:24.852053723 +0000 UTC m=+4902.136647198" Nov 21 15:30:26 crc kubenswrapper[4897]: I1121 15:30:26.930786 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-frgrd" Nov 21 15:30:26 crc kubenswrapper[4897]: I1121 15:30:26.931567 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-frgrd" Nov 21 15:30:28 crc kubenswrapper[4897]: I1121 15:30:28.011401 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-frgrd" podUID="d360d107-e163-45c9-8fa6-d5f47c78a9cc" containerName="registry-server" probeResult="failure" output=< Nov 21 15:30:28 crc kubenswrapper[4897]: timeout: failed to connect service ":50051" within 1s Nov 21 15:30:28 crc kubenswrapper[4897]: > Nov 21 15:30:29 crc kubenswrapper[4897]: I1121 15:30:29.286882 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-bnzxb" Nov 21 15:30:29 crc kubenswrapper[4897]: I1121 15:30:29.287206 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-bnzxb" Nov 21 15:30:29 crc kubenswrapper[4897]: I1121 15:30:29.341111 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-bnzxb" Nov 21 15:30:29 crc kubenswrapper[4897]: I1121 15:30:29.927723 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-bnzxb" Nov 21 15:30:29 crc kubenswrapper[4897]: I1121 15:30:29.980657 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bnzxb"] Nov 21 15:30:31 crc kubenswrapper[4897]: I1121 15:30:31.220090 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-zdspd" podUID="8f40194d-4a22-45a2-8892-d66931e738c5" containerName="registry-server" probeResult="failure" output=< Nov 21 15:30:31 crc kubenswrapper[4897]: timeout: failed to connect service ":50051" within 1s Nov 21 15:30:31 crc kubenswrapper[4897]: > Nov 21 15:30:31 crc kubenswrapper[4897]: I1121 15:30:31.899593 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-bnzxb" podUID="60077aeb-0865-4eee-b44e-60be41cf4326" containerName="registry-server" containerID="cri-o://dd37f145a4fd7c1393bab3fc534103bebc9ee7b44dad417e1f638d1f0a79455b" gracePeriod=2 Nov 21 15:30:32 crc kubenswrapper[4897]: I1121 15:30:32.912247 4897 generic.go:334] "Generic (PLEG): container finished" podID="60077aeb-0865-4eee-b44e-60be41cf4326" containerID="dd37f145a4fd7c1393bab3fc534103bebc9ee7b44dad417e1f638d1f0a79455b" exitCode=0 Nov 21 15:30:32 crc kubenswrapper[4897]: I1121 15:30:32.912281 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bnzxb" event={"ID":"60077aeb-0865-4eee-b44e-60be41cf4326","Type":"ContainerDied","Data":"dd37f145a4fd7c1393bab3fc534103bebc9ee7b44dad417e1f638d1f0a79455b"} Nov 21 15:30:33 crc kubenswrapper[4897]: I1121 15:30:33.927836 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bnzxb" event={"ID":"60077aeb-0865-4eee-b44e-60be41cf4326","Type":"ContainerDied","Data":"0ff91d0b041ab50bf32d36a9ea1f384479355021186f5676bb619fc69cb28a87"} Nov 21 15:30:33 crc kubenswrapper[4897]: I1121 15:30:33.928109 4897 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0ff91d0b041ab50bf32d36a9ea1f384479355021186f5676bb619fc69cb28a87" Nov 21 15:30:33 crc kubenswrapper[4897]: I1121 15:30:33.947060 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bnzxb" Nov 21 15:30:34 crc kubenswrapper[4897]: I1121 15:30:34.120586 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/60077aeb-0865-4eee-b44e-60be41cf4326-catalog-content\") pod \"60077aeb-0865-4eee-b44e-60be41cf4326\" (UID: \"60077aeb-0865-4eee-b44e-60be41cf4326\") " Nov 21 15:30:34 crc kubenswrapper[4897]: I1121 15:30:34.120653 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ktcn8\" (UniqueName: \"kubernetes.io/projected/60077aeb-0865-4eee-b44e-60be41cf4326-kube-api-access-ktcn8\") pod \"60077aeb-0865-4eee-b44e-60be41cf4326\" (UID: \"60077aeb-0865-4eee-b44e-60be41cf4326\") " Nov 21 15:30:34 crc kubenswrapper[4897]: I1121 15:30:34.120787 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/60077aeb-0865-4eee-b44e-60be41cf4326-utilities\") pod \"60077aeb-0865-4eee-b44e-60be41cf4326\" (UID: \"60077aeb-0865-4eee-b44e-60be41cf4326\") " Nov 21 15:30:34 crc kubenswrapper[4897]: I1121 15:30:34.122003 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/60077aeb-0865-4eee-b44e-60be41cf4326-utilities" (OuterVolumeSpecName: "utilities") pod "60077aeb-0865-4eee-b44e-60be41cf4326" (UID: "60077aeb-0865-4eee-b44e-60be41cf4326"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 15:30:34 crc kubenswrapper[4897]: I1121 15:30:34.126845 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60077aeb-0865-4eee-b44e-60be41cf4326-kube-api-access-ktcn8" (OuterVolumeSpecName: "kube-api-access-ktcn8") pod "60077aeb-0865-4eee-b44e-60be41cf4326" (UID: "60077aeb-0865-4eee-b44e-60be41cf4326"). InnerVolumeSpecName "kube-api-access-ktcn8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 15:30:34 crc kubenswrapper[4897]: I1121 15:30:34.224919 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ktcn8\" (UniqueName: \"kubernetes.io/projected/60077aeb-0865-4eee-b44e-60be41cf4326-kube-api-access-ktcn8\") on node \"crc\" DevicePath \"\"" Nov 21 15:30:34 crc kubenswrapper[4897]: I1121 15:30:34.225213 4897 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/60077aeb-0865-4eee-b44e-60be41cf4326-utilities\") on node \"crc\" DevicePath \"\"" Nov 21 15:30:34 crc kubenswrapper[4897]: I1121 15:30:34.259864 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/60077aeb-0865-4eee-b44e-60be41cf4326-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "60077aeb-0865-4eee-b44e-60be41cf4326" (UID: "60077aeb-0865-4eee-b44e-60be41cf4326"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 15:30:34 crc kubenswrapper[4897]: I1121 15:30:34.332994 4897 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/60077aeb-0865-4eee-b44e-60be41cf4326-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 21 15:30:34 crc kubenswrapper[4897]: I1121 15:30:34.941731 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bnzxb" Nov 21 15:30:34 crc kubenswrapper[4897]: I1121 15:30:34.989762 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bnzxb"] Nov 21 15:30:34 crc kubenswrapper[4897]: I1121 15:30:34.999238 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-bnzxb"] Nov 21 15:30:36 crc kubenswrapper[4897]: I1121 15:30:36.101574 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="60077aeb-0865-4eee-b44e-60be41cf4326" path="/var/lib/kubelet/pods/60077aeb-0865-4eee-b44e-60be41cf4326/volumes" Nov 21 15:30:37 crc kubenswrapper[4897]: I1121 15:30:37.993901 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-frgrd" podUID="d360d107-e163-45c9-8fa6-d5f47c78a9cc" containerName="registry-server" probeResult="failure" output=< Nov 21 15:30:37 crc kubenswrapper[4897]: timeout: failed to connect service ":50051" within 1s Nov 21 15:30:37 crc kubenswrapper[4897]: > Nov 21 15:30:40 crc kubenswrapper[4897]: I1121 15:30:40.236758 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-zdspd" Nov 21 15:30:40 crc kubenswrapper[4897]: I1121 15:30:40.290722 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-zdspd" Nov 21 15:30:40 crc kubenswrapper[4897]: I1121 15:30:40.488210 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-zdspd"] Nov 21 15:30:42 crc kubenswrapper[4897]: I1121 15:30:42.008122 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-zdspd" podUID="8f40194d-4a22-45a2-8892-d66931e738c5" containerName="registry-server" containerID="cri-o://478b9ee4cfff944b273e3eea0c75cf9302ae644e035bc8d7d2e4b7e2987ba6ae" gracePeriod=2 Nov 21 15:30:42 crc kubenswrapper[4897]: I1121 15:30:42.472174 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zdspd" Nov 21 15:30:42 crc kubenswrapper[4897]: I1121 15:30:42.631701 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f40194d-4a22-45a2-8892-d66931e738c5-utilities\") pod \"8f40194d-4a22-45a2-8892-d66931e738c5\" (UID: \"8f40194d-4a22-45a2-8892-d66931e738c5\") " Nov 21 15:30:42 crc kubenswrapper[4897]: I1121 15:30:42.631898 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f40194d-4a22-45a2-8892-d66931e738c5-catalog-content\") pod \"8f40194d-4a22-45a2-8892-d66931e738c5\" (UID: \"8f40194d-4a22-45a2-8892-d66931e738c5\") " Nov 21 15:30:42 crc kubenswrapper[4897]: I1121 15:30:42.631982 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5c7kn\" (UniqueName: \"kubernetes.io/projected/8f40194d-4a22-45a2-8892-d66931e738c5-kube-api-access-5c7kn\") pod \"8f40194d-4a22-45a2-8892-d66931e738c5\" (UID: \"8f40194d-4a22-45a2-8892-d66931e738c5\") " Nov 21 15:30:42 crc kubenswrapper[4897]: I1121 15:30:42.632547 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f40194d-4a22-45a2-8892-d66931e738c5-utilities" (OuterVolumeSpecName: "utilities") pod "8f40194d-4a22-45a2-8892-d66931e738c5" (UID: "8f40194d-4a22-45a2-8892-d66931e738c5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 15:30:42 crc kubenswrapper[4897]: I1121 15:30:42.632685 4897 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f40194d-4a22-45a2-8892-d66931e738c5-utilities\") on node \"crc\" DevicePath \"\"" Nov 21 15:30:42 crc kubenswrapper[4897]: I1121 15:30:42.640565 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f40194d-4a22-45a2-8892-d66931e738c5-kube-api-access-5c7kn" (OuterVolumeSpecName: "kube-api-access-5c7kn") pod "8f40194d-4a22-45a2-8892-d66931e738c5" (UID: "8f40194d-4a22-45a2-8892-d66931e738c5"). InnerVolumeSpecName "kube-api-access-5c7kn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 15:30:42 crc kubenswrapper[4897]: I1121 15:30:42.717324 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f40194d-4a22-45a2-8892-d66931e738c5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8f40194d-4a22-45a2-8892-d66931e738c5" (UID: "8f40194d-4a22-45a2-8892-d66931e738c5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 15:30:42 crc kubenswrapper[4897]: I1121 15:30:42.735256 4897 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f40194d-4a22-45a2-8892-d66931e738c5-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 21 15:30:42 crc kubenswrapper[4897]: I1121 15:30:42.735292 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5c7kn\" (UniqueName: \"kubernetes.io/projected/8f40194d-4a22-45a2-8892-d66931e738c5-kube-api-access-5c7kn\") on node \"crc\" DevicePath \"\"" Nov 21 15:30:43 crc kubenswrapper[4897]: I1121 15:30:43.020275 4897 generic.go:334] "Generic (PLEG): container finished" podID="8f40194d-4a22-45a2-8892-d66931e738c5" containerID="478b9ee4cfff944b273e3eea0c75cf9302ae644e035bc8d7d2e4b7e2987ba6ae" exitCode=0 Nov 21 15:30:43 crc kubenswrapper[4897]: I1121 15:30:43.020338 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zdspd" Nov 21 15:30:43 crc kubenswrapper[4897]: I1121 15:30:43.020364 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zdspd" event={"ID":"8f40194d-4a22-45a2-8892-d66931e738c5","Type":"ContainerDied","Data":"478b9ee4cfff944b273e3eea0c75cf9302ae644e035bc8d7d2e4b7e2987ba6ae"} Nov 21 15:30:43 crc kubenswrapper[4897]: I1121 15:30:43.020856 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zdspd" event={"ID":"8f40194d-4a22-45a2-8892-d66931e738c5","Type":"ContainerDied","Data":"e163e2ff7dc58b335051a7c2cf6b96d0a60c64f065b1c084ff7f9b3ee214c6f4"} Nov 21 15:30:43 crc kubenswrapper[4897]: I1121 15:30:43.020892 4897 scope.go:117] "RemoveContainer" containerID="478b9ee4cfff944b273e3eea0c75cf9302ae644e035bc8d7d2e4b7e2987ba6ae" Nov 21 15:30:43 crc kubenswrapper[4897]: I1121 15:30:43.066221 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-zdspd"] Nov 21 15:30:43 crc kubenswrapper[4897]: I1121 15:30:43.066940 4897 scope.go:117] "RemoveContainer" containerID="61d1447cb9750806e7956941105921d5f366a6ebbc6c96540330f603fca858fe" Nov 21 15:30:43 crc kubenswrapper[4897]: I1121 15:30:43.080091 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-zdspd"] Nov 21 15:30:43 crc kubenswrapper[4897]: I1121 15:30:43.103696 4897 scope.go:117] "RemoveContainer" containerID="9f42ad5bb82bb271bbb11b79ecf595b05c81deef1e6a991a86b36dc7782b67d5" Nov 21 15:30:43 crc kubenswrapper[4897]: I1121 15:30:43.157352 4897 scope.go:117] "RemoveContainer" containerID="478b9ee4cfff944b273e3eea0c75cf9302ae644e035bc8d7d2e4b7e2987ba6ae" Nov 21 15:30:43 crc kubenswrapper[4897]: E1121 15:30:43.158010 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"478b9ee4cfff944b273e3eea0c75cf9302ae644e035bc8d7d2e4b7e2987ba6ae\": container with ID starting with 478b9ee4cfff944b273e3eea0c75cf9302ae644e035bc8d7d2e4b7e2987ba6ae not found: ID does not exist" containerID="478b9ee4cfff944b273e3eea0c75cf9302ae644e035bc8d7d2e4b7e2987ba6ae" Nov 21 15:30:43 crc kubenswrapper[4897]: I1121 15:30:43.158117 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"478b9ee4cfff944b273e3eea0c75cf9302ae644e035bc8d7d2e4b7e2987ba6ae"} err="failed to get container status \"478b9ee4cfff944b273e3eea0c75cf9302ae644e035bc8d7d2e4b7e2987ba6ae\": rpc error: code = NotFound desc = could not find container \"478b9ee4cfff944b273e3eea0c75cf9302ae644e035bc8d7d2e4b7e2987ba6ae\": container with ID starting with 478b9ee4cfff944b273e3eea0c75cf9302ae644e035bc8d7d2e4b7e2987ba6ae not found: ID does not exist" Nov 21 15:30:43 crc kubenswrapper[4897]: I1121 15:30:43.158243 4897 scope.go:117] "RemoveContainer" containerID="61d1447cb9750806e7956941105921d5f366a6ebbc6c96540330f603fca858fe" Nov 21 15:30:43 crc kubenswrapper[4897]: E1121 15:30:43.158597 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"61d1447cb9750806e7956941105921d5f366a6ebbc6c96540330f603fca858fe\": container with ID starting with 61d1447cb9750806e7956941105921d5f366a6ebbc6c96540330f603fca858fe not found: ID does not exist" containerID="61d1447cb9750806e7956941105921d5f366a6ebbc6c96540330f603fca858fe" Nov 21 15:30:43 crc kubenswrapper[4897]: I1121 15:30:43.158717 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61d1447cb9750806e7956941105921d5f366a6ebbc6c96540330f603fca858fe"} err="failed to get container status \"61d1447cb9750806e7956941105921d5f366a6ebbc6c96540330f603fca858fe\": rpc error: code = NotFound desc = could not find container \"61d1447cb9750806e7956941105921d5f366a6ebbc6c96540330f603fca858fe\": container with ID starting with 61d1447cb9750806e7956941105921d5f366a6ebbc6c96540330f603fca858fe not found: ID does not exist" Nov 21 15:30:43 crc kubenswrapper[4897]: I1121 15:30:43.158814 4897 scope.go:117] "RemoveContainer" containerID="9f42ad5bb82bb271bbb11b79ecf595b05c81deef1e6a991a86b36dc7782b67d5" Nov 21 15:30:43 crc kubenswrapper[4897]: E1121 15:30:43.159676 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9f42ad5bb82bb271bbb11b79ecf595b05c81deef1e6a991a86b36dc7782b67d5\": container with ID starting with 9f42ad5bb82bb271bbb11b79ecf595b05c81deef1e6a991a86b36dc7782b67d5 not found: ID does not exist" containerID="9f42ad5bb82bb271bbb11b79ecf595b05c81deef1e6a991a86b36dc7782b67d5" Nov 21 15:30:43 crc kubenswrapper[4897]: I1121 15:30:43.159772 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f42ad5bb82bb271bbb11b79ecf595b05c81deef1e6a991a86b36dc7782b67d5"} err="failed to get container status \"9f42ad5bb82bb271bbb11b79ecf595b05c81deef1e6a991a86b36dc7782b67d5\": rpc error: code = NotFound desc = could not find container \"9f42ad5bb82bb271bbb11b79ecf595b05c81deef1e6a991a86b36dc7782b67d5\": container with ID starting with 9f42ad5bb82bb271bbb11b79ecf595b05c81deef1e6a991a86b36dc7782b67d5 not found: ID does not exist" Nov 21 15:30:44 crc kubenswrapper[4897]: I1121 15:30:44.102230 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f40194d-4a22-45a2-8892-d66931e738c5" path="/var/lib/kubelet/pods/8f40194d-4a22-45a2-8892-d66931e738c5/volumes" Nov 21 15:30:46 crc kubenswrapper[4897]: I1121 15:30:46.998295 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-frgrd" Nov 21 15:30:47 crc kubenswrapper[4897]: I1121 15:30:47.756176 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-frgrd" Nov 21 15:30:47 crc kubenswrapper[4897]: I1121 15:30:47.888902 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-frgrd"] Nov 21 15:30:48 crc kubenswrapper[4897]: I1121 15:30:48.076173 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-frgrd" podUID="d360d107-e163-45c9-8fa6-d5f47c78a9cc" containerName="registry-server" containerID="cri-o://e48d9667959f644ef2ed5f8fca5c3a8a67489ac32bfe162586f8b39633975614" gracePeriod=2 Nov 21 15:30:48 crc kubenswrapper[4897]: I1121 15:30:48.627936 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-frgrd" Nov 21 15:30:48 crc kubenswrapper[4897]: I1121 15:30:48.801841 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d360d107-e163-45c9-8fa6-d5f47c78a9cc-catalog-content\") pod \"d360d107-e163-45c9-8fa6-d5f47c78a9cc\" (UID: \"d360d107-e163-45c9-8fa6-d5f47c78a9cc\") " Nov 21 15:30:48 crc kubenswrapper[4897]: I1121 15:30:48.802102 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-92chg\" (UniqueName: \"kubernetes.io/projected/d360d107-e163-45c9-8fa6-d5f47c78a9cc-kube-api-access-92chg\") pod \"d360d107-e163-45c9-8fa6-d5f47c78a9cc\" (UID: \"d360d107-e163-45c9-8fa6-d5f47c78a9cc\") " Nov 21 15:30:48 crc kubenswrapper[4897]: I1121 15:30:48.802141 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d360d107-e163-45c9-8fa6-d5f47c78a9cc-utilities\") pod \"d360d107-e163-45c9-8fa6-d5f47c78a9cc\" (UID: \"d360d107-e163-45c9-8fa6-d5f47c78a9cc\") " Nov 21 15:30:48 crc kubenswrapper[4897]: I1121 15:30:48.802801 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d360d107-e163-45c9-8fa6-d5f47c78a9cc-utilities" (OuterVolumeSpecName: "utilities") pod "d360d107-e163-45c9-8fa6-d5f47c78a9cc" (UID: "d360d107-e163-45c9-8fa6-d5f47c78a9cc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 15:30:48 crc kubenswrapper[4897]: I1121 15:30:48.802930 4897 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d360d107-e163-45c9-8fa6-d5f47c78a9cc-utilities\") on node \"crc\" DevicePath \"\"" Nov 21 15:30:48 crc kubenswrapper[4897]: I1121 15:30:48.810536 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d360d107-e163-45c9-8fa6-d5f47c78a9cc-kube-api-access-92chg" (OuterVolumeSpecName: "kube-api-access-92chg") pod "d360d107-e163-45c9-8fa6-d5f47c78a9cc" (UID: "d360d107-e163-45c9-8fa6-d5f47c78a9cc"). InnerVolumeSpecName "kube-api-access-92chg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 15:30:48 crc kubenswrapper[4897]: I1121 15:30:48.852855 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d360d107-e163-45c9-8fa6-d5f47c78a9cc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d360d107-e163-45c9-8fa6-d5f47c78a9cc" (UID: "d360d107-e163-45c9-8fa6-d5f47c78a9cc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 15:30:48 crc kubenswrapper[4897]: I1121 15:30:48.904550 4897 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d360d107-e163-45c9-8fa6-d5f47c78a9cc-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 21 15:30:48 crc kubenswrapper[4897]: I1121 15:30:48.904586 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-92chg\" (UniqueName: \"kubernetes.io/projected/d360d107-e163-45c9-8fa6-d5f47c78a9cc-kube-api-access-92chg\") on node \"crc\" DevicePath \"\"" Nov 21 15:30:49 crc kubenswrapper[4897]: I1121 15:30:49.092019 4897 generic.go:334] "Generic (PLEG): container finished" podID="d360d107-e163-45c9-8fa6-d5f47c78a9cc" containerID="e48d9667959f644ef2ed5f8fca5c3a8a67489ac32bfe162586f8b39633975614" exitCode=0 Nov 21 15:30:49 crc kubenswrapper[4897]: I1121 15:30:49.092078 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-frgrd" Nov 21 15:30:49 crc kubenswrapper[4897]: I1121 15:30:49.092089 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-frgrd" event={"ID":"d360d107-e163-45c9-8fa6-d5f47c78a9cc","Type":"ContainerDied","Data":"e48d9667959f644ef2ed5f8fca5c3a8a67489ac32bfe162586f8b39633975614"} Nov 21 15:30:49 crc kubenswrapper[4897]: I1121 15:30:49.093052 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-frgrd" event={"ID":"d360d107-e163-45c9-8fa6-d5f47c78a9cc","Type":"ContainerDied","Data":"6ad11fb2779ce57171120eb8a2064304a553b0b2751cbb34bb1d7b926728cb39"} Nov 21 15:30:49 crc kubenswrapper[4897]: I1121 15:30:49.093079 4897 scope.go:117] "RemoveContainer" containerID="e48d9667959f644ef2ed5f8fca5c3a8a67489ac32bfe162586f8b39633975614" Nov 21 15:30:49 crc kubenswrapper[4897]: I1121 15:30:49.131249 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-frgrd"] Nov 21 15:30:49 crc kubenswrapper[4897]: I1121 15:30:49.132237 4897 scope.go:117] "RemoveContainer" containerID="2986347fe6fceca60228697b3b4f7cf09c6ccee53de91a6b0e75d1a75313b43b" Nov 21 15:30:49 crc kubenswrapper[4897]: I1121 15:30:49.144179 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-frgrd"] Nov 21 15:30:49 crc kubenswrapper[4897]: I1121 15:30:49.156833 4897 scope.go:117] "RemoveContainer" containerID="0d1a541ac7151fc5c88b28da7b0f36d9efb37de9100a43f8719fbfe359263a66" Nov 21 15:30:49 crc kubenswrapper[4897]: I1121 15:30:49.231472 4897 scope.go:117] "RemoveContainer" containerID="e48d9667959f644ef2ed5f8fca5c3a8a67489ac32bfe162586f8b39633975614" Nov 21 15:30:49 crc kubenswrapper[4897]: E1121 15:30:49.232073 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e48d9667959f644ef2ed5f8fca5c3a8a67489ac32bfe162586f8b39633975614\": container with ID starting with e48d9667959f644ef2ed5f8fca5c3a8a67489ac32bfe162586f8b39633975614 not found: ID does not exist" containerID="e48d9667959f644ef2ed5f8fca5c3a8a67489ac32bfe162586f8b39633975614" Nov 21 15:30:49 crc kubenswrapper[4897]: I1121 15:30:49.232109 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e48d9667959f644ef2ed5f8fca5c3a8a67489ac32bfe162586f8b39633975614"} err="failed to get container status \"e48d9667959f644ef2ed5f8fca5c3a8a67489ac32bfe162586f8b39633975614\": rpc error: code = NotFound desc = could not find container \"e48d9667959f644ef2ed5f8fca5c3a8a67489ac32bfe162586f8b39633975614\": container with ID starting with e48d9667959f644ef2ed5f8fca5c3a8a67489ac32bfe162586f8b39633975614 not found: ID does not exist" Nov 21 15:30:49 crc kubenswrapper[4897]: I1121 15:30:49.232137 4897 scope.go:117] "RemoveContainer" containerID="2986347fe6fceca60228697b3b4f7cf09c6ccee53de91a6b0e75d1a75313b43b" Nov 21 15:30:49 crc kubenswrapper[4897]: E1121 15:30:49.232814 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2986347fe6fceca60228697b3b4f7cf09c6ccee53de91a6b0e75d1a75313b43b\": container with ID starting with 2986347fe6fceca60228697b3b4f7cf09c6ccee53de91a6b0e75d1a75313b43b not found: ID does not exist" containerID="2986347fe6fceca60228697b3b4f7cf09c6ccee53de91a6b0e75d1a75313b43b" Nov 21 15:30:49 crc kubenswrapper[4897]: I1121 15:30:49.232840 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2986347fe6fceca60228697b3b4f7cf09c6ccee53de91a6b0e75d1a75313b43b"} err="failed to get container status \"2986347fe6fceca60228697b3b4f7cf09c6ccee53de91a6b0e75d1a75313b43b\": rpc error: code = NotFound desc = could not find container \"2986347fe6fceca60228697b3b4f7cf09c6ccee53de91a6b0e75d1a75313b43b\": container with ID starting with 2986347fe6fceca60228697b3b4f7cf09c6ccee53de91a6b0e75d1a75313b43b not found: ID does not exist" Nov 21 15:30:49 crc kubenswrapper[4897]: I1121 15:30:49.232854 4897 scope.go:117] "RemoveContainer" containerID="0d1a541ac7151fc5c88b28da7b0f36d9efb37de9100a43f8719fbfe359263a66" Nov 21 15:30:49 crc kubenswrapper[4897]: E1121 15:30:49.233189 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0d1a541ac7151fc5c88b28da7b0f36d9efb37de9100a43f8719fbfe359263a66\": container with ID starting with 0d1a541ac7151fc5c88b28da7b0f36d9efb37de9100a43f8719fbfe359263a66 not found: ID does not exist" containerID="0d1a541ac7151fc5c88b28da7b0f36d9efb37de9100a43f8719fbfe359263a66" Nov 21 15:30:49 crc kubenswrapper[4897]: I1121 15:30:49.233225 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d1a541ac7151fc5c88b28da7b0f36d9efb37de9100a43f8719fbfe359263a66"} err="failed to get container status \"0d1a541ac7151fc5c88b28da7b0f36d9efb37de9100a43f8719fbfe359263a66\": rpc error: code = NotFound desc = could not find container \"0d1a541ac7151fc5c88b28da7b0f36d9efb37de9100a43f8719fbfe359263a66\": container with ID starting with 0d1a541ac7151fc5c88b28da7b0f36d9efb37de9100a43f8719fbfe359263a66 not found: ID does not exist" Nov 21 15:30:50 crc kubenswrapper[4897]: I1121 15:30:50.106913 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d360d107-e163-45c9-8fa6-d5f47c78a9cc" path="/var/lib/kubelet/pods/d360d107-e163-45c9-8fa6-d5f47c78a9cc/volumes" Nov 21 15:31:58 crc kubenswrapper[4897]: E1121 15:31:58.433003 4897 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.233:51230->38.102.83.233:43735: write tcp 38.102.83.233:51230->38.102.83.233:43735: write: broken pipe Nov 21 15:32:04 crc kubenswrapper[4897]: I1121 15:32:04.371277 4897 patch_prober.go:28] interesting pod/machine-config-daemon-krv5b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 21 15:32:04 crc kubenswrapper[4897]: I1121 15:32:04.371790 4897 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 21 15:32:34 crc kubenswrapper[4897]: I1121 15:32:34.370395 4897 patch_prober.go:28] interesting pod/machine-config-daemon-krv5b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 21 15:32:34 crc kubenswrapper[4897]: I1121 15:32:34.371092 4897 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 21 15:33:04 crc kubenswrapper[4897]: I1121 15:33:04.370977 4897 patch_prober.go:28] interesting pod/machine-config-daemon-krv5b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 21 15:33:04 crc kubenswrapper[4897]: I1121 15:33:04.371540 4897 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 21 15:33:04 crc kubenswrapper[4897]: I1121 15:33:04.371588 4897 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" Nov 21 15:33:04 crc kubenswrapper[4897]: I1121 15:33:04.372407 4897 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"abf30c46658df7e750da25f3feeb4a134367a6c61c5b5221d6703974a1dd4d2e"} pod="openshift-machine-config-operator/machine-config-daemon-krv5b" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 21 15:33:04 crc kubenswrapper[4897]: I1121 15:33:04.372457 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" containerName="machine-config-daemon" containerID="cri-o://abf30c46658df7e750da25f3feeb4a134367a6c61c5b5221d6703974a1dd4d2e" gracePeriod=600 Nov 21 15:33:04 crc kubenswrapper[4897]: I1121 15:33:04.595610 4897 generic.go:334] "Generic (PLEG): container finished" podID="e7670227-d280-4847-b882-754429f56b0f" containerID="abf30c46658df7e750da25f3feeb4a134367a6c61c5b5221d6703974a1dd4d2e" exitCode=0 Nov 21 15:33:04 crc kubenswrapper[4897]: I1121 15:33:04.595653 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" event={"ID":"e7670227-d280-4847-b882-754429f56b0f","Type":"ContainerDied","Data":"abf30c46658df7e750da25f3feeb4a134367a6c61c5b5221d6703974a1dd4d2e"} Nov 21 15:33:04 crc kubenswrapper[4897]: I1121 15:33:04.595690 4897 scope.go:117] "RemoveContainer" containerID="2a7c1a6f8fe489b93e0267f838232cae90d796b8213c4aa28e9ba6aeb840a22d" Nov 21 15:33:04 crc kubenswrapper[4897]: E1121 15:33:04.928607 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:33:05 crc kubenswrapper[4897]: I1121 15:33:05.609117 4897 scope.go:117] "RemoveContainer" containerID="abf30c46658df7e750da25f3feeb4a134367a6c61c5b5221d6703974a1dd4d2e" Nov 21 15:33:05 crc kubenswrapper[4897]: E1121 15:33:05.609714 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:33:18 crc kubenswrapper[4897]: I1121 15:33:18.089948 4897 scope.go:117] "RemoveContainer" containerID="abf30c46658df7e750da25f3feeb4a134367a6c61c5b5221d6703974a1dd4d2e" Nov 21 15:33:18 crc kubenswrapper[4897]: E1121 15:33:18.090906 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:33:29 crc kubenswrapper[4897]: I1121 15:33:29.089730 4897 scope.go:117] "RemoveContainer" containerID="abf30c46658df7e750da25f3feeb4a134367a6c61c5b5221d6703974a1dd4d2e" Nov 21 15:33:29 crc kubenswrapper[4897]: E1121 15:33:29.090764 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:33:44 crc kubenswrapper[4897]: I1121 15:33:44.102638 4897 scope.go:117] "RemoveContainer" containerID="abf30c46658df7e750da25f3feeb4a134367a6c61c5b5221d6703974a1dd4d2e" Nov 21 15:33:44 crc kubenswrapper[4897]: E1121 15:33:44.103813 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:33:51 crc kubenswrapper[4897]: I1121 15:33:51.805468 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-v6j6v"] Nov 21 15:33:51 crc kubenswrapper[4897]: E1121 15:33:51.807793 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d360d107-e163-45c9-8fa6-d5f47c78a9cc" containerName="extract-utilities" Nov 21 15:33:51 crc kubenswrapper[4897]: I1121 15:33:51.807825 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="d360d107-e163-45c9-8fa6-d5f47c78a9cc" containerName="extract-utilities" Nov 21 15:33:51 crc kubenswrapper[4897]: E1121 15:33:51.807873 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60077aeb-0865-4eee-b44e-60be41cf4326" containerName="registry-server" Nov 21 15:33:51 crc kubenswrapper[4897]: I1121 15:33:51.807882 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="60077aeb-0865-4eee-b44e-60be41cf4326" containerName="registry-server" Nov 21 15:33:51 crc kubenswrapper[4897]: E1121 15:33:51.807901 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60077aeb-0865-4eee-b44e-60be41cf4326" containerName="extract-content" Nov 21 15:33:51 crc kubenswrapper[4897]: I1121 15:33:51.807907 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="60077aeb-0865-4eee-b44e-60be41cf4326" containerName="extract-content" Nov 21 15:33:51 crc kubenswrapper[4897]: E1121 15:33:51.807930 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60077aeb-0865-4eee-b44e-60be41cf4326" containerName="extract-utilities" Nov 21 15:33:51 crc kubenswrapper[4897]: I1121 15:33:51.807935 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="60077aeb-0865-4eee-b44e-60be41cf4326" containerName="extract-utilities" Nov 21 15:33:51 crc kubenswrapper[4897]: E1121 15:33:51.807951 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f40194d-4a22-45a2-8892-d66931e738c5" containerName="extract-utilities" Nov 21 15:33:51 crc kubenswrapper[4897]: I1121 15:33:51.807956 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f40194d-4a22-45a2-8892-d66931e738c5" containerName="extract-utilities" Nov 21 15:33:51 crc kubenswrapper[4897]: E1121 15:33:51.807971 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d360d107-e163-45c9-8fa6-d5f47c78a9cc" containerName="extract-content" Nov 21 15:33:51 crc kubenswrapper[4897]: I1121 15:33:51.807977 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="d360d107-e163-45c9-8fa6-d5f47c78a9cc" containerName="extract-content" Nov 21 15:33:51 crc kubenswrapper[4897]: E1121 15:33:51.807986 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f40194d-4a22-45a2-8892-d66931e738c5" containerName="registry-server" Nov 21 15:33:51 crc kubenswrapper[4897]: I1121 15:33:51.807992 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f40194d-4a22-45a2-8892-d66931e738c5" containerName="registry-server" Nov 21 15:33:51 crc kubenswrapper[4897]: E1121 15:33:51.808002 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d360d107-e163-45c9-8fa6-d5f47c78a9cc" containerName="registry-server" Nov 21 15:33:51 crc kubenswrapper[4897]: I1121 15:33:51.808007 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="d360d107-e163-45c9-8fa6-d5f47c78a9cc" containerName="registry-server" Nov 21 15:33:51 crc kubenswrapper[4897]: E1121 15:33:51.808048 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f40194d-4a22-45a2-8892-d66931e738c5" containerName="extract-content" Nov 21 15:33:51 crc kubenswrapper[4897]: I1121 15:33:51.808054 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f40194d-4a22-45a2-8892-d66931e738c5" containerName="extract-content" Nov 21 15:33:51 crc kubenswrapper[4897]: I1121 15:33:51.808341 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f40194d-4a22-45a2-8892-d66931e738c5" containerName="registry-server" Nov 21 15:33:51 crc kubenswrapper[4897]: I1121 15:33:51.808362 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="d360d107-e163-45c9-8fa6-d5f47c78a9cc" containerName="registry-server" Nov 21 15:33:51 crc kubenswrapper[4897]: I1121 15:33:51.808377 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="60077aeb-0865-4eee-b44e-60be41cf4326" containerName="registry-server" Nov 21 15:33:51 crc kubenswrapper[4897]: I1121 15:33:51.810545 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v6j6v" Nov 21 15:33:51 crc kubenswrapper[4897]: I1121 15:33:51.836960 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-v6j6v"] Nov 21 15:33:51 crc kubenswrapper[4897]: I1121 15:33:51.984598 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b9b715c-1305-4968-b49f-488ce8c9688c-catalog-content\") pod \"community-operators-v6j6v\" (UID: \"7b9b715c-1305-4968-b49f-488ce8c9688c\") " pod="openshift-marketplace/community-operators-v6j6v" Nov 21 15:33:51 crc kubenswrapper[4897]: I1121 15:33:51.984764 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b9b715c-1305-4968-b49f-488ce8c9688c-utilities\") pod \"community-operators-v6j6v\" (UID: \"7b9b715c-1305-4968-b49f-488ce8c9688c\") " pod="openshift-marketplace/community-operators-v6j6v" Nov 21 15:33:51 crc kubenswrapper[4897]: I1121 15:33:51.984808 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bldnl\" (UniqueName: \"kubernetes.io/projected/7b9b715c-1305-4968-b49f-488ce8c9688c-kube-api-access-bldnl\") pod \"community-operators-v6j6v\" (UID: \"7b9b715c-1305-4968-b49f-488ce8c9688c\") " pod="openshift-marketplace/community-operators-v6j6v" Nov 21 15:33:52 crc kubenswrapper[4897]: I1121 15:33:52.087219 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b9b715c-1305-4968-b49f-488ce8c9688c-catalog-content\") pod \"community-operators-v6j6v\" (UID: \"7b9b715c-1305-4968-b49f-488ce8c9688c\") " pod="openshift-marketplace/community-operators-v6j6v" Nov 21 15:33:52 crc kubenswrapper[4897]: I1121 15:33:52.087315 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b9b715c-1305-4968-b49f-488ce8c9688c-utilities\") pod \"community-operators-v6j6v\" (UID: \"7b9b715c-1305-4968-b49f-488ce8c9688c\") " pod="openshift-marketplace/community-operators-v6j6v" Nov 21 15:33:52 crc kubenswrapper[4897]: I1121 15:33:52.087341 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bldnl\" (UniqueName: \"kubernetes.io/projected/7b9b715c-1305-4968-b49f-488ce8c9688c-kube-api-access-bldnl\") pod \"community-operators-v6j6v\" (UID: \"7b9b715c-1305-4968-b49f-488ce8c9688c\") " pod="openshift-marketplace/community-operators-v6j6v" Nov 21 15:33:52 crc kubenswrapper[4897]: I1121 15:33:52.087877 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b9b715c-1305-4968-b49f-488ce8c9688c-catalog-content\") pod \"community-operators-v6j6v\" (UID: \"7b9b715c-1305-4968-b49f-488ce8c9688c\") " pod="openshift-marketplace/community-operators-v6j6v" Nov 21 15:33:52 crc kubenswrapper[4897]: I1121 15:33:52.087905 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b9b715c-1305-4968-b49f-488ce8c9688c-utilities\") pod \"community-operators-v6j6v\" (UID: \"7b9b715c-1305-4968-b49f-488ce8c9688c\") " pod="openshift-marketplace/community-operators-v6j6v" Nov 21 15:33:52 crc kubenswrapper[4897]: I1121 15:33:52.116117 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bldnl\" (UniqueName: \"kubernetes.io/projected/7b9b715c-1305-4968-b49f-488ce8c9688c-kube-api-access-bldnl\") pod \"community-operators-v6j6v\" (UID: \"7b9b715c-1305-4968-b49f-488ce8c9688c\") " pod="openshift-marketplace/community-operators-v6j6v" Nov 21 15:33:52 crc kubenswrapper[4897]: I1121 15:33:52.139223 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v6j6v" Nov 21 15:33:52 crc kubenswrapper[4897]: I1121 15:33:52.820066 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-v6j6v"] Nov 21 15:33:53 crc kubenswrapper[4897]: I1121 15:33:53.100415 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v6j6v" event={"ID":"7b9b715c-1305-4968-b49f-488ce8c9688c","Type":"ContainerStarted","Data":"4324b18d2a4ec02cbdd25fbbeb1c9e09f031a08c590d07175f8c2a71944c5368"} Nov 21 15:33:53 crc kubenswrapper[4897]: I1121 15:33:53.100797 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v6j6v" event={"ID":"7b9b715c-1305-4968-b49f-488ce8c9688c","Type":"ContainerStarted","Data":"48aa7c946c87d3a6d397504bf97a947e22d665592d8c7c2bff61f6ebe9b65769"} Nov 21 15:33:54 crc kubenswrapper[4897]: I1121 15:33:54.111634 4897 generic.go:334] "Generic (PLEG): container finished" podID="7b9b715c-1305-4968-b49f-488ce8c9688c" containerID="4324b18d2a4ec02cbdd25fbbeb1c9e09f031a08c590d07175f8c2a71944c5368" exitCode=0 Nov 21 15:33:54 crc kubenswrapper[4897]: I1121 15:33:54.111706 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v6j6v" event={"ID":"7b9b715c-1305-4968-b49f-488ce8c9688c","Type":"ContainerDied","Data":"4324b18d2a4ec02cbdd25fbbeb1c9e09f031a08c590d07175f8c2a71944c5368"} Nov 21 15:33:55 crc kubenswrapper[4897]: I1121 15:33:55.125727 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v6j6v" event={"ID":"7b9b715c-1305-4968-b49f-488ce8c9688c","Type":"ContainerStarted","Data":"036705e478c45896e7b67526aacba76928ac9a4c1c1e35414d0e0e4adfd6b38c"} Nov 21 15:33:56 crc kubenswrapper[4897]: I1121 15:33:56.089431 4897 scope.go:117] "RemoveContainer" containerID="abf30c46658df7e750da25f3feeb4a134367a6c61c5b5221d6703974a1dd4d2e" Nov 21 15:33:56 crc kubenswrapper[4897]: E1121 15:33:56.090035 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:33:57 crc kubenswrapper[4897]: I1121 15:33:57.177057 4897 generic.go:334] "Generic (PLEG): container finished" podID="7b9b715c-1305-4968-b49f-488ce8c9688c" containerID="036705e478c45896e7b67526aacba76928ac9a4c1c1e35414d0e0e4adfd6b38c" exitCode=0 Nov 21 15:33:57 crc kubenswrapper[4897]: I1121 15:33:57.177639 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v6j6v" event={"ID":"7b9b715c-1305-4968-b49f-488ce8c9688c","Type":"ContainerDied","Data":"036705e478c45896e7b67526aacba76928ac9a4c1c1e35414d0e0e4adfd6b38c"} Nov 21 15:33:58 crc kubenswrapper[4897]: I1121 15:33:58.191904 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v6j6v" event={"ID":"7b9b715c-1305-4968-b49f-488ce8c9688c","Type":"ContainerStarted","Data":"b31a4cad4f7e0bad138a5e80f7c1df147846cc303ce2ee9af988db3d48595701"} Nov 21 15:33:58 crc kubenswrapper[4897]: I1121 15:33:58.223775 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-v6j6v" podStartSLOduration=2.7232390669999997 podStartE2EDuration="7.223750352s" podCreationTimestamp="2025-11-21 15:33:51 +0000 UTC" firstStartedPulling="2025-11-21 15:33:53.102982318 +0000 UTC m=+5110.387575793" lastFinishedPulling="2025-11-21 15:33:57.603493603 +0000 UTC m=+5114.888087078" observedRunningTime="2025-11-21 15:33:58.209044293 +0000 UTC m=+5115.493637768" watchObservedRunningTime="2025-11-21 15:33:58.223750352 +0000 UTC m=+5115.508343827" Nov 21 15:34:02 crc kubenswrapper[4897]: I1121 15:34:02.141131 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-v6j6v" Nov 21 15:34:02 crc kubenswrapper[4897]: I1121 15:34:02.142802 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-v6j6v" Nov 21 15:34:02 crc kubenswrapper[4897]: I1121 15:34:02.219694 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-v6j6v" Nov 21 15:34:02 crc kubenswrapper[4897]: I1121 15:34:02.301238 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-v6j6v" Nov 21 15:34:02 crc kubenswrapper[4897]: I1121 15:34:02.466587 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-v6j6v"] Nov 21 15:34:04 crc kubenswrapper[4897]: I1121 15:34:04.312678 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-v6j6v" podUID="7b9b715c-1305-4968-b49f-488ce8c9688c" containerName="registry-server" containerID="cri-o://b31a4cad4f7e0bad138a5e80f7c1df147846cc303ce2ee9af988db3d48595701" gracePeriod=2 Nov 21 15:34:04 crc kubenswrapper[4897]: I1121 15:34:04.910541 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v6j6v" Nov 21 15:34:05 crc kubenswrapper[4897]: I1121 15:34:05.024559 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b9b715c-1305-4968-b49f-488ce8c9688c-catalog-content\") pod \"7b9b715c-1305-4968-b49f-488ce8c9688c\" (UID: \"7b9b715c-1305-4968-b49f-488ce8c9688c\") " Nov 21 15:34:05 crc kubenswrapper[4897]: I1121 15:34:05.025089 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b9b715c-1305-4968-b49f-488ce8c9688c-utilities\") pod \"7b9b715c-1305-4968-b49f-488ce8c9688c\" (UID: \"7b9b715c-1305-4968-b49f-488ce8c9688c\") " Nov 21 15:34:05 crc kubenswrapper[4897]: I1121 15:34:05.025455 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bldnl\" (UniqueName: \"kubernetes.io/projected/7b9b715c-1305-4968-b49f-488ce8c9688c-kube-api-access-bldnl\") pod \"7b9b715c-1305-4968-b49f-488ce8c9688c\" (UID: \"7b9b715c-1305-4968-b49f-488ce8c9688c\") " Nov 21 15:34:05 crc kubenswrapper[4897]: I1121 15:34:05.026130 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7b9b715c-1305-4968-b49f-488ce8c9688c-utilities" (OuterVolumeSpecName: "utilities") pod "7b9b715c-1305-4968-b49f-488ce8c9688c" (UID: "7b9b715c-1305-4968-b49f-488ce8c9688c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 15:34:05 crc kubenswrapper[4897]: I1121 15:34:05.027212 4897 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b9b715c-1305-4968-b49f-488ce8c9688c-utilities\") on node \"crc\" DevicePath \"\"" Nov 21 15:34:05 crc kubenswrapper[4897]: I1121 15:34:05.036448 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b9b715c-1305-4968-b49f-488ce8c9688c-kube-api-access-bldnl" (OuterVolumeSpecName: "kube-api-access-bldnl") pod "7b9b715c-1305-4968-b49f-488ce8c9688c" (UID: "7b9b715c-1305-4968-b49f-488ce8c9688c"). InnerVolumeSpecName "kube-api-access-bldnl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 15:34:05 crc kubenswrapper[4897]: I1121 15:34:05.099095 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7b9b715c-1305-4968-b49f-488ce8c9688c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7b9b715c-1305-4968-b49f-488ce8c9688c" (UID: "7b9b715c-1305-4968-b49f-488ce8c9688c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 15:34:05 crc kubenswrapper[4897]: I1121 15:34:05.129667 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bldnl\" (UniqueName: \"kubernetes.io/projected/7b9b715c-1305-4968-b49f-488ce8c9688c-kube-api-access-bldnl\") on node \"crc\" DevicePath \"\"" Nov 21 15:34:05 crc kubenswrapper[4897]: I1121 15:34:05.130657 4897 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b9b715c-1305-4968-b49f-488ce8c9688c-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 21 15:34:05 crc kubenswrapper[4897]: I1121 15:34:05.329671 4897 generic.go:334] "Generic (PLEG): container finished" podID="7b9b715c-1305-4968-b49f-488ce8c9688c" containerID="b31a4cad4f7e0bad138a5e80f7c1df147846cc303ce2ee9af988db3d48595701" exitCode=0 Nov 21 15:34:05 crc kubenswrapper[4897]: I1121 15:34:05.329742 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v6j6v" event={"ID":"7b9b715c-1305-4968-b49f-488ce8c9688c","Type":"ContainerDied","Data":"b31a4cad4f7e0bad138a5e80f7c1df147846cc303ce2ee9af988db3d48595701"} Nov 21 15:34:05 crc kubenswrapper[4897]: I1121 15:34:05.330263 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v6j6v" event={"ID":"7b9b715c-1305-4968-b49f-488ce8c9688c","Type":"ContainerDied","Data":"48aa7c946c87d3a6d397504bf97a947e22d665592d8c7c2bff61f6ebe9b65769"} Nov 21 15:34:05 crc kubenswrapper[4897]: I1121 15:34:05.330307 4897 scope.go:117] "RemoveContainer" containerID="b31a4cad4f7e0bad138a5e80f7c1df147846cc303ce2ee9af988db3d48595701" Nov 21 15:34:05 crc kubenswrapper[4897]: I1121 15:34:05.329785 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v6j6v" Nov 21 15:34:05 crc kubenswrapper[4897]: I1121 15:34:05.377827 4897 scope.go:117] "RemoveContainer" containerID="036705e478c45896e7b67526aacba76928ac9a4c1c1e35414d0e0e4adfd6b38c" Nov 21 15:34:05 crc kubenswrapper[4897]: I1121 15:34:05.391849 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-v6j6v"] Nov 21 15:34:05 crc kubenswrapper[4897]: I1121 15:34:05.402266 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-v6j6v"] Nov 21 15:34:05 crc kubenswrapper[4897]: I1121 15:34:05.408031 4897 scope.go:117] "RemoveContainer" containerID="4324b18d2a4ec02cbdd25fbbeb1c9e09f031a08c590d07175f8c2a71944c5368" Nov 21 15:34:05 crc kubenswrapper[4897]: I1121 15:34:05.472952 4897 scope.go:117] "RemoveContainer" containerID="b31a4cad4f7e0bad138a5e80f7c1df147846cc303ce2ee9af988db3d48595701" Nov 21 15:34:05 crc kubenswrapper[4897]: E1121 15:34:05.473583 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b31a4cad4f7e0bad138a5e80f7c1df147846cc303ce2ee9af988db3d48595701\": container with ID starting with b31a4cad4f7e0bad138a5e80f7c1df147846cc303ce2ee9af988db3d48595701 not found: ID does not exist" containerID="b31a4cad4f7e0bad138a5e80f7c1df147846cc303ce2ee9af988db3d48595701" Nov 21 15:34:05 crc kubenswrapper[4897]: I1121 15:34:05.473625 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b31a4cad4f7e0bad138a5e80f7c1df147846cc303ce2ee9af988db3d48595701"} err="failed to get container status \"b31a4cad4f7e0bad138a5e80f7c1df147846cc303ce2ee9af988db3d48595701\": rpc error: code = NotFound desc = could not find container \"b31a4cad4f7e0bad138a5e80f7c1df147846cc303ce2ee9af988db3d48595701\": container with ID starting with b31a4cad4f7e0bad138a5e80f7c1df147846cc303ce2ee9af988db3d48595701 not found: ID does not exist" Nov 21 15:34:05 crc kubenswrapper[4897]: I1121 15:34:05.473648 4897 scope.go:117] "RemoveContainer" containerID="036705e478c45896e7b67526aacba76928ac9a4c1c1e35414d0e0e4adfd6b38c" Nov 21 15:34:05 crc kubenswrapper[4897]: E1121 15:34:05.474174 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"036705e478c45896e7b67526aacba76928ac9a4c1c1e35414d0e0e4adfd6b38c\": container with ID starting with 036705e478c45896e7b67526aacba76928ac9a4c1c1e35414d0e0e4adfd6b38c not found: ID does not exist" containerID="036705e478c45896e7b67526aacba76928ac9a4c1c1e35414d0e0e4adfd6b38c" Nov 21 15:34:05 crc kubenswrapper[4897]: I1121 15:34:05.474203 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"036705e478c45896e7b67526aacba76928ac9a4c1c1e35414d0e0e4adfd6b38c"} err="failed to get container status \"036705e478c45896e7b67526aacba76928ac9a4c1c1e35414d0e0e4adfd6b38c\": rpc error: code = NotFound desc = could not find container \"036705e478c45896e7b67526aacba76928ac9a4c1c1e35414d0e0e4adfd6b38c\": container with ID starting with 036705e478c45896e7b67526aacba76928ac9a4c1c1e35414d0e0e4adfd6b38c not found: ID does not exist" Nov 21 15:34:05 crc kubenswrapper[4897]: I1121 15:34:05.474221 4897 scope.go:117] "RemoveContainer" containerID="4324b18d2a4ec02cbdd25fbbeb1c9e09f031a08c590d07175f8c2a71944c5368" Nov 21 15:34:05 crc kubenswrapper[4897]: E1121 15:34:05.474688 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4324b18d2a4ec02cbdd25fbbeb1c9e09f031a08c590d07175f8c2a71944c5368\": container with ID starting with 4324b18d2a4ec02cbdd25fbbeb1c9e09f031a08c590d07175f8c2a71944c5368 not found: ID does not exist" containerID="4324b18d2a4ec02cbdd25fbbeb1c9e09f031a08c590d07175f8c2a71944c5368" Nov 21 15:34:05 crc kubenswrapper[4897]: I1121 15:34:05.474729 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4324b18d2a4ec02cbdd25fbbeb1c9e09f031a08c590d07175f8c2a71944c5368"} err="failed to get container status \"4324b18d2a4ec02cbdd25fbbeb1c9e09f031a08c590d07175f8c2a71944c5368\": rpc error: code = NotFound desc = could not find container \"4324b18d2a4ec02cbdd25fbbeb1c9e09f031a08c590d07175f8c2a71944c5368\": container with ID starting with 4324b18d2a4ec02cbdd25fbbeb1c9e09f031a08c590d07175f8c2a71944c5368 not found: ID does not exist" Nov 21 15:34:06 crc kubenswrapper[4897]: I1121 15:34:06.103459 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7b9b715c-1305-4968-b49f-488ce8c9688c" path="/var/lib/kubelet/pods/7b9b715c-1305-4968-b49f-488ce8c9688c/volumes" Nov 21 15:34:07 crc kubenswrapper[4897]: I1121 15:34:07.089464 4897 scope.go:117] "RemoveContainer" containerID="abf30c46658df7e750da25f3feeb4a134367a6c61c5b5221d6703974a1dd4d2e" Nov 21 15:34:07 crc kubenswrapper[4897]: E1121 15:34:07.090081 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:34:22 crc kubenswrapper[4897]: I1121 15:34:22.089776 4897 scope.go:117] "RemoveContainer" containerID="abf30c46658df7e750da25f3feeb4a134367a6c61c5b5221d6703974a1dd4d2e" Nov 21 15:34:22 crc kubenswrapper[4897]: E1121 15:34:22.092396 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:34:35 crc kubenswrapper[4897]: I1121 15:34:35.089152 4897 scope.go:117] "RemoveContainer" containerID="abf30c46658df7e750da25f3feeb4a134367a6c61c5b5221d6703974a1dd4d2e" Nov 21 15:34:35 crc kubenswrapper[4897]: E1121 15:34:35.090344 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:34:48 crc kubenswrapper[4897]: I1121 15:34:48.089634 4897 scope.go:117] "RemoveContainer" containerID="abf30c46658df7e750da25f3feeb4a134367a6c61c5b5221d6703974a1dd4d2e" Nov 21 15:34:48 crc kubenswrapper[4897]: E1121 15:34:48.090758 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:35:03 crc kubenswrapper[4897]: I1121 15:35:03.088890 4897 scope.go:117] "RemoveContainer" containerID="abf30c46658df7e750da25f3feeb4a134367a6c61c5b5221d6703974a1dd4d2e" Nov 21 15:35:03 crc kubenswrapper[4897]: E1121 15:35:03.089917 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:35:16 crc kubenswrapper[4897]: I1121 15:35:16.089187 4897 scope.go:117] "RemoveContainer" containerID="abf30c46658df7e750da25f3feeb4a134367a6c61c5b5221d6703974a1dd4d2e" Nov 21 15:35:16 crc kubenswrapper[4897]: E1121 15:35:16.090040 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:35:27 crc kubenswrapper[4897]: I1121 15:35:27.090708 4897 scope.go:117] "RemoveContainer" containerID="abf30c46658df7e750da25f3feeb4a134367a6c61c5b5221d6703974a1dd4d2e" Nov 21 15:35:27 crc kubenswrapper[4897]: E1121 15:35:27.091607 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:35:40 crc kubenswrapper[4897]: I1121 15:35:40.089030 4897 scope.go:117] "RemoveContainer" containerID="abf30c46658df7e750da25f3feeb4a134367a6c61c5b5221d6703974a1dd4d2e" Nov 21 15:35:40 crc kubenswrapper[4897]: E1121 15:35:40.089975 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:35:54 crc kubenswrapper[4897]: I1121 15:35:54.102320 4897 scope.go:117] "RemoveContainer" containerID="abf30c46658df7e750da25f3feeb4a134367a6c61c5b5221d6703974a1dd4d2e" Nov 21 15:35:54 crc kubenswrapper[4897]: E1121 15:35:54.103201 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:36:05 crc kubenswrapper[4897]: I1121 15:36:05.089437 4897 scope.go:117] "RemoveContainer" containerID="abf30c46658df7e750da25f3feeb4a134367a6c61c5b5221d6703974a1dd4d2e" Nov 21 15:36:05 crc kubenswrapper[4897]: E1121 15:36:05.090304 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:36:05 crc kubenswrapper[4897]: I1121 15:36:05.941255 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Nov 21 15:36:05 crc kubenswrapper[4897]: E1121 15:36:05.943064 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b9b715c-1305-4968-b49f-488ce8c9688c" containerName="extract-utilities" Nov 21 15:36:05 crc kubenswrapper[4897]: I1121 15:36:05.943086 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b9b715c-1305-4968-b49f-488ce8c9688c" containerName="extract-utilities" Nov 21 15:36:05 crc kubenswrapper[4897]: E1121 15:36:05.943132 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b9b715c-1305-4968-b49f-488ce8c9688c" containerName="extract-content" Nov 21 15:36:05 crc kubenswrapper[4897]: I1121 15:36:05.943145 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b9b715c-1305-4968-b49f-488ce8c9688c" containerName="extract-content" Nov 21 15:36:05 crc kubenswrapper[4897]: E1121 15:36:05.943167 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b9b715c-1305-4968-b49f-488ce8c9688c" containerName="registry-server" Nov 21 15:36:05 crc kubenswrapper[4897]: I1121 15:36:05.943176 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b9b715c-1305-4968-b49f-488ce8c9688c" containerName="registry-server" Nov 21 15:36:05 crc kubenswrapper[4897]: I1121 15:36:05.944059 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b9b715c-1305-4968-b49f-488ce8c9688c" containerName="registry-server" Nov 21 15:36:05 crc kubenswrapper[4897]: I1121 15:36:05.946033 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Nov 21 15:36:05 crc kubenswrapper[4897]: I1121 15:36:05.946198 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Nov 21 15:36:05 crc kubenswrapper[4897]: I1121 15:36:05.951293 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-8gssp" Nov 21 15:36:05 crc kubenswrapper[4897]: I1121 15:36:05.951740 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Nov 21 15:36:05 crc kubenswrapper[4897]: I1121 15:36:05.952131 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Nov 21 15:36:05 crc kubenswrapper[4897]: I1121 15:36:05.952420 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Nov 21 15:36:06 crc kubenswrapper[4897]: I1121 15:36:06.093962 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/b3a6d915-d988-45e8-a007-d135548ef81e-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"b3a6d915-d988-45e8-a007-d135548ef81e\") " pod="openstack/tempest-tests-tempest" Nov 21 15:36:06 crc kubenswrapper[4897]: I1121 15:36:06.094003 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/b3a6d915-d988-45e8-a007-d135548ef81e-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"b3a6d915-d988-45e8-a007-d135548ef81e\") " pod="openstack/tempest-tests-tempest" Nov 21 15:36:06 crc kubenswrapper[4897]: I1121 15:36:06.094213 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"tempest-tests-tempest\" (UID: \"b3a6d915-d988-45e8-a007-d135548ef81e\") " pod="openstack/tempest-tests-tempest" Nov 21 15:36:06 crc kubenswrapper[4897]: I1121 15:36:06.094317 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/b3a6d915-d988-45e8-a007-d135548ef81e-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"b3a6d915-d988-45e8-a007-d135548ef81e\") " pod="openstack/tempest-tests-tempest" Nov 21 15:36:06 crc kubenswrapper[4897]: I1121 15:36:06.094553 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b3a6d915-d988-45e8-a007-d135548ef81e-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"b3a6d915-d988-45e8-a007-d135548ef81e\") " pod="openstack/tempest-tests-tempest" Nov 21 15:36:06 crc kubenswrapper[4897]: I1121 15:36:06.094715 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b3a6d915-d988-45e8-a007-d135548ef81e-config-data\") pod \"tempest-tests-tempest\" (UID: \"b3a6d915-d988-45e8-a007-d135548ef81e\") " pod="openstack/tempest-tests-tempest" Nov 21 15:36:06 crc kubenswrapper[4897]: I1121 15:36:06.094789 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4b626\" (UniqueName: \"kubernetes.io/projected/b3a6d915-d988-45e8-a007-d135548ef81e-kube-api-access-4b626\") pod \"tempest-tests-tempest\" (UID: \"b3a6d915-d988-45e8-a007-d135548ef81e\") " pod="openstack/tempest-tests-tempest" Nov 21 15:36:06 crc kubenswrapper[4897]: I1121 15:36:06.094885 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/b3a6d915-d988-45e8-a007-d135548ef81e-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"b3a6d915-d988-45e8-a007-d135548ef81e\") " pod="openstack/tempest-tests-tempest" Nov 21 15:36:06 crc kubenswrapper[4897]: I1121 15:36:06.095042 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/b3a6d915-d988-45e8-a007-d135548ef81e-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"b3a6d915-d988-45e8-a007-d135548ef81e\") " pod="openstack/tempest-tests-tempest" Nov 21 15:36:06 crc kubenswrapper[4897]: I1121 15:36:06.197639 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/b3a6d915-d988-45e8-a007-d135548ef81e-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"b3a6d915-d988-45e8-a007-d135548ef81e\") " pod="openstack/tempest-tests-tempest" Nov 21 15:36:06 crc kubenswrapper[4897]: I1121 15:36:06.197721 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/b3a6d915-d988-45e8-a007-d135548ef81e-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"b3a6d915-d988-45e8-a007-d135548ef81e\") " pod="openstack/tempest-tests-tempest" Nov 21 15:36:06 crc kubenswrapper[4897]: I1121 15:36:06.197786 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"tempest-tests-tempest\" (UID: \"b3a6d915-d988-45e8-a007-d135548ef81e\") " pod="openstack/tempest-tests-tempest" Nov 21 15:36:06 crc kubenswrapper[4897]: I1121 15:36:06.197816 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/b3a6d915-d988-45e8-a007-d135548ef81e-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"b3a6d915-d988-45e8-a007-d135548ef81e\") " pod="openstack/tempest-tests-tempest" Nov 21 15:36:06 crc kubenswrapper[4897]: I1121 15:36:06.197890 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b3a6d915-d988-45e8-a007-d135548ef81e-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"b3a6d915-d988-45e8-a007-d135548ef81e\") " pod="openstack/tempest-tests-tempest" Nov 21 15:36:06 crc kubenswrapper[4897]: I1121 15:36:06.197946 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b3a6d915-d988-45e8-a007-d135548ef81e-config-data\") pod \"tempest-tests-tempest\" (UID: \"b3a6d915-d988-45e8-a007-d135548ef81e\") " pod="openstack/tempest-tests-tempest" Nov 21 15:36:06 crc kubenswrapper[4897]: I1121 15:36:06.197976 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4b626\" (UniqueName: \"kubernetes.io/projected/b3a6d915-d988-45e8-a007-d135548ef81e-kube-api-access-4b626\") pod \"tempest-tests-tempest\" (UID: \"b3a6d915-d988-45e8-a007-d135548ef81e\") " pod="openstack/tempest-tests-tempest" Nov 21 15:36:06 crc kubenswrapper[4897]: I1121 15:36:06.198014 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/b3a6d915-d988-45e8-a007-d135548ef81e-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"b3a6d915-d988-45e8-a007-d135548ef81e\") " pod="openstack/tempest-tests-tempest" Nov 21 15:36:06 crc kubenswrapper[4897]: I1121 15:36:06.198087 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/b3a6d915-d988-45e8-a007-d135548ef81e-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"b3a6d915-d988-45e8-a007-d135548ef81e\") " pod="openstack/tempest-tests-tempest" Nov 21 15:36:06 crc kubenswrapper[4897]: I1121 15:36:06.199079 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/b3a6d915-d988-45e8-a007-d135548ef81e-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"b3a6d915-d988-45e8-a007-d135548ef81e\") " pod="openstack/tempest-tests-tempest" Nov 21 15:36:06 crc kubenswrapper[4897]: I1121 15:36:06.199270 4897 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"tempest-tests-tempest\" (UID: \"b3a6d915-d988-45e8-a007-d135548ef81e\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/tempest-tests-tempest" Nov 21 15:36:06 crc kubenswrapper[4897]: I1121 15:36:06.199733 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/b3a6d915-d988-45e8-a007-d135548ef81e-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"b3a6d915-d988-45e8-a007-d135548ef81e\") " pod="openstack/tempest-tests-tempest" Nov 21 15:36:06 crc kubenswrapper[4897]: I1121 15:36:06.200287 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b3a6d915-d988-45e8-a007-d135548ef81e-config-data\") pod \"tempest-tests-tempest\" (UID: \"b3a6d915-d988-45e8-a007-d135548ef81e\") " pod="openstack/tempest-tests-tempest" Nov 21 15:36:06 crc kubenswrapper[4897]: I1121 15:36:06.200535 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/b3a6d915-d988-45e8-a007-d135548ef81e-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"b3a6d915-d988-45e8-a007-d135548ef81e\") " pod="openstack/tempest-tests-tempest" Nov 21 15:36:06 crc kubenswrapper[4897]: I1121 15:36:06.204495 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/b3a6d915-d988-45e8-a007-d135548ef81e-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"b3a6d915-d988-45e8-a007-d135548ef81e\") " pod="openstack/tempest-tests-tempest" Nov 21 15:36:06 crc kubenswrapper[4897]: I1121 15:36:06.216377 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b3a6d915-d988-45e8-a007-d135548ef81e-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"b3a6d915-d988-45e8-a007-d135548ef81e\") " pod="openstack/tempest-tests-tempest" Nov 21 15:36:06 crc kubenswrapper[4897]: I1121 15:36:06.217099 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/b3a6d915-d988-45e8-a007-d135548ef81e-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"b3a6d915-d988-45e8-a007-d135548ef81e\") " pod="openstack/tempest-tests-tempest" Nov 21 15:36:06 crc kubenswrapper[4897]: I1121 15:36:06.220219 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4b626\" (UniqueName: \"kubernetes.io/projected/b3a6d915-d988-45e8-a007-d135548ef81e-kube-api-access-4b626\") pod \"tempest-tests-tempest\" (UID: \"b3a6d915-d988-45e8-a007-d135548ef81e\") " pod="openstack/tempest-tests-tempest" Nov 21 15:36:06 crc kubenswrapper[4897]: I1121 15:36:06.234424 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"tempest-tests-tempest\" (UID: \"b3a6d915-d988-45e8-a007-d135548ef81e\") " pod="openstack/tempest-tests-tempest" Nov 21 15:36:06 crc kubenswrapper[4897]: I1121 15:36:06.270621 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Nov 21 15:36:06 crc kubenswrapper[4897]: I1121 15:36:06.733309 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Nov 21 15:36:06 crc kubenswrapper[4897]: I1121 15:36:06.733583 4897 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 21 15:36:07 crc kubenswrapper[4897]: I1121 15:36:07.655323 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"b3a6d915-d988-45e8-a007-d135548ef81e","Type":"ContainerStarted","Data":"8c6ceb51ed1cd6a4980a2c051cfafeab46e377ccd32e396d02240218c1d4c841"} Nov 21 15:36:18 crc kubenswrapper[4897]: I1121 15:36:18.089151 4897 scope.go:117] "RemoveContainer" containerID="abf30c46658df7e750da25f3feeb4a134367a6c61c5b5221d6703974a1dd4d2e" Nov 21 15:36:18 crc kubenswrapper[4897]: E1121 15:36:18.090084 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:36:23 crc kubenswrapper[4897]: I1121 15:36:23.574479 4897 scope.go:117] "RemoveContainer" containerID="75ccd25815f0b39c2b26a41228c4e231b7c3345ed2d38c4157327c4e0eb1f0f2" Nov 21 15:36:29 crc kubenswrapper[4897]: I1121 15:36:29.089635 4897 scope.go:117] "RemoveContainer" containerID="abf30c46658df7e750da25f3feeb4a134367a6c61c5b5221d6703974a1dd4d2e" Nov 21 15:36:29 crc kubenswrapper[4897]: E1121 15:36:29.090756 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:36:43 crc kubenswrapper[4897]: I1121 15:36:43.406306 4897 scope.go:117] "RemoveContainer" containerID="0d91752aeaed8f1597a4e93218c0e815e67f9d1f3e450156fd9a2a5fdc3079f9" Nov 21 15:36:43 crc kubenswrapper[4897]: E1121 15:36:43.479850 4897 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Nov 21 15:36:43 crc kubenswrapper[4897]: E1121 15:36:43.480236 4897 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4b626,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(b3a6d915-d988-45e8-a007-d135548ef81e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 21 15:36:43 crc kubenswrapper[4897]: E1121 15:36:43.482472 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="b3a6d915-d988-45e8-a007-d135548ef81e" Nov 21 15:36:44 crc kubenswrapper[4897]: E1121 15:36:44.082194 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="b3a6d915-d988-45e8-a007-d135548ef81e" Nov 21 15:36:44 crc kubenswrapper[4897]: I1121 15:36:44.088578 4897 scope.go:117] "RemoveContainer" containerID="abf30c46658df7e750da25f3feeb4a134367a6c61c5b5221d6703974a1dd4d2e" Nov 21 15:36:44 crc kubenswrapper[4897]: E1121 15:36:44.089048 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:36:57 crc kubenswrapper[4897]: I1121 15:36:57.595167 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Nov 21 15:36:58 crc kubenswrapper[4897]: I1121 15:36:58.090000 4897 scope.go:117] "RemoveContainer" containerID="abf30c46658df7e750da25f3feeb4a134367a6c61c5b5221d6703974a1dd4d2e" Nov 21 15:36:58 crc kubenswrapper[4897]: E1121 15:36:58.090580 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:36:59 crc kubenswrapper[4897]: I1121 15:36:59.258670 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"b3a6d915-d988-45e8-a007-d135548ef81e","Type":"ContainerStarted","Data":"b8a02b6161e216b0fbfd9fc93366f53197a472ac54762a52ff31a8d73a41a20e"} Nov 21 15:36:59 crc kubenswrapper[4897]: I1121 15:36:59.287337 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=4.427588813 podStartE2EDuration="55.287317963s" podCreationTimestamp="2025-11-21 15:36:04 +0000 UTC" firstStartedPulling="2025-11-21 15:36:06.733155757 +0000 UTC m=+5244.017749232" lastFinishedPulling="2025-11-21 15:36:57.592884907 +0000 UTC m=+5294.877478382" observedRunningTime="2025-11-21 15:36:59.283059308 +0000 UTC m=+5296.567652783" watchObservedRunningTime="2025-11-21 15:36:59.287317963 +0000 UTC m=+5296.571911438" Nov 21 15:37:11 crc kubenswrapper[4897]: I1121 15:37:11.089531 4897 scope.go:117] "RemoveContainer" containerID="abf30c46658df7e750da25f3feeb4a134367a6c61c5b5221d6703974a1dd4d2e" Nov 21 15:37:11 crc kubenswrapper[4897]: E1121 15:37:11.090466 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:37:25 crc kubenswrapper[4897]: I1121 15:37:25.089944 4897 scope.go:117] "RemoveContainer" containerID="abf30c46658df7e750da25f3feeb4a134367a6c61c5b5221d6703974a1dd4d2e" Nov 21 15:37:25 crc kubenswrapper[4897]: E1121 15:37:25.091103 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:37:39 crc kubenswrapper[4897]: I1121 15:37:39.089822 4897 scope.go:117] "RemoveContainer" containerID="abf30c46658df7e750da25f3feeb4a134367a6c61c5b5221d6703974a1dd4d2e" Nov 21 15:37:39 crc kubenswrapper[4897]: E1121 15:37:39.091694 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:37:43 crc kubenswrapper[4897]: I1121 15:37:43.498477 4897 scope.go:117] "RemoveContainer" containerID="dd37f145a4fd7c1393bab3fc534103bebc9ee7b44dad417e1f638d1f0a79455b" Nov 21 15:37:51 crc kubenswrapper[4897]: I1121 15:37:51.092707 4897 scope.go:117] "RemoveContainer" containerID="abf30c46658df7e750da25f3feeb4a134367a6c61c5b5221d6703974a1dd4d2e" Nov 21 15:37:51 crc kubenswrapper[4897]: E1121 15:37:51.094242 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:38:05 crc kubenswrapper[4897]: I1121 15:38:05.089952 4897 scope.go:117] "RemoveContainer" containerID="abf30c46658df7e750da25f3feeb4a134367a6c61c5b5221d6703974a1dd4d2e" Nov 21 15:38:06 crc kubenswrapper[4897]: I1121 15:38:06.024300 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" event={"ID":"e7670227-d280-4847-b882-754429f56b0f","Type":"ContainerStarted","Data":"b75f9d31a9263f80ca2b82ff3c2e842f5c5ccc5e8cefc8a57960d6be99c63908"} Nov 21 15:40:07 crc kubenswrapper[4897]: I1121 15:40:07.050178 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-sx7s8"] Nov 21 15:40:07 crc kubenswrapper[4897]: I1121 15:40:07.065483 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sx7s8" Nov 21 15:40:07 crc kubenswrapper[4897]: I1121 15:40:07.072328 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-sx7s8"] Nov 21 15:40:07 crc kubenswrapper[4897]: I1121 15:40:07.165801 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lmdgc\" (UniqueName: \"kubernetes.io/projected/b7401e3a-fb83-453c-8d45-e2efb34c1ed8-kube-api-access-lmdgc\") pod \"redhat-operators-sx7s8\" (UID: \"b7401e3a-fb83-453c-8d45-e2efb34c1ed8\") " pod="openshift-marketplace/redhat-operators-sx7s8" Nov 21 15:40:07 crc kubenswrapper[4897]: I1121 15:40:07.166020 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7401e3a-fb83-453c-8d45-e2efb34c1ed8-utilities\") pod \"redhat-operators-sx7s8\" (UID: \"b7401e3a-fb83-453c-8d45-e2efb34c1ed8\") " pod="openshift-marketplace/redhat-operators-sx7s8" Nov 21 15:40:07 crc kubenswrapper[4897]: I1121 15:40:07.166089 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7401e3a-fb83-453c-8d45-e2efb34c1ed8-catalog-content\") pod \"redhat-operators-sx7s8\" (UID: \"b7401e3a-fb83-453c-8d45-e2efb34c1ed8\") " pod="openshift-marketplace/redhat-operators-sx7s8" Nov 21 15:40:07 crc kubenswrapper[4897]: I1121 15:40:07.268208 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lmdgc\" (UniqueName: \"kubernetes.io/projected/b7401e3a-fb83-453c-8d45-e2efb34c1ed8-kube-api-access-lmdgc\") pod \"redhat-operators-sx7s8\" (UID: \"b7401e3a-fb83-453c-8d45-e2efb34c1ed8\") " pod="openshift-marketplace/redhat-operators-sx7s8" Nov 21 15:40:07 crc kubenswrapper[4897]: I1121 15:40:07.268359 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7401e3a-fb83-453c-8d45-e2efb34c1ed8-utilities\") pod \"redhat-operators-sx7s8\" (UID: \"b7401e3a-fb83-453c-8d45-e2efb34c1ed8\") " pod="openshift-marketplace/redhat-operators-sx7s8" Nov 21 15:40:07 crc kubenswrapper[4897]: I1121 15:40:07.268432 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7401e3a-fb83-453c-8d45-e2efb34c1ed8-catalog-content\") pod \"redhat-operators-sx7s8\" (UID: \"b7401e3a-fb83-453c-8d45-e2efb34c1ed8\") " pod="openshift-marketplace/redhat-operators-sx7s8" Nov 21 15:40:07 crc kubenswrapper[4897]: I1121 15:40:07.271789 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7401e3a-fb83-453c-8d45-e2efb34c1ed8-utilities\") pod \"redhat-operators-sx7s8\" (UID: \"b7401e3a-fb83-453c-8d45-e2efb34c1ed8\") " pod="openshift-marketplace/redhat-operators-sx7s8" Nov 21 15:40:07 crc kubenswrapper[4897]: I1121 15:40:07.271864 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7401e3a-fb83-453c-8d45-e2efb34c1ed8-catalog-content\") pod \"redhat-operators-sx7s8\" (UID: \"b7401e3a-fb83-453c-8d45-e2efb34c1ed8\") " pod="openshift-marketplace/redhat-operators-sx7s8" Nov 21 15:40:07 crc kubenswrapper[4897]: I1121 15:40:07.308578 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lmdgc\" (UniqueName: \"kubernetes.io/projected/b7401e3a-fb83-453c-8d45-e2efb34c1ed8-kube-api-access-lmdgc\") pod \"redhat-operators-sx7s8\" (UID: \"b7401e3a-fb83-453c-8d45-e2efb34c1ed8\") " pod="openshift-marketplace/redhat-operators-sx7s8" Nov 21 15:40:07 crc kubenswrapper[4897]: I1121 15:40:07.391741 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sx7s8" Nov 21 15:40:08 crc kubenswrapper[4897]: I1121 15:40:08.772403 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-sx7s8"] Nov 21 15:40:09 crc kubenswrapper[4897]: I1121 15:40:09.514338 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sx7s8" event={"ID":"b7401e3a-fb83-453c-8d45-e2efb34c1ed8","Type":"ContainerDied","Data":"c3063e4d1319bd1064fd3400965a617a512f3bdc7748b610bf9f289c17d3c7e6"} Nov 21 15:40:09 crc kubenswrapper[4897]: I1121 15:40:09.515316 4897 generic.go:334] "Generic (PLEG): container finished" podID="b7401e3a-fb83-453c-8d45-e2efb34c1ed8" containerID="c3063e4d1319bd1064fd3400965a617a512f3bdc7748b610bf9f289c17d3c7e6" exitCode=0 Nov 21 15:40:09 crc kubenswrapper[4897]: I1121 15:40:09.515464 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sx7s8" event={"ID":"b7401e3a-fb83-453c-8d45-e2efb34c1ed8","Type":"ContainerStarted","Data":"a1e7a2c121e0e3164812bcc091317673c7eb90841dda2ed6a77c5c8777686f8e"} Nov 21 15:40:10 crc kubenswrapper[4897]: I1121 15:40:10.538348 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sx7s8" event={"ID":"b7401e3a-fb83-453c-8d45-e2efb34c1ed8","Type":"ContainerStarted","Data":"10ad7d280923a05bee6368e831b8cefa05f89e75fb7e07e94ed29b703252bd97"} Nov 21 15:40:15 crc kubenswrapper[4897]: I1121 15:40:15.602802 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sx7s8" event={"ID":"b7401e3a-fb83-453c-8d45-e2efb34c1ed8","Type":"ContainerDied","Data":"10ad7d280923a05bee6368e831b8cefa05f89e75fb7e07e94ed29b703252bd97"} Nov 21 15:40:15 crc kubenswrapper[4897]: I1121 15:40:15.602762 4897 generic.go:334] "Generic (PLEG): container finished" podID="b7401e3a-fb83-453c-8d45-e2efb34c1ed8" containerID="10ad7d280923a05bee6368e831b8cefa05f89e75fb7e07e94ed29b703252bd97" exitCode=0 Nov 21 15:40:16 crc kubenswrapper[4897]: I1121 15:40:16.628414 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sx7s8" event={"ID":"b7401e3a-fb83-453c-8d45-e2efb34c1ed8","Type":"ContainerStarted","Data":"f2872cb69bf99e2310df6dc29eb567d214cb5d0e7e6c0d9c6247b72a36905c39"} Nov 21 15:40:16 crc kubenswrapper[4897]: I1121 15:40:16.674035 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-sx7s8" podStartSLOduration=4.146513805 podStartE2EDuration="10.672372165s" podCreationTimestamp="2025-11-21 15:40:06 +0000 UTC" firstStartedPulling="2025-11-21 15:40:09.516872011 +0000 UTC m=+5486.801465486" lastFinishedPulling="2025-11-21 15:40:16.042730371 +0000 UTC m=+5493.327323846" observedRunningTime="2025-11-21 15:40:16.655929511 +0000 UTC m=+5493.940522986" watchObservedRunningTime="2025-11-21 15:40:16.672372165 +0000 UTC m=+5493.956965640" Nov 21 15:40:17 crc kubenswrapper[4897]: I1121 15:40:17.392466 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-sx7s8" Nov 21 15:40:17 crc kubenswrapper[4897]: I1121 15:40:17.392763 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-sx7s8" Nov 21 15:40:19 crc kubenswrapper[4897]: I1121 15:40:19.078578 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-sx7s8" podUID="b7401e3a-fb83-453c-8d45-e2efb34c1ed8" containerName="registry-server" probeResult="failure" output=< Nov 21 15:40:19 crc kubenswrapper[4897]: timeout: failed to connect service ":50051" within 1s Nov 21 15:40:19 crc kubenswrapper[4897]: > Nov 21 15:40:28 crc kubenswrapper[4897]: I1121 15:40:28.447118 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-sx7s8" podUID="b7401e3a-fb83-453c-8d45-e2efb34c1ed8" containerName="registry-server" probeResult="failure" output=< Nov 21 15:40:28 crc kubenswrapper[4897]: timeout: failed to connect service ":50051" within 1s Nov 21 15:40:28 crc kubenswrapper[4897]: > Nov 21 15:40:34 crc kubenswrapper[4897]: I1121 15:40:34.370905 4897 patch_prober.go:28] interesting pod/machine-config-daemon-krv5b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 21 15:40:34 crc kubenswrapper[4897]: I1121 15:40:34.372666 4897 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 21 15:40:38 crc kubenswrapper[4897]: I1121 15:40:38.663787 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-sx7s8" podUID="b7401e3a-fb83-453c-8d45-e2efb34c1ed8" containerName="registry-server" probeResult="failure" output=< Nov 21 15:40:38 crc kubenswrapper[4897]: timeout: failed to connect service ":50051" within 1s Nov 21 15:40:38 crc kubenswrapper[4897]: > Nov 21 15:40:48 crc kubenswrapper[4897]: I1121 15:40:48.448112 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-sx7s8" podUID="b7401e3a-fb83-453c-8d45-e2efb34c1ed8" containerName="registry-server" probeResult="failure" output=< Nov 21 15:40:48 crc kubenswrapper[4897]: timeout: failed to connect service ":50051" within 1s Nov 21 15:40:48 crc kubenswrapper[4897]: > Nov 21 15:40:52 crc kubenswrapper[4897]: I1121 15:40:52.268181 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-xdmc4"] Nov 21 15:40:52 crc kubenswrapper[4897]: I1121 15:40:52.277929 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xdmc4" Nov 21 15:40:52 crc kubenswrapper[4897]: I1121 15:40:52.282401 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xdmc4"] Nov 21 15:40:52 crc kubenswrapper[4897]: I1121 15:40:52.369925 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5da2270-16fa-4d91-9110-ac7305326b10-catalog-content\") pod \"redhat-marketplace-xdmc4\" (UID: \"f5da2270-16fa-4d91-9110-ac7305326b10\") " pod="openshift-marketplace/redhat-marketplace-xdmc4" Nov 21 15:40:52 crc kubenswrapper[4897]: I1121 15:40:52.370208 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qqxtd\" (UniqueName: \"kubernetes.io/projected/f5da2270-16fa-4d91-9110-ac7305326b10-kube-api-access-qqxtd\") pod \"redhat-marketplace-xdmc4\" (UID: \"f5da2270-16fa-4d91-9110-ac7305326b10\") " pod="openshift-marketplace/redhat-marketplace-xdmc4" Nov 21 15:40:52 crc kubenswrapper[4897]: I1121 15:40:52.370336 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5da2270-16fa-4d91-9110-ac7305326b10-utilities\") pod \"redhat-marketplace-xdmc4\" (UID: \"f5da2270-16fa-4d91-9110-ac7305326b10\") " pod="openshift-marketplace/redhat-marketplace-xdmc4" Nov 21 15:40:52 crc kubenswrapper[4897]: I1121 15:40:52.472818 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5da2270-16fa-4d91-9110-ac7305326b10-catalog-content\") pod \"redhat-marketplace-xdmc4\" (UID: \"f5da2270-16fa-4d91-9110-ac7305326b10\") " pod="openshift-marketplace/redhat-marketplace-xdmc4" Nov 21 15:40:52 crc kubenswrapper[4897]: I1121 15:40:52.472945 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qqxtd\" (UniqueName: \"kubernetes.io/projected/f5da2270-16fa-4d91-9110-ac7305326b10-kube-api-access-qqxtd\") pod \"redhat-marketplace-xdmc4\" (UID: \"f5da2270-16fa-4d91-9110-ac7305326b10\") " pod="openshift-marketplace/redhat-marketplace-xdmc4" Nov 21 15:40:52 crc kubenswrapper[4897]: I1121 15:40:52.472996 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5da2270-16fa-4d91-9110-ac7305326b10-utilities\") pod \"redhat-marketplace-xdmc4\" (UID: \"f5da2270-16fa-4d91-9110-ac7305326b10\") " pod="openshift-marketplace/redhat-marketplace-xdmc4" Nov 21 15:40:52 crc kubenswrapper[4897]: I1121 15:40:52.473596 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5da2270-16fa-4d91-9110-ac7305326b10-catalog-content\") pod \"redhat-marketplace-xdmc4\" (UID: \"f5da2270-16fa-4d91-9110-ac7305326b10\") " pod="openshift-marketplace/redhat-marketplace-xdmc4" Nov 21 15:40:52 crc kubenswrapper[4897]: I1121 15:40:52.474908 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5da2270-16fa-4d91-9110-ac7305326b10-utilities\") pod \"redhat-marketplace-xdmc4\" (UID: \"f5da2270-16fa-4d91-9110-ac7305326b10\") " pod="openshift-marketplace/redhat-marketplace-xdmc4" Nov 21 15:40:52 crc kubenswrapper[4897]: I1121 15:40:52.500329 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qqxtd\" (UniqueName: \"kubernetes.io/projected/f5da2270-16fa-4d91-9110-ac7305326b10-kube-api-access-qqxtd\") pod \"redhat-marketplace-xdmc4\" (UID: \"f5da2270-16fa-4d91-9110-ac7305326b10\") " pod="openshift-marketplace/redhat-marketplace-xdmc4" Nov 21 15:40:52 crc kubenswrapper[4897]: I1121 15:40:52.601176 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xdmc4" Nov 21 15:40:53 crc kubenswrapper[4897]: I1121 15:40:53.097641 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xdmc4"] Nov 21 15:40:54 crc kubenswrapper[4897]: I1121 15:40:54.018173 4897 generic.go:334] "Generic (PLEG): container finished" podID="f5da2270-16fa-4d91-9110-ac7305326b10" containerID="65bf848c64ac7cd813802d0c5ec47a5efe05c6560592cb3cf0778d70b56681a1" exitCode=0 Nov 21 15:40:54 crc kubenswrapper[4897]: I1121 15:40:54.018368 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xdmc4" event={"ID":"f5da2270-16fa-4d91-9110-ac7305326b10","Type":"ContainerDied","Data":"65bf848c64ac7cd813802d0c5ec47a5efe05c6560592cb3cf0778d70b56681a1"} Nov 21 15:40:54 crc kubenswrapper[4897]: I1121 15:40:54.018520 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xdmc4" event={"ID":"f5da2270-16fa-4d91-9110-ac7305326b10","Type":"ContainerStarted","Data":"8f50af3f0bf4bbdefedb1336a2d0809b5ead6e3a1bfdbd74d21e98e4e677d06b"} Nov 21 15:40:56 crc kubenswrapper[4897]: I1121 15:40:56.045868 4897 generic.go:334] "Generic (PLEG): container finished" podID="f5da2270-16fa-4d91-9110-ac7305326b10" containerID="be29704301c9ca39234af62a77939773dbec2304201eb6789c9b1c467eeaf702" exitCode=0 Nov 21 15:40:56 crc kubenswrapper[4897]: I1121 15:40:56.045953 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xdmc4" event={"ID":"f5da2270-16fa-4d91-9110-ac7305326b10","Type":"ContainerDied","Data":"be29704301c9ca39234af62a77939773dbec2304201eb6789c9b1c467eeaf702"} Nov 21 15:40:57 crc kubenswrapper[4897]: I1121 15:40:57.061066 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xdmc4" event={"ID":"f5da2270-16fa-4d91-9110-ac7305326b10","Type":"ContainerStarted","Data":"90f3960cc8df6087ca42249921bcf81515300580c58eb2cb8a3b0adb482f6055"} Nov 21 15:40:57 crc kubenswrapper[4897]: I1121 15:40:57.095058 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-xdmc4" podStartSLOduration=2.66102161 podStartE2EDuration="5.09504104s" podCreationTimestamp="2025-11-21 15:40:52 +0000 UTC" firstStartedPulling="2025-11-21 15:40:54.022897417 +0000 UTC m=+5531.307490892" lastFinishedPulling="2025-11-21 15:40:56.456916847 +0000 UTC m=+5533.741510322" observedRunningTime="2025-11-21 15:40:57.091772062 +0000 UTC m=+5534.376365557" watchObservedRunningTime="2025-11-21 15:40:57.09504104 +0000 UTC m=+5534.379634515" Nov 21 15:40:57 crc kubenswrapper[4897]: I1121 15:40:57.447676 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-sx7s8" Nov 21 15:40:57 crc kubenswrapper[4897]: I1121 15:40:57.509979 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-sx7s8" Nov 21 15:40:58 crc kubenswrapper[4897]: I1121 15:40:58.638810 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-sx7s8"] Nov 21 15:40:59 crc kubenswrapper[4897]: I1121 15:40:59.081223 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-sx7s8" podUID="b7401e3a-fb83-453c-8d45-e2efb34c1ed8" containerName="registry-server" containerID="cri-o://f2872cb69bf99e2310df6dc29eb567d214cb5d0e7e6c0d9c6247b72a36905c39" gracePeriod=2 Nov 21 15:41:00 crc kubenswrapper[4897]: I1121 15:41:00.103853 4897 generic.go:334] "Generic (PLEG): container finished" podID="b7401e3a-fb83-453c-8d45-e2efb34c1ed8" containerID="f2872cb69bf99e2310df6dc29eb567d214cb5d0e7e6c0d9c6247b72a36905c39" exitCode=0 Nov 21 15:41:00 crc kubenswrapper[4897]: I1121 15:41:00.106246 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sx7s8" event={"ID":"b7401e3a-fb83-453c-8d45-e2efb34c1ed8","Type":"ContainerDied","Data":"f2872cb69bf99e2310df6dc29eb567d214cb5d0e7e6c0d9c6247b72a36905c39"} Nov 21 15:41:00 crc kubenswrapper[4897]: I1121 15:41:00.106339 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sx7s8" event={"ID":"b7401e3a-fb83-453c-8d45-e2efb34c1ed8","Type":"ContainerDied","Data":"a1e7a2c121e0e3164812bcc091317673c7eb90841dda2ed6a77c5c8777686f8e"} Nov 21 15:41:00 crc kubenswrapper[4897]: I1121 15:41:00.106824 4897 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a1e7a2c121e0e3164812bcc091317673c7eb90841dda2ed6a77c5c8777686f8e" Nov 21 15:41:00 crc kubenswrapper[4897]: I1121 15:41:00.483732 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sx7s8" Nov 21 15:41:00 crc kubenswrapper[4897]: I1121 15:41:00.566289 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7401e3a-fb83-453c-8d45-e2efb34c1ed8-catalog-content\") pod \"b7401e3a-fb83-453c-8d45-e2efb34c1ed8\" (UID: \"b7401e3a-fb83-453c-8d45-e2efb34c1ed8\") " Nov 21 15:41:00 crc kubenswrapper[4897]: I1121 15:41:00.566610 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lmdgc\" (UniqueName: \"kubernetes.io/projected/b7401e3a-fb83-453c-8d45-e2efb34c1ed8-kube-api-access-lmdgc\") pod \"b7401e3a-fb83-453c-8d45-e2efb34c1ed8\" (UID: \"b7401e3a-fb83-453c-8d45-e2efb34c1ed8\") " Nov 21 15:41:00 crc kubenswrapper[4897]: I1121 15:41:00.566709 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7401e3a-fb83-453c-8d45-e2efb34c1ed8-utilities\") pod \"b7401e3a-fb83-453c-8d45-e2efb34c1ed8\" (UID: \"b7401e3a-fb83-453c-8d45-e2efb34c1ed8\") " Nov 21 15:41:00 crc kubenswrapper[4897]: I1121 15:41:00.569745 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b7401e3a-fb83-453c-8d45-e2efb34c1ed8-utilities" (OuterVolumeSpecName: "utilities") pod "b7401e3a-fb83-453c-8d45-e2efb34c1ed8" (UID: "b7401e3a-fb83-453c-8d45-e2efb34c1ed8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 15:41:00 crc kubenswrapper[4897]: I1121 15:41:00.619001 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7401e3a-fb83-453c-8d45-e2efb34c1ed8-kube-api-access-lmdgc" (OuterVolumeSpecName: "kube-api-access-lmdgc") pod "b7401e3a-fb83-453c-8d45-e2efb34c1ed8" (UID: "b7401e3a-fb83-453c-8d45-e2efb34c1ed8"). InnerVolumeSpecName "kube-api-access-lmdgc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 15:41:00 crc kubenswrapper[4897]: I1121 15:41:00.670784 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lmdgc\" (UniqueName: \"kubernetes.io/projected/b7401e3a-fb83-453c-8d45-e2efb34c1ed8-kube-api-access-lmdgc\") on node \"crc\" DevicePath \"\"" Nov 21 15:41:00 crc kubenswrapper[4897]: I1121 15:41:00.670820 4897 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7401e3a-fb83-453c-8d45-e2efb34c1ed8-utilities\") on node \"crc\" DevicePath \"\"" Nov 21 15:41:00 crc kubenswrapper[4897]: I1121 15:41:00.689227 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b7401e3a-fb83-453c-8d45-e2efb34c1ed8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b7401e3a-fb83-453c-8d45-e2efb34c1ed8" (UID: "b7401e3a-fb83-453c-8d45-e2efb34c1ed8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 15:41:00 crc kubenswrapper[4897]: I1121 15:41:00.773173 4897 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7401e3a-fb83-453c-8d45-e2efb34c1ed8-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 21 15:41:01 crc kubenswrapper[4897]: I1121 15:41:01.114984 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sx7s8" Nov 21 15:41:01 crc kubenswrapper[4897]: I1121 15:41:01.167886 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-sx7s8"] Nov 21 15:41:01 crc kubenswrapper[4897]: I1121 15:41:01.181337 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-sx7s8"] Nov 21 15:41:02 crc kubenswrapper[4897]: I1121 15:41:02.102023 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b7401e3a-fb83-453c-8d45-e2efb34c1ed8" path="/var/lib/kubelet/pods/b7401e3a-fb83-453c-8d45-e2efb34c1ed8/volumes" Nov 21 15:41:02 crc kubenswrapper[4897]: I1121 15:41:02.601683 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-xdmc4" Nov 21 15:41:02 crc kubenswrapper[4897]: I1121 15:41:02.602114 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-xdmc4" Nov 21 15:41:02 crc kubenswrapper[4897]: I1121 15:41:02.661781 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-xdmc4" Nov 21 15:41:03 crc kubenswrapper[4897]: I1121 15:41:03.191136 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-xdmc4" Nov 21 15:41:04 crc kubenswrapper[4897]: I1121 15:41:04.041340 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xdmc4"] Nov 21 15:41:04 crc kubenswrapper[4897]: I1121 15:41:04.371259 4897 patch_prober.go:28] interesting pod/machine-config-daemon-krv5b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 21 15:41:04 crc kubenswrapper[4897]: I1121 15:41:04.371326 4897 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 21 15:41:05 crc kubenswrapper[4897]: I1121 15:41:05.151698 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-xdmc4" podUID="f5da2270-16fa-4d91-9110-ac7305326b10" containerName="registry-server" containerID="cri-o://90f3960cc8df6087ca42249921bcf81515300580c58eb2cb8a3b0adb482f6055" gracePeriod=2 Nov 21 15:41:05 crc kubenswrapper[4897]: I1121 15:41:05.768034 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xdmc4" Nov 21 15:41:05 crc kubenswrapper[4897]: I1121 15:41:05.797173 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5da2270-16fa-4d91-9110-ac7305326b10-utilities\") pod \"f5da2270-16fa-4d91-9110-ac7305326b10\" (UID: \"f5da2270-16fa-4d91-9110-ac7305326b10\") " Nov 21 15:41:05 crc kubenswrapper[4897]: I1121 15:41:05.797519 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qqxtd\" (UniqueName: \"kubernetes.io/projected/f5da2270-16fa-4d91-9110-ac7305326b10-kube-api-access-qqxtd\") pod \"f5da2270-16fa-4d91-9110-ac7305326b10\" (UID: \"f5da2270-16fa-4d91-9110-ac7305326b10\") " Nov 21 15:41:05 crc kubenswrapper[4897]: I1121 15:41:05.797596 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5da2270-16fa-4d91-9110-ac7305326b10-catalog-content\") pod \"f5da2270-16fa-4d91-9110-ac7305326b10\" (UID: \"f5da2270-16fa-4d91-9110-ac7305326b10\") " Nov 21 15:41:05 crc kubenswrapper[4897]: I1121 15:41:05.798077 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f5da2270-16fa-4d91-9110-ac7305326b10-utilities" (OuterVolumeSpecName: "utilities") pod "f5da2270-16fa-4d91-9110-ac7305326b10" (UID: "f5da2270-16fa-4d91-9110-ac7305326b10"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 15:41:05 crc kubenswrapper[4897]: I1121 15:41:05.804670 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5da2270-16fa-4d91-9110-ac7305326b10-kube-api-access-qqxtd" (OuterVolumeSpecName: "kube-api-access-qqxtd") pod "f5da2270-16fa-4d91-9110-ac7305326b10" (UID: "f5da2270-16fa-4d91-9110-ac7305326b10"). InnerVolumeSpecName "kube-api-access-qqxtd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 15:41:05 crc kubenswrapper[4897]: I1121 15:41:05.820903 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f5da2270-16fa-4d91-9110-ac7305326b10-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f5da2270-16fa-4d91-9110-ac7305326b10" (UID: "f5da2270-16fa-4d91-9110-ac7305326b10"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 15:41:05 crc kubenswrapper[4897]: I1121 15:41:05.901218 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qqxtd\" (UniqueName: \"kubernetes.io/projected/f5da2270-16fa-4d91-9110-ac7305326b10-kube-api-access-qqxtd\") on node \"crc\" DevicePath \"\"" Nov 21 15:41:05 crc kubenswrapper[4897]: I1121 15:41:05.901618 4897 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5da2270-16fa-4d91-9110-ac7305326b10-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 21 15:41:05 crc kubenswrapper[4897]: I1121 15:41:05.901638 4897 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5da2270-16fa-4d91-9110-ac7305326b10-utilities\") on node \"crc\" DevicePath \"\"" Nov 21 15:41:06 crc kubenswrapper[4897]: I1121 15:41:06.166197 4897 generic.go:334] "Generic (PLEG): container finished" podID="f5da2270-16fa-4d91-9110-ac7305326b10" containerID="90f3960cc8df6087ca42249921bcf81515300580c58eb2cb8a3b0adb482f6055" exitCode=0 Nov 21 15:41:06 crc kubenswrapper[4897]: I1121 15:41:06.166248 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xdmc4" event={"ID":"f5da2270-16fa-4d91-9110-ac7305326b10","Type":"ContainerDied","Data":"90f3960cc8df6087ca42249921bcf81515300580c58eb2cb8a3b0adb482f6055"} Nov 21 15:41:06 crc kubenswrapper[4897]: I1121 15:41:06.166316 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xdmc4" event={"ID":"f5da2270-16fa-4d91-9110-ac7305326b10","Type":"ContainerDied","Data":"8f50af3f0bf4bbdefedb1336a2d0809b5ead6e3a1bfdbd74d21e98e4e677d06b"} Nov 21 15:41:06 crc kubenswrapper[4897]: I1121 15:41:06.166266 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xdmc4" Nov 21 15:41:06 crc kubenswrapper[4897]: I1121 15:41:06.167029 4897 scope.go:117] "RemoveContainer" containerID="90f3960cc8df6087ca42249921bcf81515300580c58eb2cb8a3b0adb482f6055" Nov 21 15:41:06 crc kubenswrapper[4897]: I1121 15:41:06.195217 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xdmc4"] Nov 21 15:41:06 crc kubenswrapper[4897]: I1121 15:41:06.203829 4897 scope.go:117] "RemoveContainer" containerID="be29704301c9ca39234af62a77939773dbec2304201eb6789c9b1c467eeaf702" Nov 21 15:41:06 crc kubenswrapper[4897]: I1121 15:41:06.209358 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-xdmc4"] Nov 21 15:41:06 crc kubenswrapper[4897]: I1121 15:41:06.232479 4897 scope.go:117] "RemoveContainer" containerID="65bf848c64ac7cd813802d0c5ec47a5efe05c6560592cb3cf0778d70b56681a1" Nov 21 15:41:06 crc kubenswrapper[4897]: I1121 15:41:06.285712 4897 scope.go:117] "RemoveContainer" containerID="90f3960cc8df6087ca42249921bcf81515300580c58eb2cb8a3b0adb482f6055" Nov 21 15:41:06 crc kubenswrapper[4897]: E1121 15:41:06.288973 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"90f3960cc8df6087ca42249921bcf81515300580c58eb2cb8a3b0adb482f6055\": container with ID starting with 90f3960cc8df6087ca42249921bcf81515300580c58eb2cb8a3b0adb482f6055 not found: ID does not exist" containerID="90f3960cc8df6087ca42249921bcf81515300580c58eb2cb8a3b0adb482f6055" Nov 21 15:41:06 crc kubenswrapper[4897]: I1121 15:41:06.289023 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90f3960cc8df6087ca42249921bcf81515300580c58eb2cb8a3b0adb482f6055"} err="failed to get container status \"90f3960cc8df6087ca42249921bcf81515300580c58eb2cb8a3b0adb482f6055\": rpc error: code = NotFound desc = could not find container \"90f3960cc8df6087ca42249921bcf81515300580c58eb2cb8a3b0adb482f6055\": container with ID starting with 90f3960cc8df6087ca42249921bcf81515300580c58eb2cb8a3b0adb482f6055 not found: ID does not exist" Nov 21 15:41:06 crc kubenswrapper[4897]: I1121 15:41:06.289049 4897 scope.go:117] "RemoveContainer" containerID="be29704301c9ca39234af62a77939773dbec2304201eb6789c9b1c467eeaf702" Nov 21 15:41:06 crc kubenswrapper[4897]: E1121 15:41:06.289699 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"be29704301c9ca39234af62a77939773dbec2304201eb6789c9b1c467eeaf702\": container with ID starting with be29704301c9ca39234af62a77939773dbec2304201eb6789c9b1c467eeaf702 not found: ID does not exist" containerID="be29704301c9ca39234af62a77939773dbec2304201eb6789c9b1c467eeaf702" Nov 21 15:41:06 crc kubenswrapper[4897]: I1121 15:41:06.289753 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be29704301c9ca39234af62a77939773dbec2304201eb6789c9b1c467eeaf702"} err="failed to get container status \"be29704301c9ca39234af62a77939773dbec2304201eb6789c9b1c467eeaf702\": rpc error: code = NotFound desc = could not find container \"be29704301c9ca39234af62a77939773dbec2304201eb6789c9b1c467eeaf702\": container with ID starting with be29704301c9ca39234af62a77939773dbec2304201eb6789c9b1c467eeaf702 not found: ID does not exist" Nov 21 15:41:06 crc kubenswrapper[4897]: I1121 15:41:06.289791 4897 scope.go:117] "RemoveContainer" containerID="65bf848c64ac7cd813802d0c5ec47a5efe05c6560592cb3cf0778d70b56681a1" Nov 21 15:41:06 crc kubenswrapper[4897]: E1121 15:41:06.290137 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"65bf848c64ac7cd813802d0c5ec47a5efe05c6560592cb3cf0778d70b56681a1\": container with ID starting with 65bf848c64ac7cd813802d0c5ec47a5efe05c6560592cb3cf0778d70b56681a1 not found: ID does not exist" containerID="65bf848c64ac7cd813802d0c5ec47a5efe05c6560592cb3cf0778d70b56681a1" Nov 21 15:41:06 crc kubenswrapper[4897]: I1121 15:41:06.290188 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"65bf848c64ac7cd813802d0c5ec47a5efe05c6560592cb3cf0778d70b56681a1"} err="failed to get container status \"65bf848c64ac7cd813802d0c5ec47a5efe05c6560592cb3cf0778d70b56681a1\": rpc error: code = NotFound desc = could not find container \"65bf848c64ac7cd813802d0c5ec47a5efe05c6560592cb3cf0778d70b56681a1\": container with ID starting with 65bf848c64ac7cd813802d0c5ec47a5efe05c6560592cb3cf0778d70b56681a1 not found: ID does not exist" Nov 21 15:41:08 crc kubenswrapper[4897]: I1121 15:41:08.103383 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f5da2270-16fa-4d91-9110-ac7305326b10" path="/var/lib/kubelet/pods/f5da2270-16fa-4d91-9110-ac7305326b10/volumes" Nov 21 15:41:28 crc kubenswrapper[4897]: I1121 15:41:28.822900 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-bf59v"] Nov 21 15:41:28 crc kubenswrapper[4897]: E1121 15:41:28.823894 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7401e3a-fb83-453c-8d45-e2efb34c1ed8" containerName="registry-server" Nov 21 15:41:28 crc kubenswrapper[4897]: I1121 15:41:28.825230 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7401e3a-fb83-453c-8d45-e2efb34c1ed8" containerName="registry-server" Nov 21 15:41:28 crc kubenswrapper[4897]: E1121 15:41:28.825273 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7401e3a-fb83-453c-8d45-e2efb34c1ed8" containerName="extract-utilities" Nov 21 15:41:28 crc kubenswrapper[4897]: I1121 15:41:28.825284 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7401e3a-fb83-453c-8d45-e2efb34c1ed8" containerName="extract-utilities" Nov 21 15:41:28 crc kubenswrapper[4897]: E1121 15:41:28.825303 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5da2270-16fa-4d91-9110-ac7305326b10" containerName="registry-server" Nov 21 15:41:28 crc kubenswrapper[4897]: I1121 15:41:28.825309 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5da2270-16fa-4d91-9110-ac7305326b10" containerName="registry-server" Nov 21 15:41:28 crc kubenswrapper[4897]: E1121 15:41:28.825333 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5da2270-16fa-4d91-9110-ac7305326b10" containerName="extract-utilities" Nov 21 15:41:28 crc kubenswrapper[4897]: I1121 15:41:28.825339 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5da2270-16fa-4d91-9110-ac7305326b10" containerName="extract-utilities" Nov 21 15:41:28 crc kubenswrapper[4897]: E1121 15:41:28.825350 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7401e3a-fb83-453c-8d45-e2efb34c1ed8" containerName="extract-content" Nov 21 15:41:28 crc kubenswrapper[4897]: I1121 15:41:28.825356 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7401e3a-fb83-453c-8d45-e2efb34c1ed8" containerName="extract-content" Nov 21 15:41:28 crc kubenswrapper[4897]: E1121 15:41:28.825375 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5da2270-16fa-4d91-9110-ac7305326b10" containerName="extract-content" Nov 21 15:41:28 crc kubenswrapper[4897]: I1121 15:41:28.825380 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5da2270-16fa-4d91-9110-ac7305326b10" containerName="extract-content" Nov 21 15:41:28 crc kubenswrapper[4897]: I1121 15:41:28.825677 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7401e3a-fb83-453c-8d45-e2efb34c1ed8" containerName="registry-server" Nov 21 15:41:28 crc kubenswrapper[4897]: I1121 15:41:28.825695 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5da2270-16fa-4d91-9110-ac7305326b10" containerName="registry-server" Nov 21 15:41:28 crc kubenswrapper[4897]: I1121 15:41:28.827542 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bf59v" Nov 21 15:41:28 crc kubenswrapper[4897]: I1121 15:41:28.840464 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bf59v"] Nov 21 15:41:28 crc kubenswrapper[4897]: I1121 15:41:28.976107 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xhmgj\" (UniqueName: \"kubernetes.io/projected/68a0fea7-2445-4086-8d8b-3c38fdc3b2f8-kube-api-access-xhmgj\") pod \"certified-operators-bf59v\" (UID: \"68a0fea7-2445-4086-8d8b-3c38fdc3b2f8\") " pod="openshift-marketplace/certified-operators-bf59v" Nov 21 15:41:28 crc kubenswrapper[4897]: I1121 15:41:28.976185 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68a0fea7-2445-4086-8d8b-3c38fdc3b2f8-utilities\") pod \"certified-operators-bf59v\" (UID: \"68a0fea7-2445-4086-8d8b-3c38fdc3b2f8\") " pod="openshift-marketplace/certified-operators-bf59v" Nov 21 15:41:28 crc kubenswrapper[4897]: I1121 15:41:28.976237 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68a0fea7-2445-4086-8d8b-3c38fdc3b2f8-catalog-content\") pod \"certified-operators-bf59v\" (UID: \"68a0fea7-2445-4086-8d8b-3c38fdc3b2f8\") " pod="openshift-marketplace/certified-operators-bf59v" Nov 21 15:41:29 crc kubenswrapper[4897]: I1121 15:41:29.078780 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xhmgj\" (UniqueName: \"kubernetes.io/projected/68a0fea7-2445-4086-8d8b-3c38fdc3b2f8-kube-api-access-xhmgj\") pod \"certified-operators-bf59v\" (UID: \"68a0fea7-2445-4086-8d8b-3c38fdc3b2f8\") " pod="openshift-marketplace/certified-operators-bf59v" Nov 21 15:41:29 crc kubenswrapper[4897]: I1121 15:41:29.078840 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68a0fea7-2445-4086-8d8b-3c38fdc3b2f8-utilities\") pod \"certified-operators-bf59v\" (UID: \"68a0fea7-2445-4086-8d8b-3c38fdc3b2f8\") " pod="openshift-marketplace/certified-operators-bf59v" Nov 21 15:41:29 crc kubenswrapper[4897]: I1121 15:41:29.078865 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68a0fea7-2445-4086-8d8b-3c38fdc3b2f8-catalog-content\") pod \"certified-operators-bf59v\" (UID: \"68a0fea7-2445-4086-8d8b-3c38fdc3b2f8\") " pod="openshift-marketplace/certified-operators-bf59v" Nov 21 15:41:29 crc kubenswrapper[4897]: I1121 15:41:29.083478 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68a0fea7-2445-4086-8d8b-3c38fdc3b2f8-catalog-content\") pod \"certified-operators-bf59v\" (UID: \"68a0fea7-2445-4086-8d8b-3c38fdc3b2f8\") " pod="openshift-marketplace/certified-operators-bf59v" Nov 21 15:41:29 crc kubenswrapper[4897]: I1121 15:41:29.083872 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68a0fea7-2445-4086-8d8b-3c38fdc3b2f8-utilities\") pod \"certified-operators-bf59v\" (UID: \"68a0fea7-2445-4086-8d8b-3c38fdc3b2f8\") " pod="openshift-marketplace/certified-operators-bf59v" Nov 21 15:41:29 crc kubenswrapper[4897]: I1121 15:41:29.142064 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xhmgj\" (UniqueName: \"kubernetes.io/projected/68a0fea7-2445-4086-8d8b-3c38fdc3b2f8-kube-api-access-xhmgj\") pod \"certified-operators-bf59v\" (UID: \"68a0fea7-2445-4086-8d8b-3c38fdc3b2f8\") " pod="openshift-marketplace/certified-operators-bf59v" Nov 21 15:41:29 crc kubenswrapper[4897]: I1121 15:41:29.171413 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bf59v" Nov 21 15:41:29 crc kubenswrapper[4897]: I1121 15:41:29.887341 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bf59v"] Nov 21 15:41:30 crc kubenswrapper[4897]: I1121 15:41:30.454165 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bf59v" event={"ID":"68a0fea7-2445-4086-8d8b-3c38fdc3b2f8","Type":"ContainerStarted","Data":"268f41ee841cc7c5fb1e2454d44944f3bab8eaeb36ffaa459810037509d54c7b"} Nov 21 15:41:31 crc kubenswrapper[4897]: I1121 15:41:31.464162 4897 generic.go:334] "Generic (PLEG): container finished" podID="68a0fea7-2445-4086-8d8b-3c38fdc3b2f8" containerID="879ec92bb8ee1383f59892be473377dd0f327c6228b608381262f6223baeecfe" exitCode=0 Nov 21 15:41:31 crc kubenswrapper[4897]: I1121 15:41:31.464373 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bf59v" event={"ID":"68a0fea7-2445-4086-8d8b-3c38fdc3b2f8","Type":"ContainerDied","Data":"879ec92bb8ee1383f59892be473377dd0f327c6228b608381262f6223baeecfe"} Nov 21 15:41:31 crc kubenswrapper[4897]: I1121 15:41:31.468127 4897 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 21 15:41:32 crc kubenswrapper[4897]: I1121 15:41:32.476611 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bf59v" event={"ID":"68a0fea7-2445-4086-8d8b-3c38fdc3b2f8","Type":"ContainerStarted","Data":"3f29116f47fccacc993ae35c17add1f832eff74677b1d418bad5cdf960cdba96"} Nov 21 15:41:34 crc kubenswrapper[4897]: I1121 15:41:34.371258 4897 patch_prober.go:28] interesting pod/machine-config-daemon-krv5b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 21 15:41:34 crc kubenswrapper[4897]: I1121 15:41:34.371876 4897 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 21 15:41:34 crc kubenswrapper[4897]: I1121 15:41:34.371934 4897 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" Nov 21 15:41:34 crc kubenswrapper[4897]: I1121 15:41:34.373584 4897 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b75f9d31a9263f80ca2b82ff3c2e842f5c5ccc5e8cefc8a57960d6be99c63908"} pod="openshift-machine-config-operator/machine-config-daemon-krv5b" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 21 15:41:34 crc kubenswrapper[4897]: I1121 15:41:34.373789 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" containerName="machine-config-daemon" containerID="cri-o://b75f9d31a9263f80ca2b82ff3c2e842f5c5ccc5e8cefc8a57960d6be99c63908" gracePeriod=600 Nov 21 15:41:34 crc kubenswrapper[4897]: I1121 15:41:34.529349 4897 generic.go:334] "Generic (PLEG): container finished" podID="e7670227-d280-4847-b882-754429f56b0f" containerID="b75f9d31a9263f80ca2b82ff3c2e842f5c5ccc5e8cefc8a57960d6be99c63908" exitCode=0 Nov 21 15:41:34 crc kubenswrapper[4897]: I1121 15:41:34.529448 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" event={"ID":"e7670227-d280-4847-b882-754429f56b0f","Type":"ContainerDied","Data":"b75f9d31a9263f80ca2b82ff3c2e842f5c5ccc5e8cefc8a57960d6be99c63908"} Nov 21 15:41:34 crc kubenswrapper[4897]: I1121 15:41:34.529487 4897 scope.go:117] "RemoveContainer" containerID="abf30c46658df7e750da25f3feeb4a134367a6c61c5b5221d6703974a1dd4d2e" Nov 21 15:41:34 crc kubenswrapper[4897]: I1121 15:41:34.535804 4897 generic.go:334] "Generic (PLEG): container finished" podID="68a0fea7-2445-4086-8d8b-3c38fdc3b2f8" containerID="3f29116f47fccacc993ae35c17add1f832eff74677b1d418bad5cdf960cdba96" exitCode=0 Nov 21 15:41:34 crc kubenswrapper[4897]: I1121 15:41:34.535857 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bf59v" event={"ID":"68a0fea7-2445-4086-8d8b-3c38fdc3b2f8","Type":"ContainerDied","Data":"3f29116f47fccacc993ae35c17add1f832eff74677b1d418bad5cdf960cdba96"} Nov 21 15:41:35 crc kubenswrapper[4897]: I1121 15:41:35.547750 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" event={"ID":"e7670227-d280-4847-b882-754429f56b0f","Type":"ContainerStarted","Data":"9f6ac9e28e3e389a8236d0859923fa122a6eea6c927d5c77a01f623f7524a046"} Nov 21 15:41:36 crc kubenswrapper[4897]: I1121 15:41:36.559972 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bf59v" event={"ID":"68a0fea7-2445-4086-8d8b-3c38fdc3b2f8","Type":"ContainerStarted","Data":"a8e1929da87f066c93ea4a64080147df220bcf922bfa10b79d39232bcb7ae93f"} Nov 21 15:41:36 crc kubenswrapper[4897]: I1121 15:41:36.586659 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-bf59v" podStartSLOduration=4.094935317 podStartE2EDuration="8.586636235s" podCreationTimestamp="2025-11-21 15:41:28 +0000 UTC" firstStartedPulling="2025-11-21 15:41:31.467055471 +0000 UTC m=+5568.751648946" lastFinishedPulling="2025-11-21 15:41:35.958756389 +0000 UTC m=+5573.243349864" observedRunningTime="2025-11-21 15:41:36.578564717 +0000 UTC m=+5573.863158212" watchObservedRunningTime="2025-11-21 15:41:36.586636235 +0000 UTC m=+5573.871229710" Nov 21 15:41:39 crc kubenswrapper[4897]: I1121 15:41:39.171577 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-bf59v" Nov 21 15:41:39 crc kubenswrapper[4897]: I1121 15:41:39.172103 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-bf59v" Nov 21 15:41:39 crc kubenswrapper[4897]: I1121 15:41:39.224119 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-bf59v" Nov 21 15:41:49 crc kubenswrapper[4897]: I1121 15:41:49.277166 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-bf59v" Nov 21 15:41:49 crc kubenswrapper[4897]: I1121 15:41:49.341457 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-bf59v"] Nov 21 15:41:49 crc kubenswrapper[4897]: I1121 15:41:49.686254 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-bf59v" podUID="68a0fea7-2445-4086-8d8b-3c38fdc3b2f8" containerName="registry-server" containerID="cri-o://a8e1929da87f066c93ea4a64080147df220bcf922bfa10b79d39232bcb7ae93f" gracePeriod=2 Nov 21 15:41:50 crc kubenswrapper[4897]: I1121 15:41:50.320716 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bf59v" Nov 21 15:41:50 crc kubenswrapper[4897]: I1121 15:41:50.397617 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68a0fea7-2445-4086-8d8b-3c38fdc3b2f8-catalog-content\") pod \"68a0fea7-2445-4086-8d8b-3c38fdc3b2f8\" (UID: \"68a0fea7-2445-4086-8d8b-3c38fdc3b2f8\") " Nov 21 15:41:50 crc kubenswrapper[4897]: I1121 15:41:50.397800 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68a0fea7-2445-4086-8d8b-3c38fdc3b2f8-utilities\") pod \"68a0fea7-2445-4086-8d8b-3c38fdc3b2f8\" (UID: \"68a0fea7-2445-4086-8d8b-3c38fdc3b2f8\") " Nov 21 15:41:50 crc kubenswrapper[4897]: I1121 15:41:50.397878 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xhmgj\" (UniqueName: \"kubernetes.io/projected/68a0fea7-2445-4086-8d8b-3c38fdc3b2f8-kube-api-access-xhmgj\") pod \"68a0fea7-2445-4086-8d8b-3c38fdc3b2f8\" (UID: \"68a0fea7-2445-4086-8d8b-3c38fdc3b2f8\") " Nov 21 15:41:50 crc kubenswrapper[4897]: I1121 15:41:50.398479 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/68a0fea7-2445-4086-8d8b-3c38fdc3b2f8-utilities" (OuterVolumeSpecName: "utilities") pod "68a0fea7-2445-4086-8d8b-3c38fdc3b2f8" (UID: "68a0fea7-2445-4086-8d8b-3c38fdc3b2f8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 15:41:50 crc kubenswrapper[4897]: I1121 15:41:50.404571 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68a0fea7-2445-4086-8d8b-3c38fdc3b2f8-kube-api-access-xhmgj" (OuterVolumeSpecName: "kube-api-access-xhmgj") pod "68a0fea7-2445-4086-8d8b-3c38fdc3b2f8" (UID: "68a0fea7-2445-4086-8d8b-3c38fdc3b2f8"). InnerVolumeSpecName "kube-api-access-xhmgj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 15:41:50 crc kubenswrapper[4897]: I1121 15:41:50.450607 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/68a0fea7-2445-4086-8d8b-3c38fdc3b2f8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "68a0fea7-2445-4086-8d8b-3c38fdc3b2f8" (UID: "68a0fea7-2445-4086-8d8b-3c38fdc3b2f8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 15:41:50 crc kubenswrapper[4897]: I1121 15:41:50.500231 4897 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68a0fea7-2445-4086-8d8b-3c38fdc3b2f8-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 21 15:41:50 crc kubenswrapper[4897]: I1121 15:41:50.500261 4897 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68a0fea7-2445-4086-8d8b-3c38fdc3b2f8-utilities\") on node \"crc\" DevicePath \"\"" Nov 21 15:41:50 crc kubenswrapper[4897]: I1121 15:41:50.500272 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xhmgj\" (UniqueName: \"kubernetes.io/projected/68a0fea7-2445-4086-8d8b-3c38fdc3b2f8-kube-api-access-xhmgj\") on node \"crc\" DevicePath \"\"" Nov 21 15:41:50 crc kubenswrapper[4897]: I1121 15:41:50.700932 4897 generic.go:334] "Generic (PLEG): container finished" podID="68a0fea7-2445-4086-8d8b-3c38fdc3b2f8" containerID="a8e1929da87f066c93ea4a64080147df220bcf922bfa10b79d39232bcb7ae93f" exitCode=0 Nov 21 15:41:50 crc kubenswrapper[4897]: I1121 15:41:50.700986 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bf59v" event={"ID":"68a0fea7-2445-4086-8d8b-3c38fdc3b2f8","Type":"ContainerDied","Data":"a8e1929da87f066c93ea4a64080147df220bcf922bfa10b79d39232bcb7ae93f"} Nov 21 15:41:50 crc kubenswrapper[4897]: I1121 15:41:50.701017 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bf59v" event={"ID":"68a0fea7-2445-4086-8d8b-3c38fdc3b2f8","Type":"ContainerDied","Data":"268f41ee841cc7c5fb1e2454d44944f3bab8eaeb36ffaa459810037509d54c7b"} Nov 21 15:41:50 crc kubenswrapper[4897]: I1121 15:41:50.700992 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bf59v" Nov 21 15:41:50 crc kubenswrapper[4897]: I1121 15:41:50.701036 4897 scope.go:117] "RemoveContainer" containerID="a8e1929da87f066c93ea4a64080147df220bcf922bfa10b79d39232bcb7ae93f" Nov 21 15:41:50 crc kubenswrapper[4897]: I1121 15:41:50.731195 4897 scope.go:117] "RemoveContainer" containerID="3f29116f47fccacc993ae35c17add1f832eff74677b1d418bad5cdf960cdba96" Nov 21 15:41:50 crc kubenswrapper[4897]: I1121 15:41:50.737342 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-bf59v"] Nov 21 15:41:50 crc kubenswrapper[4897]: I1121 15:41:50.746905 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-bf59v"] Nov 21 15:41:50 crc kubenswrapper[4897]: I1121 15:41:50.766410 4897 scope.go:117] "RemoveContainer" containerID="879ec92bb8ee1383f59892be473377dd0f327c6228b608381262f6223baeecfe" Nov 21 15:41:50 crc kubenswrapper[4897]: I1121 15:41:50.814185 4897 scope.go:117] "RemoveContainer" containerID="a8e1929da87f066c93ea4a64080147df220bcf922bfa10b79d39232bcb7ae93f" Nov 21 15:41:50 crc kubenswrapper[4897]: E1121 15:41:50.814607 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a8e1929da87f066c93ea4a64080147df220bcf922bfa10b79d39232bcb7ae93f\": container with ID starting with a8e1929da87f066c93ea4a64080147df220bcf922bfa10b79d39232bcb7ae93f not found: ID does not exist" containerID="a8e1929da87f066c93ea4a64080147df220bcf922bfa10b79d39232bcb7ae93f" Nov 21 15:41:50 crc kubenswrapper[4897]: I1121 15:41:50.814649 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8e1929da87f066c93ea4a64080147df220bcf922bfa10b79d39232bcb7ae93f"} err="failed to get container status \"a8e1929da87f066c93ea4a64080147df220bcf922bfa10b79d39232bcb7ae93f\": rpc error: code = NotFound desc = could not find container \"a8e1929da87f066c93ea4a64080147df220bcf922bfa10b79d39232bcb7ae93f\": container with ID starting with a8e1929da87f066c93ea4a64080147df220bcf922bfa10b79d39232bcb7ae93f not found: ID does not exist" Nov 21 15:41:50 crc kubenswrapper[4897]: I1121 15:41:50.814675 4897 scope.go:117] "RemoveContainer" containerID="3f29116f47fccacc993ae35c17add1f832eff74677b1d418bad5cdf960cdba96" Nov 21 15:41:50 crc kubenswrapper[4897]: E1121 15:41:50.814899 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3f29116f47fccacc993ae35c17add1f832eff74677b1d418bad5cdf960cdba96\": container with ID starting with 3f29116f47fccacc993ae35c17add1f832eff74677b1d418bad5cdf960cdba96 not found: ID does not exist" containerID="3f29116f47fccacc993ae35c17add1f832eff74677b1d418bad5cdf960cdba96" Nov 21 15:41:50 crc kubenswrapper[4897]: I1121 15:41:50.814922 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f29116f47fccacc993ae35c17add1f832eff74677b1d418bad5cdf960cdba96"} err="failed to get container status \"3f29116f47fccacc993ae35c17add1f832eff74677b1d418bad5cdf960cdba96\": rpc error: code = NotFound desc = could not find container \"3f29116f47fccacc993ae35c17add1f832eff74677b1d418bad5cdf960cdba96\": container with ID starting with 3f29116f47fccacc993ae35c17add1f832eff74677b1d418bad5cdf960cdba96 not found: ID does not exist" Nov 21 15:41:50 crc kubenswrapper[4897]: I1121 15:41:50.814937 4897 scope.go:117] "RemoveContainer" containerID="879ec92bb8ee1383f59892be473377dd0f327c6228b608381262f6223baeecfe" Nov 21 15:41:50 crc kubenswrapper[4897]: E1121 15:41:50.815276 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"879ec92bb8ee1383f59892be473377dd0f327c6228b608381262f6223baeecfe\": container with ID starting with 879ec92bb8ee1383f59892be473377dd0f327c6228b608381262f6223baeecfe not found: ID does not exist" containerID="879ec92bb8ee1383f59892be473377dd0f327c6228b608381262f6223baeecfe" Nov 21 15:41:50 crc kubenswrapper[4897]: I1121 15:41:50.815302 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"879ec92bb8ee1383f59892be473377dd0f327c6228b608381262f6223baeecfe"} err="failed to get container status \"879ec92bb8ee1383f59892be473377dd0f327c6228b608381262f6223baeecfe\": rpc error: code = NotFound desc = could not find container \"879ec92bb8ee1383f59892be473377dd0f327c6228b608381262f6223baeecfe\": container with ID starting with 879ec92bb8ee1383f59892be473377dd0f327c6228b608381262f6223baeecfe not found: ID does not exist" Nov 21 15:41:52 crc kubenswrapper[4897]: I1121 15:41:52.105084 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="68a0fea7-2445-4086-8d8b-3c38fdc3b2f8" path="/var/lib/kubelet/pods/68a0fea7-2445-4086-8d8b-3c38fdc3b2f8/volumes" Nov 21 15:43:34 crc kubenswrapper[4897]: I1121 15:43:34.371418 4897 patch_prober.go:28] interesting pod/machine-config-daemon-krv5b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 21 15:43:34 crc kubenswrapper[4897]: I1121 15:43:34.372163 4897 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 21 15:44:04 crc kubenswrapper[4897]: I1121 15:44:04.370336 4897 patch_prober.go:28] interesting pod/machine-config-daemon-krv5b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 21 15:44:04 crc kubenswrapper[4897]: I1121 15:44:04.370953 4897 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 21 15:44:34 crc kubenswrapper[4897]: I1121 15:44:34.370468 4897 patch_prober.go:28] interesting pod/machine-config-daemon-krv5b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 21 15:44:34 crc kubenswrapper[4897]: I1121 15:44:34.371189 4897 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 21 15:44:34 crc kubenswrapper[4897]: I1121 15:44:34.371259 4897 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" Nov 21 15:44:34 crc kubenswrapper[4897]: I1121 15:44:34.372413 4897 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9f6ac9e28e3e389a8236d0859923fa122a6eea6c927d5c77a01f623f7524a046"} pod="openshift-machine-config-operator/machine-config-daemon-krv5b" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 21 15:44:34 crc kubenswrapper[4897]: I1121 15:44:34.372471 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" containerName="machine-config-daemon" containerID="cri-o://9f6ac9e28e3e389a8236d0859923fa122a6eea6c927d5c77a01f623f7524a046" gracePeriod=600 Nov 21 15:44:34 crc kubenswrapper[4897]: E1121 15:44:34.498414 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:44:34 crc kubenswrapper[4897]: E1121 15:44:34.584146 4897 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode7670227_d280_4847_b882_754429f56b0f.slice/crio-9f6ac9e28e3e389a8236d0859923fa122a6eea6c927d5c77a01f623f7524a046.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode7670227_d280_4847_b882_754429f56b0f.slice/crio-conmon-9f6ac9e28e3e389a8236d0859923fa122a6eea6c927d5c77a01f623f7524a046.scope\": RecentStats: unable to find data in memory cache]" Nov 21 15:44:35 crc kubenswrapper[4897]: I1121 15:44:35.383299 4897 generic.go:334] "Generic (PLEG): container finished" podID="e7670227-d280-4847-b882-754429f56b0f" containerID="9f6ac9e28e3e389a8236d0859923fa122a6eea6c927d5c77a01f623f7524a046" exitCode=0 Nov 21 15:44:35 crc kubenswrapper[4897]: I1121 15:44:35.383350 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" event={"ID":"e7670227-d280-4847-b882-754429f56b0f","Type":"ContainerDied","Data":"9f6ac9e28e3e389a8236d0859923fa122a6eea6c927d5c77a01f623f7524a046"} Nov 21 15:44:35 crc kubenswrapper[4897]: I1121 15:44:35.383388 4897 scope.go:117] "RemoveContainer" containerID="b75f9d31a9263f80ca2b82ff3c2e842f5c5ccc5e8cefc8a57960d6be99c63908" Nov 21 15:44:35 crc kubenswrapper[4897]: I1121 15:44:35.384260 4897 scope.go:117] "RemoveContainer" containerID="9f6ac9e28e3e389a8236d0859923fa122a6eea6c927d5c77a01f623f7524a046" Nov 21 15:44:35 crc kubenswrapper[4897]: E1121 15:44:35.384593 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:44:44 crc kubenswrapper[4897]: E1121 15:44:44.879076 4897 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/system.slice/rpm-ostreed.service\": RecentStats: unable to find data in memory cache]" Nov 21 15:44:46 crc kubenswrapper[4897]: I1121 15:44:46.089053 4897 scope.go:117] "RemoveContainer" containerID="9f6ac9e28e3e389a8236d0859923fa122a6eea6c927d5c77a01f623f7524a046" Nov 21 15:44:46 crc kubenswrapper[4897]: E1121 15:44:46.089798 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:45:00 crc kubenswrapper[4897]: I1121 15:45:00.090061 4897 scope.go:117] "RemoveContainer" containerID="9f6ac9e28e3e389a8236d0859923fa122a6eea6c927d5c77a01f623f7524a046" Nov 21 15:45:00 crc kubenswrapper[4897]: E1121 15:45:00.090926 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:45:00 crc kubenswrapper[4897]: I1121 15:45:00.169860 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29395665-d87ql"] Nov 21 15:45:00 crc kubenswrapper[4897]: E1121 15:45:00.170313 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68a0fea7-2445-4086-8d8b-3c38fdc3b2f8" containerName="extract-content" Nov 21 15:45:00 crc kubenswrapper[4897]: I1121 15:45:00.170331 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="68a0fea7-2445-4086-8d8b-3c38fdc3b2f8" containerName="extract-content" Nov 21 15:45:00 crc kubenswrapper[4897]: E1121 15:45:00.170387 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68a0fea7-2445-4086-8d8b-3c38fdc3b2f8" containerName="registry-server" Nov 21 15:45:00 crc kubenswrapper[4897]: I1121 15:45:00.170395 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="68a0fea7-2445-4086-8d8b-3c38fdc3b2f8" containerName="registry-server" Nov 21 15:45:00 crc kubenswrapper[4897]: E1121 15:45:00.170429 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68a0fea7-2445-4086-8d8b-3c38fdc3b2f8" containerName="extract-utilities" Nov 21 15:45:00 crc kubenswrapper[4897]: I1121 15:45:00.170437 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="68a0fea7-2445-4086-8d8b-3c38fdc3b2f8" containerName="extract-utilities" Nov 21 15:45:00 crc kubenswrapper[4897]: I1121 15:45:00.170791 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="68a0fea7-2445-4086-8d8b-3c38fdc3b2f8" containerName="registry-server" Nov 21 15:45:00 crc kubenswrapper[4897]: I1121 15:45:00.171556 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29395665-d87ql" Nov 21 15:45:00 crc kubenswrapper[4897]: I1121 15:45:00.184539 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29395665-d87ql"] Nov 21 15:45:00 crc kubenswrapper[4897]: I1121 15:45:00.190059 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 21 15:45:00 crc kubenswrapper[4897]: I1121 15:45:00.194801 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 21 15:45:00 crc kubenswrapper[4897]: I1121 15:45:00.264999 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8bb5aaa3-8c31-413c-8f29-fbf37e0e380b-secret-volume\") pod \"collect-profiles-29395665-d87ql\" (UID: \"8bb5aaa3-8c31-413c-8f29-fbf37e0e380b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29395665-d87ql" Nov 21 15:45:00 crc kubenswrapper[4897]: I1121 15:45:00.265156 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8bb5aaa3-8c31-413c-8f29-fbf37e0e380b-config-volume\") pod \"collect-profiles-29395665-d87ql\" (UID: \"8bb5aaa3-8c31-413c-8f29-fbf37e0e380b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29395665-d87ql" Nov 21 15:45:00 crc kubenswrapper[4897]: I1121 15:45:00.265471 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8k2fx\" (UniqueName: \"kubernetes.io/projected/8bb5aaa3-8c31-413c-8f29-fbf37e0e380b-kube-api-access-8k2fx\") pod \"collect-profiles-29395665-d87ql\" (UID: \"8bb5aaa3-8c31-413c-8f29-fbf37e0e380b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29395665-d87ql" Nov 21 15:45:00 crc kubenswrapper[4897]: I1121 15:45:00.368305 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8k2fx\" (UniqueName: \"kubernetes.io/projected/8bb5aaa3-8c31-413c-8f29-fbf37e0e380b-kube-api-access-8k2fx\") pod \"collect-profiles-29395665-d87ql\" (UID: \"8bb5aaa3-8c31-413c-8f29-fbf37e0e380b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29395665-d87ql" Nov 21 15:45:00 crc kubenswrapper[4897]: I1121 15:45:00.368408 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8bb5aaa3-8c31-413c-8f29-fbf37e0e380b-secret-volume\") pod \"collect-profiles-29395665-d87ql\" (UID: \"8bb5aaa3-8c31-413c-8f29-fbf37e0e380b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29395665-d87ql" Nov 21 15:45:00 crc kubenswrapper[4897]: I1121 15:45:00.368534 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8bb5aaa3-8c31-413c-8f29-fbf37e0e380b-config-volume\") pod \"collect-profiles-29395665-d87ql\" (UID: \"8bb5aaa3-8c31-413c-8f29-fbf37e0e380b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29395665-d87ql" Nov 21 15:45:00 crc kubenswrapper[4897]: I1121 15:45:00.369774 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8bb5aaa3-8c31-413c-8f29-fbf37e0e380b-config-volume\") pod \"collect-profiles-29395665-d87ql\" (UID: \"8bb5aaa3-8c31-413c-8f29-fbf37e0e380b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29395665-d87ql" Nov 21 15:45:00 crc kubenswrapper[4897]: I1121 15:45:00.375810 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8bb5aaa3-8c31-413c-8f29-fbf37e0e380b-secret-volume\") pod \"collect-profiles-29395665-d87ql\" (UID: \"8bb5aaa3-8c31-413c-8f29-fbf37e0e380b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29395665-d87ql" Nov 21 15:45:00 crc kubenswrapper[4897]: I1121 15:45:00.388983 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8k2fx\" (UniqueName: \"kubernetes.io/projected/8bb5aaa3-8c31-413c-8f29-fbf37e0e380b-kube-api-access-8k2fx\") pod \"collect-profiles-29395665-d87ql\" (UID: \"8bb5aaa3-8c31-413c-8f29-fbf37e0e380b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29395665-d87ql" Nov 21 15:45:00 crc kubenswrapper[4897]: I1121 15:45:00.515718 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29395665-d87ql" Nov 21 15:45:01 crc kubenswrapper[4897]: I1121 15:45:01.021662 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29395665-d87ql"] Nov 21 15:45:01 crc kubenswrapper[4897]: I1121 15:45:01.699737 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29395665-d87ql" event={"ID":"8bb5aaa3-8c31-413c-8f29-fbf37e0e380b","Type":"ContainerStarted","Data":"f045535ee7e2fbd48d033a6e8c54ad4f990abdf0d2db26389c5165c1bd17f40f"} Nov 21 15:45:01 crc kubenswrapper[4897]: I1121 15:45:01.700281 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29395665-d87ql" event={"ID":"8bb5aaa3-8c31-413c-8f29-fbf37e0e380b","Type":"ContainerStarted","Data":"6c20019e68c90c081f4dd7e47f501e9f909832448abade58bc2e5378fe6b5e57"} Nov 21 15:45:01 crc kubenswrapper[4897]: I1121 15:45:01.718413 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29395665-d87ql" podStartSLOduration=1.7183951309999999 podStartE2EDuration="1.718395131s" podCreationTimestamp="2025-11-21 15:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 15:45:01.712818931 +0000 UTC m=+5778.997412406" watchObservedRunningTime="2025-11-21 15:45:01.718395131 +0000 UTC m=+5779.002988606" Nov 21 15:45:02 crc kubenswrapper[4897]: I1121 15:45:02.714408 4897 generic.go:334] "Generic (PLEG): container finished" podID="8bb5aaa3-8c31-413c-8f29-fbf37e0e380b" containerID="f045535ee7e2fbd48d033a6e8c54ad4f990abdf0d2db26389c5165c1bd17f40f" exitCode=0 Nov 21 15:45:02 crc kubenswrapper[4897]: I1121 15:45:02.714682 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29395665-d87ql" event={"ID":"8bb5aaa3-8c31-413c-8f29-fbf37e0e380b","Type":"ContainerDied","Data":"f045535ee7e2fbd48d033a6e8c54ad4f990abdf0d2db26389c5165c1bd17f40f"} Nov 21 15:45:04 crc kubenswrapper[4897]: I1121 15:45:04.147264 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29395665-d87ql" Nov 21 15:45:04 crc kubenswrapper[4897]: I1121 15:45:04.259056 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8bb5aaa3-8c31-413c-8f29-fbf37e0e380b-config-volume\") pod \"8bb5aaa3-8c31-413c-8f29-fbf37e0e380b\" (UID: \"8bb5aaa3-8c31-413c-8f29-fbf37e0e380b\") " Nov 21 15:45:04 crc kubenswrapper[4897]: I1121 15:45:04.259288 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8bb5aaa3-8c31-413c-8f29-fbf37e0e380b-secret-volume\") pod \"8bb5aaa3-8c31-413c-8f29-fbf37e0e380b\" (UID: \"8bb5aaa3-8c31-413c-8f29-fbf37e0e380b\") " Nov 21 15:45:04 crc kubenswrapper[4897]: I1121 15:45:04.259499 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8k2fx\" (UniqueName: \"kubernetes.io/projected/8bb5aaa3-8c31-413c-8f29-fbf37e0e380b-kube-api-access-8k2fx\") pod \"8bb5aaa3-8c31-413c-8f29-fbf37e0e380b\" (UID: \"8bb5aaa3-8c31-413c-8f29-fbf37e0e380b\") " Nov 21 15:45:04 crc kubenswrapper[4897]: I1121 15:45:04.259945 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8bb5aaa3-8c31-413c-8f29-fbf37e0e380b-config-volume" (OuterVolumeSpecName: "config-volume") pod "8bb5aaa3-8c31-413c-8f29-fbf37e0e380b" (UID: "8bb5aaa3-8c31-413c-8f29-fbf37e0e380b"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 15:45:04 crc kubenswrapper[4897]: I1121 15:45:04.260308 4897 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8bb5aaa3-8c31-413c-8f29-fbf37e0e380b-config-volume\") on node \"crc\" DevicePath \"\"" Nov 21 15:45:04 crc kubenswrapper[4897]: I1121 15:45:04.265569 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8bb5aaa3-8c31-413c-8f29-fbf37e0e380b-kube-api-access-8k2fx" (OuterVolumeSpecName: "kube-api-access-8k2fx") pod "8bb5aaa3-8c31-413c-8f29-fbf37e0e380b" (UID: "8bb5aaa3-8c31-413c-8f29-fbf37e0e380b"). InnerVolumeSpecName "kube-api-access-8k2fx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 15:45:04 crc kubenswrapper[4897]: I1121 15:45:04.270792 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8bb5aaa3-8c31-413c-8f29-fbf37e0e380b-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "8bb5aaa3-8c31-413c-8f29-fbf37e0e380b" (UID: "8bb5aaa3-8c31-413c-8f29-fbf37e0e380b"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 15:45:04 crc kubenswrapper[4897]: I1121 15:45:04.362280 4897 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8bb5aaa3-8c31-413c-8f29-fbf37e0e380b-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 21 15:45:04 crc kubenswrapper[4897]: I1121 15:45:04.362320 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8k2fx\" (UniqueName: \"kubernetes.io/projected/8bb5aaa3-8c31-413c-8f29-fbf37e0e380b-kube-api-access-8k2fx\") on node \"crc\" DevicePath \"\"" Nov 21 15:45:04 crc kubenswrapper[4897]: I1121 15:45:04.738720 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29395665-d87ql" event={"ID":"8bb5aaa3-8c31-413c-8f29-fbf37e0e380b","Type":"ContainerDied","Data":"6c20019e68c90c081f4dd7e47f501e9f909832448abade58bc2e5378fe6b5e57"} Nov 21 15:45:04 crc kubenswrapper[4897]: I1121 15:45:04.738782 4897 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6c20019e68c90c081f4dd7e47f501e9f909832448abade58bc2e5378fe6b5e57" Nov 21 15:45:04 crc kubenswrapper[4897]: I1121 15:45:04.738793 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29395665-d87ql" Nov 21 15:45:04 crc kubenswrapper[4897]: I1121 15:45:04.798378 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29395620-84swg"] Nov 21 15:45:04 crc kubenswrapper[4897]: I1121 15:45:04.809267 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29395620-84swg"] Nov 21 15:45:06 crc kubenswrapper[4897]: I1121 15:45:06.109460 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c2518e25-e7d2-4bbe-b2fb-1aa2f933069b" path="/var/lib/kubelet/pods/c2518e25-e7d2-4bbe-b2fb-1aa2f933069b/volumes" Nov 21 15:45:13 crc kubenswrapper[4897]: I1121 15:45:13.090120 4897 scope.go:117] "RemoveContainer" containerID="9f6ac9e28e3e389a8236d0859923fa122a6eea6c927d5c77a01f623f7524a046" Nov 21 15:45:13 crc kubenswrapper[4897]: E1121 15:45:13.090937 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:45:28 crc kubenswrapper[4897]: I1121 15:45:28.090624 4897 scope.go:117] "RemoveContainer" containerID="9f6ac9e28e3e389a8236d0859923fa122a6eea6c927d5c77a01f623f7524a046" Nov 21 15:45:28 crc kubenswrapper[4897]: E1121 15:45:28.091467 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:45:43 crc kubenswrapper[4897]: I1121 15:45:43.089756 4897 scope.go:117] "RemoveContainer" containerID="9f6ac9e28e3e389a8236d0859923fa122a6eea6c927d5c77a01f623f7524a046" Nov 21 15:45:43 crc kubenswrapper[4897]: E1121 15:45:43.090488 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:45:44 crc kubenswrapper[4897]: I1121 15:45:44.164763 4897 scope.go:117] "RemoveContainer" containerID="fdaaddea473e11bf7a52648c4831c17c899e611002c030ddf6c9f4ad8d48c283" Nov 21 15:45:55 crc kubenswrapper[4897]: I1121 15:45:55.090066 4897 scope.go:117] "RemoveContainer" containerID="9f6ac9e28e3e389a8236d0859923fa122a6eea6c927d5c77a01f623f7524a046" Nov 21 15:45:55 crc kubenswrapper[4897]: E1121 15:45:55.091556 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:46:07 crc kubenswrapper[4897]: I1121 15:46:07.044359 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-xxvts"] Nov 21 15:46:07 crc kubenswrapper[4897]: E1121 15:46:07.045443 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bb5aaa3-8c31-413c-8f29-fbf37e0e380b" containerName="collect-profiles" Nov 21 15:46:07 crc kubenswrapper[4897]: I1121 15:46:07.045457 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bb5aaa3-8c31-413c-8f29-fbf37e0e380b" containerName="collect-profiles" Nov 21 15:46:07 crc kubenswrapper[4897]: I1121 15:46:07.045766 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="8bb5aaa3-8c31-413c-8f29-fbf37e0e380b" containerName="collect-profiles" Nov 21 15:46:07 crc kubenswrapper[4897]: I1121 15:46:07.048082 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xxvts" Nov 21 15:46:07 crc kubenswrapper[4897]: I1121 15:46:07.059656 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xxvts"] Nov 21 15:46:07 crc kubenswrapper[4897]: I1121 15:46:07.090065 4897 scope.go:117] "RemoveContainer" containerID="9f6ac9e28e3e389a8236d0859923fa122a6eea6c927d5c77a01f623f7524a046" Nov 21 15:46:07 crc kubenswrapper[4897]: E1121 15:46:07.090555 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:46:07 crc kubenswrapper[4897]: I1121 15:46:07.169905 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eeaab305-c370-45b0-8db9-cf907726dfd5-utilities\") pod \"community-operators-xxvts\" (UID: \"eeaab305-c370-45b0-8db9-cf907726dfd5\") " pod="openshift-marketplace/community-operators-xxvts" Nov 21 15:46:07 crc kubenswrapper[4897]: I1121 15:46:07.170058 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-swgqf\" (UniqueName: \"kubernetes.io/projected/eeaab305-c370-45b0-8db9-cf907726dfd5-kube-api-access-swgqf\") pod \"community-operators-xxvts\" (UID: \"eeaab305-c370-45b0-8db9-cf907726dfd5\") " pod="openshift-marketplace/community-operators-xxvts" Nov 21 15:46:07 crc kubenswrapper[4897]: I1121 15:46:07.170099 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eeaab305-c370-45b0-8db9-cf907726dfd5-catalog-content\") pod \"community-operators-xxvts\" (UID: \"eeaab305-c370-45b0-8db9-cf907726dfd5\") " pod="openshift-marketplace/community-operators-xxvts" Nov 21 15:46:07 crc kubenswrapper[4897]: I1121 15:46:07.274945 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-swgqf\" (UniqueName: \"kubernetes.io/projected/eeaab305-c370-45b0-8db9-cf907726dfd5-kube-api-access-swgqf\") pod \"community-operators-xxvts\" (UID: \"eeaab305-c370-45b0-8db9-cf907726dfd5\") " pod="openshift-marketplace/community-operators-xxvts" Nov 21 15:46:07 crc kubenswrapper[4897]: I1121 15:46:07.276122 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eeaab305-c370-45b0-8db9-cf907726dfd5-catalog-content\") pod \"community-operators-xxvts\" (UID: \"eeaab305-c370-45b0-8db9-cf907726dfd5\") " pod="openshift-marketplace/community-operators-xxvts" Nov 21 15:46:07 crc kubenswrapper[4897]: I1121 15:46:07.276635 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eeaab305-c370-45b0-8db9-cf907726dfd5-utilities\") pod \"community-operators-xxvts\" (UID: \"eeaab305-c370-45b0-8db9-cf907726dfd5\") " pod="openshift-marketplace/community-operators-xxvts" Nov 21 15:46:07 crc kubenswrapper[4897]: I1121 15:46:07.276636 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eeaab305-c370-45b0-8db9-cf907726dfd5-catalog-content\") pod \"community-operators-xxvts\" (UID: \"eeaab305-c370-45b0-8db9-cf907726dfd5\") " pod="openshift-marketplace/community-operators-xxvts" Nov 21 15:46:07 crc kubenswrapper[4897]: I1121 15:46:07.277179 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eeaab305-c370-45b0-8db9-cf907726dfd5-utilities\") pod \"community-operators-xxvts\" (UID: \"eeaab305-c370-45b0-8db9-cf907726dfd5\") " pod="openshift-marketplace/community-operators-xxvts" Nov 21 15:46:07 crc kubenswrapper[4897]: I1121 15:46:07.297144 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-swgqf\" (UniqueName: \"kubernetes.io/projected/eeaab305-c370-45b0-8db9-cf907726dfd5-kube-api-access-swgqf\") pod \"community-operators-xxvts\" (UID: \"eeaab305-c370-45b0-8db9-cf907726dfd5\") " pod="openshift-marketplace/community-operators-xxvts" Nov 21 15:46:07 crc kubenswrapper[4897]: I1121 15:46:07.386633 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xxvts" Nov 21 15:46:07 crc kubenswrapper[4897]: I1121 15:46:07.943495 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xxvts"] Nov 21 15:46:08 crc kubenswrapper[4897]: I1121 15:46:08.502809 4897 generic.go:334] "Generic (PLEG): container finished" podID="eeaab305-c370-45b0-8db9-cf907726dfd5" containerID="c4a120c10cec40b43ad7f30d2a72bc893f1477d84d77cd4a4ca19ef3202d9f5e" exitCode=0 Nov 21 15:46:08 crc kubenswrapper[4897]: I1121 15:46:08.502877 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xxvts" event={"ID":"eeaab305-c370-45b0-8db9-cf907726dfd5","Type":"ContainerDied","Data":"c4a120c10cec40b43ad7f30d2a72bc893f1477d84d77cd4a4ca19ef3202d9f5e"} Nov 21 15:46:08 crc kubenswrapper[4897]: I1121 15:46:08.504059 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xxvts" event={"ID":"eeaab305-c370-45b0-8db9-cf907726dfd5","Type":"ContainerStarted","Data":"bf9e54f33ddae0eec74b9dc527c9c615f3527bc0e6cb662a006ce2e5e3e00e87"} Nov 21 15:46:09 crc kubenswrapper[4897]: I1121 15:46:09.517131 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xxvts" event={"ID":"eeaab305-c370-45b0-8db9-cf907726dfd5","Type":"ContainerStarted","Data":"9a9ecb46414469b1dafbdc523ef28c38da30e87ce26b1260200443dc92f93a54"} Nov 21 15:46:11 crc kubenswrapper[4897]: I1121 15:46:11.538330 4897 generic.go:334] "Generic (PLEG): container finished" podID="eeaab305-c370-45b0-8db9-cf907726dfd5" containerID="9a9ecb46414469b1dafbdc523ef28c38da30e87ce26b1260200443dc92f93a54" exitCode=0 Nov 21 15:46:11 crc kubenswrapper[4897]: I1121 15:46:11.538439 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xxvts" event={"ID":"eeaab305-c370-45b0-8db9-cf907726dfd5","Type":"ContainerDied","Data":"9a9ecb46414469b1dafbdc523ef28c38da30e87ce26b1260200443dc92f93a54"} Nov 21 15:46:12 crc kubenswrapper[4897]: I1121 15:46:12.574057 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xxvts" event={"ID":"eeaab305-c370-45b0-8db9-cf907726dfd5","Type":"ContainerStarted","Data":"1afd7b35e88f42fc017daf5686a389f58dd42213de15617d8349b244c6a1b796"} Nov 21 15:46:12 crc kubenswrapper[4897]: I1121 15:46:12.598258 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-xxvts" podStartSLOduration=2.141677144 podStartE2EDuration="5.598238167s" podCreationTimestamp="2025-11-21 15:46:07 +0000 UTC" firstStartedPulling="2025-11-21 15:46:08.504910208 +0000 UTC m=+5845.789503683" lastFinishedPulling="2025-11-21 15:46:11.961471231 +0000 UTC m=+5849.246064706" observedRunningTime="2025-11-21 15:46:12.595664928 +0000 UTC m=+5849.880258403" watchObservedRunningTime="2025-11-21 15:46:12.598238167 +0000 UTC m=+5849.882831632" Nov 21 15:46:17 crc kubenswrapper[4897]: I1121 15:46:17.386795 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-xxvts" Nov 21 15:46:17 crc kubenswrapper[4897]: I1121 15:46:17.387372 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-xxvts" Nov 21 15:46:18 crc kubenswrapper[4897]: I1121 15:46:18.489779 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-xxvts" podUID="eeaab305-c370-45b0-8db9-cf907726dfd5" containerName="registry-server" probeResult="failure" output=< Nov 21 15:46:18 crc kubenswrapper[4897]: timeout: failed to connect service ":50051" within 1s Nov 21 15:46:18 crc kubenswrapper[4897]: > Nov 21 15:46:25 crc kubenswrapper[4897]: I1121 15:46:20.106579 4897 patch_prober.go:28] interesting pod/logging-loki-query-frontend-84558f7c9f-vltt5 container/loki-query-frontend namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.75:3101/loki/api/v1/status/buildinfo\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 21 15:46:25 crc kubenswrapper[4897]: I1121 15:46:20.106926 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-vltt5" podUID="a77cc078-3448-4f4b-86f4-90132ec0fd88" containerName="loki-query-frontend" probeResult="failure" output="Get \"https://10.217.0.75:3101/loki/api/v1/status/buildinfo\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Nov 21 15:46:25 crc kubenswrapper[4897]: I1121 15:46:21.089747 4897 scope.go:117] "RemoveContainer" containerID="9f6ac9e28e3e389a8236d0859923fa122a6eea6c927d5c77a01f623f7524a046" Nov 21 15:46:25 crc kubenswrapper[4897]: E1121 15:46:21.090231 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:46:25 crc kubenswrapper[4897]: I1121 15:46:24.720432 4897 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="c628be92-b88f-4e02-a981-4b8fe6fd7c13" containerName="ceilometer-central-agent" probeResult="failure" output="command timed out" Nov 21 15:46:25 crc kubenswrapper[4897]: I1121 15:46:25.705930 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-gzzl8" podUID="d4c39700-beb8-4272-9f54-1c501332af0d" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.111:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 21 15:46:25 crc kubenswrapper[4897]: I1121 15:46:25.706247 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-kq4ss" podUID="39c184ba-7360-4d31-99f5-239711efdf91" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.104:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 21 15:46:25 crc kubenswrapper[4897]: I1121 15:46:25.706301 4897 patch_prober.go:28] interesting pod/openshift-kube-scheduler-crc container/kube-scheduler namespace/openshift-kube-scheduler: Liveness probe status=failure output="Get \"https://192.168.126.11:10259/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 21 15:46:25 crc kubenswrapper[4897]: I1121 15:46:25.706326 4897 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podUID="3dcd261975c3d6b9a6ad6367fd4facd3" containerName="kube-scheduler" probeResult="failure" output="Get \"https://192.168.126.11:10259/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 21 15:46:25 crc kubenswrapper[4897]: I1121 15:46:25.706365 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/manila-operator-controller-manager-58f887965d-dhc5w" podUID="282cd160-d81e-4e20-866f-dee909546aa2" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.112:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 21 15:46:25 crc kubenswrapper[4897]: I1121 15:46:25.706401 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-2grhg" podUID="77f4c339-9ff1-4943-922d-88d81a73a1f1" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.113:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 21 15:46:25 crc kubenswrapper[4897]: I1121 15:46:25.706915 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-2dnk2" podUID="3b3d91dd-852c-4b65-a2c9-dbc02e79daaa" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.114:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 21 15:46:25 crc kubenswrapper[4897]: I1121 15:46:25.706974 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-bh4rl" podUID="499cc09f-fafe-47c2-87ac-bc00d79e949e" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.115:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 21 15:46:25 crc kubenswrapper[4897]: I1121 15:46:25.707014 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-2sqhj" podUID="81f3265d-f571-47a2-9ad8-b7cdf952d850" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.116:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 21 15:46:25 crc kubenswrapper[4897]: I1121 15:46:25.707050 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-pt5vd" podUID="a7d8ae48-9dd4-40d0-bab5-26e3cff24b20" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.109:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 21 15:46:25 crc kubenswrapper[4897]: I1121 15:46:25.707401 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-phs97" podUID="d1fc2027-5a14-4de3-8c0e-c4ecee66c40b" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.110:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 21 15:46:25 crc kubenswrapper[4897]: I1121 15:46:25.708065 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-smlb7" podUID="4b3dcab3-52d5-4f1d-bde6-2ae9803a0a8f" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.119:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 21 15:46:25 crc kubenswrapper[4897]: I1121 15:46:25.708117 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/swift-operator-controller-manager-d656998f4-p8zts" podUID="a031ee6b-00fb-4795-9e6d-6e21068c405e" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.120:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 21 15:46:25 crc kubenswrapper[4897]: I1121 15:46:25.708155 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/telemetry-operator-controller-manager-77797865fc-9n8nt" podUID="1557d794-d052-44ad-b908-4766d1abe234" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.121:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 21 15:46:25 crc kubenswrapper[4897]: I1121 15:46:25.708190 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-glmx7" podUID="09df9b2c-a507-4a35-b646-a2dc45e1eb71" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.118:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 21 15:46:25 crc kubenswrapper[4897]: I1121 15:46:25.708400 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/test-operator-controller-manager-b4c496f69-9fkd7" podUID="e9edea52-e8de-4e6b-b672-fa71b01d053c" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.122:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 21 15:46:25 crc kubenswrapper[4897]: I1121 15:46:25.708825 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-jrlfk" podUID="40748f20-acda-4ced-ac8b-2e741dae0c66" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.123:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 21 15:46:25 crc kubenswrapper[4897]: I1121 15:46:25.708881 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/openstack-operator-controller-manager-6c5c9b94-7hk8h" podUID="6b47cf09-64d3-4ae5-a7d7-d0125af9d456" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.124:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 21 15:46:28 crc kubenswrapper[4897]: I1121 15:46:28.447341 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-xxvts" podUID="eeaab305-c370-45b0-8db9-cf907726dfd5" containerName="registry-server" probeResult="failure" output=< Nov 21 15:46:28 crc kubenswrapper[4897]: timeout: failed to connect service ":50051" within 1s Nov 21 15:46:28 crc kubenswrapper[4897]: > Nov 21 15:46:33 crc kubenswrapper[4897]: I1121 15:46:33.089031 4897 scope.go:117] "RemoveContainer" containerID="9f6ac9e28e3e389a8236d0859923fa122a6eea6c927d5c77a01f623f7524a046" Nov 21 15:46:33 crc kubenswrapper[4897]: E1121 15:46:33.089854 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:46:38 crc kubenswrapper[4897]: I1121 15:46:38.440625 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-xxvts" podUID="eeaab305-c370-45b0-8db9-cf907726dfd5" containerName="registry-server" probeResult="failure" output=< Nov 21 15:46:38 crc kubenswrapper[4897]: timeout: failed to connect service ":50051" within 1s Nov 21 15:46:38 crc kubenswrapper[4897]: > Nov 21 15:46:44 crc kubenswrapper[4897]: I1121 15:46:44.261904 4897 scope.go:117] "RemoveContainer" containerID="f2872cb69bf99e2310df6dc29eb567d214cb5d0e7e6c0d9c6247b72a36905c39" Nov 21 15:46:44 crc kubenswrapper[4897]: I1121 15:46:44.288189 4897 scope.go:117] "RemoveContainer" containerID="10ad7d280923a05bee6368e831b8cefa05f89e75fb7e07e94ed29b703252bd97" Nov 21 15:46:44 crc kubenswrapper[4897]: I1121 15:46:44.324246 4897 scope.go:117] "RemoveContainer" containerID="c3063e4d1319bd1064fd3400965a617a512f3bdc7748b610bf9f289c17d3c7e6" Nov 21 15:46:47 crc kubenswrapper[4897]: I1121 15:46:47.444566 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-xxvts" Nov 21 15:46:47 crc kubenswrapper[4897]: I1121 15:46:47.503109 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-xxvts" Nov 21 15:46:47 crc kubenswrapper[4897]: I1121 15:46:47.683946 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xxvts"] Nov 21 15:46:48 crc kubenswrapper[4897]: I1121 15:46:48.089839 4897 scope.go:117] "RemoveContainer" containerID="9f6ac9e28e3e389a8236d0859923fa122a6eea6c927d5c77a01f623f7524a046" Nov 21 15:46:48 crc kubenswrapper[4897]: E1121 15:46:48.090243 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:46:48 crc kubenswrapper[4897]: I1121 15:46:48.978184 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-xxvts" podUID="eeaab305-c370-45b0-8db9-cf907726dfd5" containerName="registry-server" containerID="cri-o://1afd7b35e88f42fc017daf5686a389f58dd42213de15617d8349b244c6a1b796" gracePeriod=2 Nov 21 15:46:49 crc kubenswrapper[4897]: I1121 15:46:49.480427 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xxvts" Nov 21 15:46:49 crc kubenswrapper[4897]: I1121 15:46:49.643238 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-swgqf\" (UniqueName: \"kubernetes.io/projected/eeaab305-c370-45b0-8db9-cf907726dfd5-kube-api-access-swgqf\") pod \"eeaab305-c370-45b0-8db9-cf907726dfd5\" (UID: \"eeaab305-c370-45b0-8db9-cf907726dfd5\") " Nov 21 15:46:49 crc kubenswrapper[4897]: I1121 15:46:49.643424 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eeaab305-c370-45b0-8db9-cf907726dfd5-utilities\") pod \"eeaab305-c370-45b0-8db9-cf907726dfd5\" (UID: \"eeaab305-c370-45b0-8db9-cf907726dfd5\") " Nov 21 15:46:49 crc kubenswrapper[4897]: I1121 15:46:49.643674 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eeaab305-c370-45b0-8db9-cf907726dfd5-catalog-content\") pod \"eeaab305-c370-45b0-8db9-cf907726dfd5\" (UID: \"eeaab305-c370-45b0-8db9-cf907726dfd5\") " Nov 21 15:46:49 crc kubenswrapper[4897]: I1121 15:46:49.644210 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eeaab305-c370-45b0-8db9-cf907726dfd5-utilities" (OuterVolumeSpecName: "utilities") pod "eeaab305-c370-45b0-8db9-cf907726dfd5" (UID: "eeaab305-c370-45b0-8db9-cf907726dfd5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 15:46:49 crc kubenswrapper[4897]: I1121 15:46:49.661316 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eeaab305-c370-45b0-8db9-cf907726dfd5-kube-api-access-swgqf" (OuterVolumeSpecName: "kube-api-access-swgqf") pod "eeaab305-c370-45b0-8db9-cf907726dfd5" (UID: "eeaab305-c370-45b0-8db9-cf907726dfd5"). InnerVolumeSpecName "kube-api-access-swgqf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 15:46:49 crc kubenswrapper[4897]: I1121 15:46:49.702923 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eeaab305-c370-45b0-8db9-cf907726dfd5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "eeaab305-c370-45b0-8db9-cf907726dfd5" (UID: "eeaab305-c370-45b0-8db9-cf907726dfd5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 15:46:49 crc kubenswrapper[4897]: I1121 15:46:49.746994 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-swgqf\" (UniqueName: \"kubernetes.io/projected/eeaab305-c370-45b0-8db9-cf907726dfd5-kube-api-access-swgqf\") on node \"crc\" DevicePath \"\"" Nov 21 15:46:49 crc kubenswrapper[4897]: I1121 15:46:49.747035 4897 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eeaab305-c370-45b0-8db9-cf907726dfd5-utilities\") on node \"crc\" DevicePath \"\"" Nov 21 15:46:49 crc kubenswrapper[4897]: I1121 15:46:49.747045 4897 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eeaab305-c370-45b0-8db9-cf907726dfd5-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 21 15:46:49 crc kubenswrapper[4897]: I1121 15:46:49.992338 4897 generic.go:334] "Generic (PLEG): container finished" podID="eeaab305-c370-45b0-8db9-cf907726dfd5" containerID="1afd7b35e88f42fc017daf5686a389f58dd42213de15617d8349b244c6a1b796" exitCode=0 Nov 21 15:46:49 crc kubenswrapper[4897]: I1121 15:46:49.992400 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xxvts" event={"ID":"eeaab305-c370-45b0-8db9-cf907726dfd5","Type":"ContainerDied","Data":"1afd7b35e88f42fc017daf5686a389f58dd42213de15617d8349b244c6a1b796"} Nov 21 15:46:49 crc kubenswrapper[4897]: I1121 15:46:49.992430 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xxvts" Nov 21 15:46:49 crc kubenswrapper[4897]: I1121 15:46:49.992450 4897 scope.go:117] "RemoveContainer" containerID="1afd7b35e88f42fc017daf5686a389f58dd42213de15617d8349b244c6a1b796" Nov 21 15:46:49 crc kubenswrapper[4897]: I1121 15:46:49.992438 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xxvts" event={"ID":"eeaab305-c370-45b0-8db9-cf907726dfd5","Type":"ContainerDied","Data":"bf9e54f33ddae0eec74b9dc527c9c615f3527bc0e6cb662a006ce2e5e3e00e87"} Nov 21 15:46:50 crc kubenswrapper[4897]: I1121 15:46:50.022263 4897 scope.go:117] "RemoveContainer" containerID="9a9ecb46414469b1dafbdc523ef28c38da30e87ce26b1260200443dc92f93a54" Nov 21 15:46:50 crc kubenswrapper[4897]: I1121 15:46:50.034353 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xxvts"] Nov 21 15:46:50 crc kubenswrapper[4897]: I1121 15:46:50.044136 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-xxvts"] Nov 21 15:46:50 crc kubenswrapper[4897]: I1121 15:46:50.053592 4897 scope.go:117] "RemoveContainer" containerID="c4a120c10cec40b43ad7f30d2a72bc893f1477d84d77cd4a4ca19ef3202d9f5e" Nov 21 15:46:50 crc kubenswrapper[4897]: I1121 15:46:50.105390 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eeaab305-c370-45b0-8db9-cf907726dfd5" path="/var/lib/kubelet/pods/eeaab305-c370-45b0-8db9-cf907726dfd5/volumes" Nov 21 15:46:50 crc kubenswrapper[4897]: I1121 15:46:50.110554 4897 scope.go:117] "RemoveContainer" containerID="1afd7b35e88f42fc017daf5686a389f58dd42213de15617d8349b244c6a1b796" Nov 21 15:46:50 crc kubenswrapper[4897]: E1121 15:46:50.110998 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1afd7b35e88f42fc017daf5686a389f58dd42213de15617d8349b244c6a1b796\": container with ID starting with 1afd7b35e88f42fc017daf5686a389f58dd42213de15617d8349b244c6a1b796 not found: ID does not exist" containerID="1afd7b35e88f42fc017daf5686a389f58dd42213de15617d8349b244c6a1b796" Nov 21 15:46:50 crc kubenswrapper[4897]: I1121 15:46:50.111025 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1afd7b35e88f42fc017daf5686a389f58dd42213de15617d8349b244c6a1b796"} err="failed to get container status \"1afd7b35e88f42fc017daf5686a389f58dd42213de15617d8349b244c6a1b796\": rpc error: code = NotFound desc = could not find container \"1afd7b35e88f42fc017daf5686a389f58dd42213de15617d8349b244c6a1b796\": container with ID starting with 1afd7b35e88f42fc017daf5686a389f58dd42213de15617d8349b244c6a1b796 not found: ID does not exist" Nov 21 15:46:50 crc kubenswrapper[4897]: I1121 15:46:50.111049 4897 scope.go:117] "RemoveContainer" containerID="9a9ecb46414469b1dafbdc523ef28c38da30e87ce26b1260200443dc92f93a54" Nov 21 15:46:50 crc kubenswrapper[4897]: E1121 15:46:50.111379 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9a9ecb46414469b1dafbdc523ef28c38da30e87ce26b1260200443dc92f93a54\": container with ID starting with 9a9ecb46414469b1dafbdc523ef28c38da30e87ce26b1260200443dc92f93a54 not found: ID does not exist" containerID="9a9ecb46414469b1dafbdc523ef28c38da30e87ce26b1260200443dc92f93a54" Nov 21 15:46:50 crc kubenswrapper[4897]: I1121 15:46:50.111405 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a9ecb46414469b1dafbdc523ef28c38da30e87ce26b1260200443dc92f93a54"} err="failed to get container status \"9a9ecb46414469b1dafbdc523ef28c38da30e87ce26b1260200443dc92f93a54\": rpc error: code = NotFound desc = could not find container \"9a9ecb46414469b1dafbdc523ef28c38da30e87ce26b1260200443dc92f93a54\": container with ID starting with 9a9ecb46414469b1dafbdc523ef28c38da30e87ce26b1260200443dc92f93a54 not found: ID does not exist" Nov 21 15:46:50 crc kubenswrapper[4897]: I1121 15:46:50.111426 4897 scope.go:117] "RemoveContainer" containerID="c4a120c10cec40b43ad7f30d2a72bc893f1477d84d77cd4a4ca19ef3202d9f5e" Nov 21 15:46:50 crc kubenswrapper[4897]: E1121 15:46:50.111867 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c4a120c10cec40b43ad7f30d2a72bc893f1477d84d77cd4a4ca19ef3202d9f5e\": container with ID starting with c4a120c10cec40b43ad7f30d2a72bc893f1477d84d77cd4a4ca19ef3202d9f5e not found: ID does not exist" containerID="c4a120c10cec40b43ad7f30d2a72bc893f1477d84d77cd4a4ca19ef3202d9f5e" Nov 21 15:46:50 crc kubenswrapper[4897]: I1121 15:46:50.111898 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c4a120c10cec40b43ad7f30d2a72bc893f1477d84d77cd4a4ca19ef3202d9f5e"} err="failed to get container status \"c4a120c10cec40b43ad7f30d2a72bc893f1477d84d77cd4a4ca19ef3202d9f5e\": rpc error: code = NotFound desc = could not find container \"c4a120c10cec40b43ad7f30d2a72bc893f1477d84d77cd4a4ca19ef3202d9f5e\": container with ID starting with c4a120c10cec40b43ad7f30d2a72bc893f1477d84d77cd4a4ca19ef3202d9f5e not found: ID does not exist" Nov 21 15:47:02 crc kubenswrapper[4897]: I1121 15:47:02.089820 4897 scope.go:117] "RemoveContainer" containerID="9f6ac9e28e3e389a8236d0859923fa122a6eea6c927d5c77a01f623f7524a046" Nov 21 15:47:02 crc kubenswrapper[4897]: E1121 15:47:02.090583 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:47:15 crc kubenswrapper[4897]: I1121 15:47:15.090292 4897 scope.go:117] "RemoveContainer" containerID="9f6ac9e28e3e389a8236d0859923fa122a6eea6c927d5c77a01f623f7524a046" Nov 21 15:47:15 crc kubenswrapper[4897]: E1121 15:47:15.091260 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:47:28 crc kubenswrapper[4897]: I1121 15:47:28.089481 4897 scope.go:117] "RemoveContainer" containerID="9f6ac9e28e3e389a8236d0859923fa122a6eea6c927d5c77a01f623f7524a046" Nov 21 15:47:28 crc kubenswrapper[4897]: E1121 15:47:28.090806 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:47:41 crc kubenswrapper[4897]: I1121 15:47:41.089626 4897 scope.go:117] "RemoveContainer" containerID="9f6ac9e28e3e389a8236d0859923fa122a6eea6c927d5c77a01f623f7524a046" Nov 21 15:47:41 crc kubenswrapper[4897]: E1121 15:47:41.090468 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:47:56 crc kubenswrapper[4897]: I1121 15:47:56.088706 4897 scope.go:117] "RemoveContainer" containerID="9f6ac9e28e3e389a8236d0859923fa122a6eea6c927d5c77a01f623f7524a046" Nov 21 15:47:56 crc kubenswrapper[4897]: E1121 15:47:56.089394 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:48:10 crc kubenswrapper[4897]: I1121 15:48:10.089484 4897 scope.go:117] "RemoveContainer" containerID="9f6ac9e28e3e389a8236d0859923fa122a6eea6c927d5c77a01f623f7524a046" Nov 21 15:48:10 crc kubenswrapper[4897]: E1121 15:48:10.090245 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:48:24 crc kubenswrapper[4897]: I1121 15:48:24.098979 4897 scope.go:117] "RemoveContainer" containerID="9f6ac9e28e3e389a8236d0859923fa122a6eea6c927d5c77a01f623f7524a046" Nov 21 15:48:24 crc kubenswrapper[4897]: E1121 15:48:24.100128 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:48:37 crc kubenswrapper[4897]: I1121 15:48:37.090241 4897 scope.go:117] "RemoveContainer" containerID="9f6ac9e28e3e389a8236d0859923fa122a6eea6c927d5c77a01f623f7524a046" Nov 21 15:48:37 crc kubenswrapper[4897]: E1121 15:48:37.092133 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:48:50 crc kubenswrapper[4897]: I1121 15:48:50.089812 4897 scope.go:117] "RemoveContainer" containerID="9f6ac9e28e3e389a8236d0859923fa122a6eea6c927d5c77a01f623f7524a046" Nov 21 15:48:50 crc kubenswrapper[4897]: E1121 15:48:50.090673 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:49:01 crc kubenswrapper[4897]: I1121 15:49:01.089105 4897 scope.go:117] "RemoveContainer" containerID="9f6ac9e28e3e389a8236d0859923fa122a6eea6c927d5c77a01f623f7524a046" Nov 21 15:49:01 crc kubenswrapper[4897]: E1121 15:49:01.090003 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:49:16 crc kubenswrapper[4897]: I1121 15:49:16.089362 4897 scope.go:117] "RemoveContainer" containerID="9f6ac9e28e3e389a8236d0859923fa122a6eea6c927d5c77a01f623f7524a046" Nov 21 15:49:16 crc kubenswrapper[4897]: E1121 15:49:16.090294 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:49:28 crc kubenswrapper[4897]: I1121 15:49:28.090999 4897 scope.go:117] "RemoveContainer" containerID="9f6ac9e28e3e389a8236d0859923fa122a6eea6c927d5c77a01f623f7524a046" Nov 21 15:49:28 crc kubenswrapper[4897]: E1121 15:49:28.091853 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:49:40 crc kubenswrapper[4897]: I1121 15:49:40.090303 4897 scope.go:117] "RemoveContainer" containerID="9f6ac9e28e3e389a8236d0859923fa122a6eea6c927d5c77a01f623f7524a046" Nov 21 15:49:40 crc kubenswrapper[4897]: I1121 15:49:40.903154 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" event={"ID":"e7670227-d280-4847-b882-754429f56b0f","Type":"ContainerStarted","Data":"050e0baf907b19cdd602d0b93f62921dca8e2aa352d875c030b097331cbb3b00"} Nov 21 15:51:30 crc kubenswrapper[4897]: I1121 15:51:30.859954 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-bpsgg"] Nov 21 15:51:30 crc kubenswrapper[4897]: E1121 15:51:30.861027 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eeaab305-c370-45b0-8db9-cf907726dfd5" containerName="registry-server" Nov 21 15:51:30 crc kubenswrapper[4897]: I1121 15:51:30.861045 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="eeaab305-c370-45b0-8db9-cf907726dfd5" containerName="registry-server" Nov 21 15:51:30 crc kubenswrapper[4897]: E1121 15:51:30.861074 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eeaab305-c370-45b0-8db9-cf907726dfd5" containerName="extract-content" Nov 21 15:51:30 crc kubenswrapper[4897]: I1121 15:51:30.861080 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="eeaab305-c370-45b0-8db9-cf907726dfd5" containerName="extract-content" Nov 21 15:51:30 crc kubenswrapper[4897]: E1121 15:51:30.861114 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eeaab305-c370-45b0-8db9-cf907726dfd5" containerName="extract-utilities" Nov 21 15:51:30 crc kubenswrapper[4897]: I1121 15:51:30.861123 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="eeaab305-c370-45b0-8db9-cf907726dfd5" containerName="extract-utilities" Nov 21 15:51:30 crc kubenswrapper[4897]: I1121 15:51:30.861343 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="eeaab305-c370-45b0-8db9-cf907726dfd5" containerName="registry-server" Nov 21 15:51:30 crc kubenswrapper[4897]: I1121 15:51:30.863371 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bpsgg" Nov 21 15:51:30 crc kubenswrapper[4897]: I1121 15:51:30.876955 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bpsgg"] Nov 21 15:51:30 crc kubenswrapper[4897]: I1121 15:51:30.922420 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hlpvn\" (UniqueName: \"kubernetes.io/projected/7ad1e6a8-2797-4538-974d-f32d2f359e09-kube-api-access-hlpvn\") pod \"certified-operators-bpsgg\" (UID: \"7ad1e6a8-2797-4538-974d-f32d2f359e09\") " pod="openshift-marketplace/certified-operators-bpsgg" Nov 21 15:51:30 crc kubenswrapper[4897]: I1121 15:51:30.922556 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ad1e6a8-2797-4538-974d-f32d2f359e09-utilities\") pod \"certified-operators-bpsgg\" (UID: \"7ad1e6a8-2797-4538-974d-f32d2f359e09\") " pod="openshift-marketplace/certified-operators-bpsgg" Nov 21 15:51:30 crc kubenswrapper[4897]: I1121 15:51:30.922685 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ad1e6a8-2797-4538-974d-f32d2f359e09-catalog-content\") pod \"certified-operators-bpsgg\" (UID: \"7ad1e6a8-2797-4538-974d-f32d2f359e09\") " pod="openshift-marketplace/certified-operators-bpsgg" Nov 21 15:51:31 crc kubenswrapper[4897]: I1121 15:51:31.024561 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ad1e6a8-2797-4538-974d-f32d2f359e09-catalog-content\") pod \"certified-operators-bpsgg\" (UID: \"7ad1e6a8-2797-4538-974d-f32d2f359e09\") " pod="openshift-marketplace/certified-operators-bpsgg" Nov 21 15:51:31 crc kubenswrapper[4897]: I1121 15:51:31.024931 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hlpvn\" (UniqueName: \"kubernetes.io/projected/7ad1e6a8-2797-4538-974d-f32d2f359e09-kube-api-access-hlpvn\") pod \"certified-operators-bpsgg\" (UID: \"7ad1e6a8-2797-4538-974d-f32d2f359e09\") " pod="openshift-marketplace/certified-operators-bpsgg" Nov 21 15:51:31 crc kubenswrapper[4897]: I1121 15:51:31.025000 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ad1e6a8-2797-4538-974d-f32d2f359e09-utilities\") pod \"certified-operators-bpsgg\" (UID: \"7ad1e6a8-2797-4538-974d-f32d2f359e09\") " pod="openshift-marketplace/certified-operators-bpsgg" Nov 21 15:51:31 crc kubenswrapper[4897]: I1121 15:51:31.025101 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ad1e6a8-2797-4538-974d-f32d2f359e09-catalog-content\") pod \"certified-operators-bpsgg\" (UID: \"7ad1e6a8-2797-4538-974d-f32d2f359e09\") " pod="openshift-marketplace/certified-operators-bpsgg" Nov 21 15:51:31 crc kubenswrapper[4897]: I1121 15:51:31.025570 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ad1e6a8-2797-4538-974d-f32d2f359e09-utilities\") pod \"certified-operators-bpsgg\" (UID: \"7ad1e6a8-2797-4538-974d-f32d2f359e09\") " pod="openshift-marketplace/certified-operators-bpsgg" Nov 21 15:51:31 crc kubenswrapper[4897]: I1121 15:51:31.050708 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hlpvn\" (UniqueName: \"kubernetes.io/projected/7ad1e6a8-2797-4538-974d-f32d2f359e09-kube-api-access-hlpvn\") pod \"certified-operators-bpsgg\" (UID: \"7ad1e6a8-2797-4538-974d-f32d2f359e09\") " pod="openshift-marketplace/certified-operators-bpsgg" Nov 21 15:51:31 crc kubenswrapper[4897]: I1121 15:51:31.187120 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bpsgg" Nov 21 15:51:31 crc kubenswrapper[4897]: I1121 15:51:31.855017 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bpsgg"] Nov 21 15:51:32 crc kubenswrapper[4897]: I1121 15:51:32.132894 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bpsgg" event={"ID":"7ad1e6a8-2797-4538-974d-f32d2f359e09","Type":"ContainerStarted","Data":"b2f09e6a8d1e06de82b357832e0e4078b5230ef59481a3fa86e630b2f49419ad"} Nov 21 15:51:33 crc kubenswrapper[4897]: I1121 15:51:33.150825 4897 generic.go:334] "Generic (PLEG): container finished" podID="7ad1e6a8-2797-4538-974d-f32d2f359e09" containerID="337abc5aa074d461f045cc9d6bdae0fed0961878059bed94c65c7673756966d1" exitCode=0 Nov 21 15:51:33 crc kubenswrapper[4897]: I1121 15:51:33.150945 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bpsgg" event={"ID":"7ad1e6a8-2797-4538-974d-f32d2f359e09","Type":"ContainerDied","Data":"337abc5aa074d461f045cc9d6bdae0fed0961878059bed94c65c7673756966d1"} Nov 21 15:51:33 crc kubenswrapper[4897]: I1121 15:51:33.155215 4897 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 21 15:51:35 crc kubenswrapper[4897]: I1121 15:51:35.171945 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bpsgg" event={"ID":"7ad1e6a8-2797-4538-974d-f32d2f359e09","Type":"ContainerStarted","Data":"8eb5b06ab6e8f4259b558f138962f3cc0782206dc083313c76840d8171082470"} Nov 21 15:51:39 crc kubenswrapper[4897]: I1121 15:51:39.215008 4897 generic.go:334] "Generic (PLEG): container finished" podID="7ad1e6a8-2797-4538-974d-f32d2f359e09" containerID="8eb5b06ab6e8f4259b558f138962f3cc0782206dc083313c76840d8171082470" exitCode=0 Nov 21 15:51:39 crc kubenswrapper[4897]: I1121 15:51:39.215140 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bpsgg" event={"ID":"7ad1e6a8-2797-4538-974d-f32d2f359e09","Type":"ContainerDied","Data":"8eb5b06ab6e8f4259b558f138962f3cc0782206dc083313c76840d8171082470"} Nov 21 15:51:40 crc kubenswrapper[4897]: I1121 15:51:40.228942 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bpsgg" event={"ID":"7ad1e6a8-2797-4538-974d-f32d2f359e09","Type":"ContainerStarted","Data":"f93f01dfe2c108568a1ffc2a92c102b704943ee9fd7edf99967fcbc2e8ced39b"} Nov 21 15:51:40 crc kubenswrapper[4897]: I1121 15:51:40.254953 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-bpsgg" podStartSLOduration=3.577254219 podStartE2EDuration="10.254928069s" podCreationTimestamp="2025-11-21 15:51:30 +0000 UTC" firstStartedPulling="2025-11-21 15:51:33.154931195 +0000 UTC m=+6170.439524670" lastFinishedPulling="2025-11-21 15:51:39.832605035 +0000 UTC m=+6177.117198520" observedRunningTime="2025-11-21 15:51:40.245461394 +0000 UTC m=+6177.530054859" watchObservedRunningTime="2025-11-21 15:51:40.254928069 +0000 UTC m=+6177.539521554" Nov 21 15:51:41 crc kubenswrapper[4897]: I1121 15:51:41.187971 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-bpsgg" Nov 21 15:51:41 crc kubenswrapper[4897]: I1121 15:51:41.188489 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-bpsgg" Nov 21 15:51:41 crc kubenswrapper[4897]: I1121 15:51:41.514080 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-ll8pq"] Nov 21 15:51:41 crc kubenswrapper[4897]: I1121 15:51:41.516828 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ll8pq" Nov 21 15:51:41 crc kubenswrapper[4897]: I1121 15:51:41.539227 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ll8pq"] Nov 21 15:51:41 crc kubenswrapper[4897]: I1121 15:51:41.698098 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc2009f9-78da-4ba6-8af1-ceee9d597068-catalog-content\") pod \"redhat-operators-ll8pq\" (UID: \"fc2009f9-78da-4ba6-8af1-ceee9d597068\") " pod="openshift-marketplace/redhat-operators-ll8pq" Nov 21 15:51:41 crc kubenswrapper[4897]: I1121 15:51:41.698460 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vsjjt\" (UniqueName: \"kubernetes.io/projected/fc2009f9-78da-4ba6-8af1-ceee9d597068-kube-api-access-vsjjt\") pod \"redhat-operators-ll8pq\" (UID: \"fc2009f9-78da-4ba6-8af1-ceee9d597068\") " pod="openshift-marketplace/redhat-operators-ll8pq" Nov 21 15:51:41 crc kubenswrapper[4897]: I1121 15:51:41.698738 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc2009f9-78da-4ba6-8af1-ceee9d597068-utilities\") pod \"redhat-operators-ll8pq\" (UID: \"fc2009f9-78da-4ba6-8af1-ceee9d597068\") " pod="openshift-marketplace/redhat-operators-ll8pq" Nov 21 15:51:41 crc kubenswrapper[4897]: I1121 15:51:41.802457 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vsjjt\" (UniqueName: \"kubernetes.io/projected/fc2009f9-78da-4ba6-8af1-ceee9d597068-kube-api-access-vsjjt\") pod \"redhat-operators-ll8pq\" (UID: \"fc2009f9-78da-4ba6-8af1-ceee9d597068\") " pod="openshift-marketplace/redhat-operators-ll8pq" Nov 21 15:51:41 crc kubenswrapper[4897]: I1121 15:51:41.802649 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc2009f9-78da-4ba6-8af1-ceee9d597068-utilities\") pod \"redhat-operators-ll8pq\" (UID: \"fc2009f9-78da-4ba6-8af1-ceee9d597068\") " pod="openshift-marketplace/redhat-operators-ll8pq" Nov 21 15:51:41 crc kubenswrapper[4897]: I1121 15:51:41.802711 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc2009f9-78da-4ba6-8af1-ceee9d597068-catalog-content\") pod \"redhat-operators-ll8pq\" (UID: \"fc2009f9-78da-4ba6-8af1-ceee9d597068\") " pod="openshift-marketplace/redhat-operators-ll8pq" Nov 21 15:51:41 crc kubenswrapper[4897]: I1121 15:51:41.804275 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc2009f9-78da-4ba6-8af1-ceee9d597068-catalog-content\") pod \"redhat-operators-ll8pq\" (UID: \"fc2009f9-78da-4ba6-8af1-ceee9d597068\") " pod="openshift-marketplace/redhat-operators-ll8pq" Nov 21 15:51:41 crc kubenswrapper[4897]: I1121 15:51:41.804442 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc2009f9-78da-4ba6-8af1-ceee9d597068-utilities\") pod \"redhat-operators-ll8pq\" (UID: \"fc2009f9-78da-4ba6-8af1-ceee9d597068\") " pod="openshift-marketplace/redhat-operators-ll8pq" Nov 21 15:51:41 crc kubenswrapper[4897]: I1121 15:51:41.828053 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vsjjt\" (UniqueName: \"kubernetes.io/projected/fc2009f9-78da-4ba6-8af1-ceee9d597068-kube-api-access-vsjjt\") pod \"redhat-operators-ll8pq\" (UID: \"fc2009f9-78da-4ba6-8af1-ceee9d597068\") " pod="openshift-marketplace/redhat-operators-ll8pq" Nov 21 15:51:41 crc kubenswrapper[4897]: I1121 15:51:41.853838 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ll8pq" Nov 21 15:51:42 crc kubenswrapper[4897]: I1121 15:51:42.246885 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-bpsgg" podUID="7ad1e6a8-2797-4538-974d-f32d2f359e09" containerName="registry-server" probeResult="failure" output=< Nov 21 15:51:42 crc kubenswrapper[4897]: timeout: failed to connect service ":50051" within 1s Nov 21 15:51:42 crc kubenswrapper[4897]: > Nov 21 15:51:42 crc kubenswrapper[4897]: I1121 15:51:42.440832 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ll8pq"] Nov 21 15:51:42 crc kubenswrapper[4897]: W1121 15:51:42.448138 4897 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfc2009f9_78da_4ba6_8af1_ceee9d597068.slice/crio-3463c58f8e828b61c4767026ac8bc6e2ff4bcb35cf6a867fc9ab6f377dbba6c3 WatchSource:0}: Error finding container 3463c58f8e828b61c4767026ac8bc6e2ff4bcb35cf6a867fc9ab6f377dbba6c3: Status 404 returned error can't find the container with id 3463c58f8e828b61c4767026ac8bc6e2ff4bcb35cf6a867fc9ab6f377dbba6c3 Nov 21 15:51:43 crc kubenswrapper[4897]: I1121 15:51:43.302330 4897 generic.go:334] "Generic (PLEG): container finished" podID="fc2009f9-78da-4ba6-8af1-ceee9d597068" containerID="69ee742e986b3cefbadcfe1b7dcef2141e5828e991b371095119891e29f924e1" exitCode=0 Nov 21 15:51:43 crc kubenswrapper[4897]: I1121 15:51:43.302436 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ll8pq" event={"ID":"fc2009f9-78da-4ba6-8af1-ceee9d597068","Type":"ContainerDied","Data":"69ee742e986b3cefbadcfe1b7dcef2141e5828e991b371095119891e29f924e1"} Nov 21 15:51:43 crc kubenswrapper[4897]: I1121 15:51:43.302727 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ll8pq" event={"ID":"fc2009f9-78da-4ba6-8af1-ceee9d597068","Type":"ContainerStarted","Data":"3463c58f8e828b61c4767026ac8bc6e2ff4bcb35cf6a867fc9ab6f377dbba6c3"} Nov 21 15:51:44 crc kubenswrapper[4897]: I1121 15:51:44.319052 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ll8pq" event={"ID":"fc2009f9-78da-4ba6-8af1-ceee9d597068","Type":"ContainerStarted","Data":"e368c4d05068b8b7ce809473ba403753947becad88908615236570f9a9ee7aab"} Nov 21 15:51:51 crc kubenswrapper[4897]: I1121 15:51:51.237102 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-bpsgg" Nov 21 15:51:51 crc kubenswrapper[4897]: I1121 15:51:51.298753 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-bpsgg" Nov 21 15:51:51 crc kubenswrapper[4897]: I1121 15:51:51.415566 4897 generic.go:334] "Generic (PLEG): container finished" podID="fc2009f9-78da-4ba6-8af1-ceee9d597068" containerID="e368c4d05068b8b7ce809473ba403753947becad88908615236570f9a9ee7aab" exitCode=0 Nov 21 15:51:51 crc kubenswrapper[4897]: I1121 15:51:51.415615 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ll8pq" event={"ID":"fc2009f9-78da-4ba6-8af1-ceee9d597068","Type":"ContainerDied","Data":"e368c4d05068b8b7ce809473ba403753947becad88908615236570f9a9ee7aab"} Nov 21 15:51:51 crc kubenswrapper[4897]: I1121 15:51:51.503956 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-bpsgg"] Nov 21 15:51:52 crc kubenswrapper[4897]: I1121 15:51:52.452666 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-bpsgg" podUID="7ad1e6a8-2797-4538-974d-f32d2f359e09" containerName="registry-server" containerID="cri-o://f93f01dfe2c108568a1ffc2a92c102b704943ee9fd7edf99967fcbc2e8ced39b" gracePeriod=2 Nov 21 15:51:52 crc kubenswrapper[4897]: I1121 15:51:52.454267 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ll8pq" event={"ID":"fc2009f9-78da-4ba6-8af1-ceee9d597068","Type":"ContainerStarted","Data":"39bba0429418b078fabf1e942ea403bf78858552691deaee8ecb5ee739eef389"} Nov 21 15:51:52 crc kubenswrapper[4897]: I1121 15:51:52.485375 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-ll8pq" podStartSLOduration=2.777847326 podStartE2EDuration="11.485354571s" podCreationTimestamp="2025-11-21 15:51:41 +0000 UTC" firstStartedPulling="2025-11-21 15:51:43.306122777 +0000 UTC m=+6180.590716252" lastFinishedPulling="2025-11-21 15:51:52.013630022 +0000 UTC m=+6189.298223497" observedRunningTime="2025-11-21 15:51:52.476034529 +0000 UTC m=+6189.760628014" watchObservedRunningTime="2025-11-21 15:51:52.485354571 +0000 UTC m=+6189.769948046" Nov 21 15:51:52 crc kubenswrapper[4897]: I1121 15:51:52.960066 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bpsgg" Nov 21 15:51:53 crc kubenswrapper[4897]: I1121 15:51:53.121358 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ad1e6a8-2797-4538-974d-f32d2f359e09-utilities\") pod \"7ad1e6a8-2797-4538-974d-f32d2f359e09\" (UID: \"7ad1e6a8-2797-4538-974d-f32d2f359e09\") " Nov 21 15:51:53 crc kubenswrapper[4897]: I1121 15:51:53.121428 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hlpvn\" (UniqueName: \"kubernetes.io/projected/7ad1e6a8-2797-4538-974d-f32d2f359e09-kube-api-access-hlpvn\") pod \"7ad1e6a8-2797-4538-974d-f32d2f359e09\" (UID: \"7ad1e6a8-2797-4538-974d-f32d2f359e09\") " Nov 21 15:51:53 crc kubenswrapper[4897]: I1121 15:51:53.121634 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ad1e6a8-2797-4538-974d-f32d2f359e09-catalog-content\") pod \"7ad1e6a8-2797-4538-974d-f32d2f359e09\" (UID: \"7ad1e6a8-2797-4538-974d-f32d2f359e09\") " Nov 21 15:51:53 crc kubenswrapper[4897]: I1121 15:51:53.122060 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7ad1e6a8-2797-4538-974d-f32d2f359e09-utilities" (OuterVolumeSpecName: "utilities") pod "7ad1e6a8-2797-4538-974d-f32d2f359e09" (UID: "7ad1e6a8-2797-4538-974d-f32d2f359e09"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 15:51:53 crc kubenswrapper[4897]: I1121 15:51:53.122463 4897 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ad1e6a8-2797-4538-974d-f32d2f359e09-utilities\") on node \"crc\" DevicePath \"\"" Nov 21 15:51:53 crc kubenswrapper[4897]: I1121 15:51:53.130261 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ad1e6a8-2797-4538-974d-f32d2f359e09-kube-api-access-hlpvn" (OuterVolumeSpecName: "kube-api-access-hlpvn") pod "7ad1e6a8-2797-4538-974d-f32d2f359e09" (UID: "7ad1e6a8-2797-4538-974d-f32d2f359e09"). InnerVolumeSpecName "kube-api-access-hlpvn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 15:51:53 crc kubenswrapper[4897]: I1121 15:51:53.161823 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7ad1e6a8-2797-4538-974d-f32d2f359e09-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7ad1e6a8-2797-4538-974d-f32d2f359e09" (UID: "7ad1e6a8-2797-4538-974d-f32d2f359e09"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 15:51:53 crc kubenswrapper[4897]: I1121 15:51:53.224657 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hlpvn\" (UniqueName: \"kubernetes.io/projected/7ad1e6a8-2797-4538-974d-f32d2f359e09-kube-api-access-hlpvn\") on node \"crc\" DevicePath \"\"" Nov 21 15:51:53 crc kubenswrapper[4897]: I1121 15:51:53.224690 4897 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ad1e6a8-2797-4538-974d-f32d2f359e09-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 21 15:51:53 crc kubenswrapper[4897]: I1121 15:51:53.465921 4897 generic.go:334] "Generic (PLEG): container finished" podID="7ad1e6a8-2797-4538-974d-f32d2f359e09" containerID="f93f01dfe2c108568a1ffc2a92c102b704943ee9fd7edf99967fcbc2e8ced39b" exitCode=0 Nov 21 15:51:53 crc kubenswrapper[4897]: I1121 15:51:53.465985 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bpsgg" event={"ID":"7ad1e6a8-2797-4538-974d-f32d2f359e09","Type":"ContainerDied","Data":"f93f01dfe2c108568a1ffc2a92c102b704943ee9fd7edf99967fcbc2e8ced39b"} Nov 21 15:51:53 crc kubenswrapper[4897]: I1121 15:51:53.466013 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bpsgg" event={"ID":"7ad1e6a8-2797-4538-974d-f32d2f359e09","Type":"ContainerDied","Data":"b2f09e6a8d1e06de82b357832e0e4078b5230ef59481a3fa86e630b2f49419ad"} Nov 21 15:51:53 crc kubenswrapper[4897]: I1121 15:51:53.466020 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bpsgg" Nov 21 15:51:53 crc kubenswrapper[4897]: I1121 15:51:53.466030 4897 scope.go:117] "RemoveContainer" containerID="f93f01dfe2c108568a1ffc2a92c102b704943ee9fd7edf99967fcbc2e8ced39b" Nov 21 15:51:53 crc kubenswrapper[4897]: I1121 15:51:53.502407 4897 scope.go:117] "RemoveContainer" containerID="8eb5b06ab6e8f4259b558f138962f3cc0782206dc083313c76840d8171082470" Nov 21 15:51:53 crc kubenswrapper[4897]: I1121 15:51:53.508908 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-bpsgg"] Nov 21 15:51:53 crc kubenswrapper[4897]: I1121 15:51:53.518971 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-bpsgg"] Nov 21 15:51:53 crc kubenswrapper[4897]: I1121 15:51:53.522139 4897 scope.go:117] "RemoveContainer" containerID="337abc5aa074d461f045cc9d6bdae0fed0961878059bed94c65c7673756966d1" Nov 21 15:51:53 crc kubenswrapper[4897]: I1121 15:51:53.600402 4897 scope.go:117] "RemoveContainer" containerID="f93f01dfe2c108568a1ffc2a92c102b704943ee9fd7edf99967fcbc2e8ced39b" Nov 21 15:51:53 crc kubenswrapper[4897]: E1121 15:51:53.601010 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f93f01dfe2c108568a1ffc2a92c102b704943ee9fd7edf99967fcbc2e8ced39b\": container with ID starting with f93f01dfe2c108568a1ffc2a92c102b704943ee9fd7edf99967fcbc2e8ced39b not found: ID does not exist" containerID="f93f01dfe2c108568a1ffc2a92c102b704943ee9fd7edf99967fcbc2e8ced39b" Nov 21 15:51:53 crc kubenswrapper[4897]: I1121 15:51:53.601057 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f93f01dfe2c108568a1ffc2a92c102b704943ee9fd7edf99967fcbc2e8ced39b"} err="failed to get container status \"f93f01dfe2c108568a1ffc2a92c102b704943ee9fd7edf99967fcbc2e8ced39b\": rpc error: code = NotFound desc = could not find container \"f93f01dfe2c108568a1ffc2a92c102b704943ee9fd7edf99967fcbc2e8ced39b\": container with ID starting with f93f01dfe2c108568a1ffc2a92c102b704943ee9fd7edf99967fcbc2e8ced39b not found: ID does not exist" Nov 21 15:51:53 crc kubenswrapper[4897]: I1121 15:51:53.601084 4897 scope.go:117] "RemoveContainer" containerID="8eb5b06ab6e8f4259b558f138962f3cc0782206dc083313c76840d8171082470" Nov 21 15:51:53 crc kubenswrapper[4897]: E1121 15:51:53.601538 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8eb5b06ab6e8f4259b558f138962f3cc0782206dc083313c76840d8171082470\": container with ID starting with 8eb5b06ab6e8f4259b558f138962f3cc0782206dc083313c76840d8171082470 not found: ID does not exist" containerID="8eb5b06ab6e8f4259b558f138962f3cc0782206dc083313c76840d8171082470" Nov 21 15:51:53 crc kubenswrapper[4897]: I1121 15:51:53.601596 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8eb5b06ab6e8f4259b558f138962f3cc0782206dc083313c76840d8171082470"} err="failed to get container status \"8eb5b06ab6e8f4259b558f138962f3cc0782206dc083313c76840d8171082470\": rpc error: code = NotFound desc = could not find container \"8eb5b06ab6e8f4259b558f138962f3cc0782206dc083313c76840d8171082470\": container with ID starting with 8eb5b06ab6e8f4259b558f138962f3cc0782206dc083313c76840d8171082470 not found: ID does not exist" Nov 21 15:51:53 crc kubenswrapper[4897]: I1121 15:51:53.601625 4897 scope.go:117] "RemoveContainer" containerID="337abc5aa074d461f045cc9d6bdae0fed0961878059bed94c65c7673756966d1" Nov 21 15:51:53 crc kubenswrapper[4897]: E1121 15:51:53.601963 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"337abc5aa074d461f045cc9d6bdae0fed0961878059bed94c65c7673756966d1\": container with ID starting with 337abc5aa074d461f045cc9d6bdae0fed0961878059bed94c65c7673756966d1 not found: ID does not exist" containerID="337abc5aa074d461f045cc9d6bdae0fed0961878059bed94c65c7673756966d1" Nov 21 15:51:53 crc kubenswrapper[4897]: I1121 15:51:53.601994 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"337abc5aa074d461f045cc9d6bdae0fed0961878059bed94c65c7673756966d1"} err="failed to get container status \"337abc5aa074d461f045cc9d6bdae0fed0961878059bed94c65c7673756966d1\": rpc error: code = NotFound desc = could not find container \"337abc5aa074d461f045cc9d6bdae0fed0961878059bed94c65c7673756966d1\": container with ID starting with 337abc5aa074d461f045cc9d6bdae0fed0961878059bed94c65c7673756966d1 not found: ID does not exist" Nov 21 15:51:54 crc kubenswrapper[4897]: I1121 15:51:54.111756 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ad1e6a8-2797-4538-974d-f32d2f359e09" path="/var/lib/kubelet/pods/7ad1e6a8-2797-4538-974d-f32d2f359e09/volumes" Nov 21 15:52:01 crc kubenswrapper[4897]: I1121 15:52:01.854652 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-ll8pq" Nov 21 15:52:01 crc kubenswrapper[4897]: I1121 15:52:01.855194 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-ll8pq" Nov 21 15:52:02 crc kubenswrapper[4897]: I1121 15:52:02.909037 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-ll8pq" podUID="fc2009f9-78da-4ba6-8af1-ceee9d597068" containerName="registry-server" probeResult="failure" output=< Nov 21 15:52:02 crc kubenswrapper[4897]: timeout: failed to connect service ":50051" within 1s Nov 21 15:52:02 crc kubenswrapper[4897]: > Nov 21 15:52:04 crc kubenswrapper[4897]: I1121 15:52:04.370433 4897 patch_prober.go:28] interesting pod/machine-config-daemon-krv5b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 21 15:52:04 crc kubenswrapper[4897]: I1121 15:52:04.370963 4897 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 21 15:52:12 crc kubenswrapper[4897]: I1121 15:52:12.901948 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-ll8pq" podUID="fc2009f9-78da-4ba6-8af1-ceee9d597068" containerName="registry-server" probeResult="failure" output=< Nov 21 15:52:12 crc kubenswrapper[4897]: timeout: failed to connect service ":50051" within 1s Nov 21 15:52:12 crc kubenswrapper[4897]: > Nov 21 15:52:22 crc kubenswrapper[4897]: I1121 15:52:22.897911 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-ll8pq" podUID="fc2009f9-78da-4ba6-8af1-ceee9d597068" containerName="registry-server" probeResult="failure" output=< Nov 21 15:52:22 crc kubenswrapper[4897]: timeout: failed to connect service ":50051" within 1s Nov 21 15:52:22 crc kubenswrapper[4897]: > Nov 21 15:52:31 crc kubenswrapper[4897]: I1121 15:52:31.921135 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-ll8pq" Nov 21 15:52:31 crc kubenswrapper[4897]: I1121 15:52:31.974020 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-ll8pq" Nov 21 15:52:32 crc kubenswrapper[4897]: I1121 15:52:32.171600 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-ll8pq"] Nov 21 15:52:33 crc kubenswrapper[4897]: I1121 15:52:33.925483 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-ll8pq" podUID="fc2009f9-78da-4ba6-8af1-ceee9d597068" containerName="registry-server" containerID="cri-o://39bba0429418b078fabf1e942ea403bf78858552691deaee8ecb5ee739eef389" gracePeriod=2 Nov 21 15:52:34 crc kubenswrapper[4897]: I1121 15:52:34.371290 4897 patch_prober.go:28] interesting pod/machine-config-daemon-krv5b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 21 15:52:34 crc kubenswrapper[4897]: I1121 15:52:34.371655 4897 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 21 15:52:34 crc kubenswrapper[4897]: I1121 15:52:34.535990 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ll8pq" Nov 21 15:52:34 crc kubenswrapper[4897]: I1121 15:52:34.697798 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc2009f9-78da-4ba6-8af1-ceee9d597068-catalog-content\") pod \"fc2009f9-78da-4ba6-8af1-ceee9d597068\" (UID: \"fc2009f9-78da-4ba6-8af1-ceee9d597068\") " Nov 21 15:52:34 crc kubenswrapper[4897]: I1121 15:52:34.698190 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vsjjt\" (UniqueName: \"kubernetes.io/projected/fc2009f9-78da-4ba6-8af1-ceee9d597068-kube-api-access-vsjjt\") pod \"fc2009f9-78da-4ba6-8af1-ceee9d597068\" (UID: \"fc2009f9-78da-4ba6-8af1-ceee9d597068\") " Nov 21 15:52:34 crc kubenswrapper[4897]: I1121 15:52:34.698334 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc2009f9-78da-4ba6-8af1-ceee9d597068-utilities\") pod \"fc2009f9-78da-4ba6-8af1-ceee9d597068\" (UID: \"fc2009f9-78da-4ba6-8af1-ceee9d597068\") " Nov 21 15:52:34 crc kubenswrapper[4897]: I1121 15:52:34.703075 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fc2009f9-78da-4ba6-8af1-ceee9d597068-utilities" (OuterVolumeSpecName: "utilities") pod "fc2009f9-78da-4ba6-8af1-ceee9d597068" (UID: "fc2009f9-78da-4ba6-8af1-ceee9d597068"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 15:52:34 crc kubenswrapper[4897]: I1121 15:52:34.705826 4897 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc2009f9-78da-4ba6-8af1-ceee9d597068-utilities\") on node \"crc\" DevicePath \"\"" Nov 21 15:52:34 crc kubenswrapper[4897]: I1121 15:52:34.726332 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc2009f9-78da-4ba6-8af1-ceee9d597068-kube-api-access-vsjjt" (OuterVolumeSpecName: "kube-api-access-vsjjt") pod "fc2009f9-78da-4ba6-8af1-ceee9d597068" (UID: "fc2009f9-78da-4ba6-8af1-ceee9d597068"). InnerVolumeSpecName "kube-api-access-vsjjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 15:52:34 crc kubenswrapper[4897]: I1121 15:52:34.800757 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fc2009f9-78da-4ba6-8af1-ceee9d597068-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fc2009f9-78da-4ba6-8af1-ceee9d597068" (UID: "fc2009f9-78da-4ba6-8af1-ceee9d597068"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 15:52:34 crc kubenswrapper[4897]: I1121 15:52:34.808004 4897 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc2009f9-78da-4ba6-8af1-ceee9d597068-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 21 15:52:34 crc kubenswrapper[4897]: I1121 15:52:34.808061 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vsjjt\" (UniqueName: \"kubernetes.io/projected/fc2009f9-78da-4ba6-8af1-ceee9d597068-kube-api-access-vsjjt\") on node \"crc\" DevicePath \"\"" Nov 21 15:52:34 crc kubenswrapper[4897]: I1121 15:52:34.938070 4897 generic.go:334] "Generic (PLEG): container finished" podID="fc2009f9-78da-4ba6-8af1-ceee9d597068" containerID="39bba0429418b078fabf1e942ea403bf78858552691deaee8ecb5ee739eef389" exitCode=0 Nov 21 15:52:34 crc kubenswrapper[4897]: I1121 15:52:34.938136 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ll8pq" event={"ID":"fc2009f9-78da-4ba6-8af1-ceee9d597068","Type":"ContainerDied","Data":"39bba0429418b078fabf1e942ea403bf78858552691deaee8ecb5ee739eef389"} Nov 21 15:52:34 crc kubenswrapper[4897]: I1121 15:52:34.938169 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ll8pq" event={"ID":"fc2009f9-78da-4ba6-8af1-ceee9d597068","Type":"ContainerDied","Data":"3463c58f8e828b61c4767026ac8bc6e2ff4bcb35cf6a867fc9ab6f377dbba6c3"} Nov 21 15:52:34 crc kubenswrapper[4897]: I1121 15:52:34.938194 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ll8pq" Nov 21 15:52:34 crc kubenswrapper[4897]: I1121 15:52:34.938206 4897 scope.go:117] "RemoveContainer" containerID="39bba0429418b078fabf1e942ea403bf78858552691deaee8ecb5ee739eef389" Nov 21 15:52:34 crc kubenswrapper[4897]: I1121 15:52:34.966190 4897 scope.go:117] "RemoveContainer" containerID="e368c4d05068b8b7ce809473ba403753947becad88908615236570f9a9ee7aab" Nov 21 15:52:34 crc kubenswrapper[4897]: I1121 15:52:34.972419 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-ll8pq"] Nov 21 15:52:34 crc kubenswrapper[4897]: I1121 15:52:34.984404 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-ll8pq"] Nov 21 15:52:35 crc kubenswrapper[4897]: I1121 15:52:35.008412 4897 scope.go:117] "RemoveContainer" containerID="69ee742e986b3cefbadcfe1b7dcef2141e5828e991b371095119891e29f924e1" Nov 21 15:52:35 crc kubenswrapper[4897]: I1121 15:52:35.045592 4897 scope.go:117] "RemoveContainer" containerID="39bba0429418b078fabf1e942ea403bf78858552691deaee8ecb5ee739eef389" Nov 21 15:52:35 crc kubenswrapper[4897]: E1121 15:52:35.046045 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"39bba0429418b078fabf1e942ea403bf78858552691deaee8ecb5ee739eef389\": container with ID starting with 39bba0429418b078fabf1e942ea403bf78858552691deaee8ecb5ee739eef389 not found: ID does not exist" containerID="39bba0429418b078fabf1e942ea403bf78858552691deaee8ecb5ee739eef389" Nov 21 15:52:35 crc kubenswrapper[4897]: I1121 15:52:35.046093 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39bba0429418b078fabf1e942ea403bf78858552691deaee8ecb5ee739eef389"} err="failed to get container status \"39bba0429418b078fabf1e942ea403bf78858552691deaee8ecb5ee739eef389\": rpc error: code = NotFound desc = could not find container \"39bba0429418b078fabf1e942ea403bf78858552691deaee8ecb5ee739eef389\": container with ID starting with 39bba0429418b078fabf1e942ea403bf78858552691deaee8ecb5ee739eef389 not found: ID does not exist" Nov 21 15:52:35 crc kubenswrapper[4897]: I1121 15:52:35.046130 4897 scope.go:117] "RemoveContainer" containerID="e368c4d05068b8b7ce809473ba403753947becad88908615236570f9a9ee7aab" Nov 21 15:52:35 crc kubenswrapper[4897]: E1121 15:52:35.046414 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e368c4d05068b8b7ce809473ba403753947becad88908615236570f9a9ee7aab\": container with ID starting with e368c4d05068b8b7ce809473ba403753947becad88908615236570f9a9ee7aab not found: ID does not exist" containerID="e368c4d05068b8b7ce809473ba403753947becad88908615236570f9a9ee7aab" Nov 21 15:52:35 crc kubenswrapper[4897]: I1121 15:52:35.046448 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e368c4d05068b8b7ce809473ba403753947becad88908615236570f9a9ee7aab"} err="failed to get container status \"e368c4d05068b8b7ce809473ba403753947becad88908615236570f9a9ee7aab\": rpc error: code = NotFound desc = could not find container \"e368c4d05068b8b7ce809473ba403753947becad88908615236570f9a9ee7aab\": container with ID starting with e368c4d05068b8b7ce809473ba403753947becad88908615236570f9a9ee7aab not found: ID does not exist" Nov 21 15:52:35 crc kubenswrapper[4897]: I1121 15:52:35.046468 4897 scope.go:117] "RemoveContainer" containerID="69ee742e986b3cefbadcfe1b7dcef2141e5828e991b371095119891e29f924e1" Nov 21 15:52:35 crc kubenswrapper[4897]: E1121 15:52:35.046716 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"69ee742e986b3cefbadcfe1b7dcef2141e5828e991b371095119891e29f924e1\": container with ID starting with 69ee742e986b3cefbadcfe1b7dcef2141e5828e991b371095119891e29f924e1 not found: ID does not exist" containerID="69ee742e986b3cefbadcfe1b7dcef2141e5828e991b371095119891e29f924e1" Nov 21 15:52:35 crc kubenswrapper[4897]: I1121 15:52:35.046768 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"69ee742e986b3cefbadcfe1b7dcef2141e5828e991b371095119891e29f924e1"} err="failed to get container status \"69ee742e986b3cefbadcfe1b7dcef2141e5828e991b371095119891e29f924e1\": rpc error: code = NotFound desc = could not find container \"69ee742e986b3cefbadcfe1b7dcef2141e5828e991b371095119891e29f924e1\": container with ID starting with 69ee742e986b3cefbadcfe1b7dcef2141e5828e991b371095119891e29f924e1 not found: ID does not exist" Nov 21 15:52:36 crc kubenswrapper[4897]: I1121 15:52:36.101226 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fc2009f9-78da-4ba6-8af1-ceee9d597068" path="/var/lib/kubelet/pods/fc2009f9-78da-4ba6-8af1-ceee9d597068/volumes" Nov 21 15:53:04 crc kubenswrapper[4897]: I1121 15:53:04.370663 4897 patch_prober.go:28] interesting pod/machine-config-daemon-krv5b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 21 15:53:04 crc kubenswrapper[4897]: I1121 15:53:04.371143 4897 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 21 15:53:04 crc kubenswrapper[4897]: I1121 15:53:04.371198 4897 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" Nov 21 15:53:04 crc kubenswrapper[4897]: I1121 15:53:04.372157 4897 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"050e0baf907b19cdd602d0b93f62921dca8e2aa352d875c030b097331cbb3b00"} pod="openshift-machine-config-operator/machine-config-daemon-krv5b" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 21 15:53:04 crc kubenswrapper[4897]: I1121 15:53:04.372217 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" containerName="machine-config-daemon" containerID="cri-o://050e0baf907b19cdd602d0b93f62921dca8e2aa352d875c030b097331cbb3b00" gracePeriod=600 Nov 21 15:53:05 crc kubenswrapper[4897]: I1121 15:53:05.315368 4897 generic.go:334] "Generic (PLEG): container finished" podID="e7670227-d280-4847-b882-754429f56b0f" containerID="050e0baf907b19cdd602d0b93f62921dca8e2aa352d875c030b097331cbb3b00" exitCode=0 Nov 21 15:53:05 crc kubenswrapper[4897]: I1121 15:53:05.315458 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" event={"ID":"e7670227-d280-4847-b882-754429f56b0f","Type":"ContainerDied","Data":"050e0baf907b19cdd602d0b93f62921dca8e2aa352d875c030b097331cbb3b00"} Nov 21 15:53:05 crc kubenswrapper[4897]: I1121 15:53:05.315826 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" event={"ID":"e7670227-d280-4847-b882-754429f56b0f","Type":"ContainerStarted","Data":"0a45fdbba95fc7088c173ed5486d2f902ed547b99722f52d109c8eadeadc8938"} Nov 21 15:53:05 crc kubenswrapper[4897]: I1121 15:53:05.315860 4897 scope.go:117] "RemoveContainer" containerID="9f6ac9e28e3e389a8236d0859923fa122a6eea6c927d5c77a01f623f7524a046" Nov 21 15:54:28 crc kubenswrapper[4897]: I1121 15:54:28.071070 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-w5bqr"] Nov 21 15:54:28 crc kubenswrapper[4897]: E1121 15:54:28.072345 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ad1e6a8-2797-4538-974d-f32d2f359e09" containerName="extract-utilities" Nov 21 15:54:28 crc kubenswrapper[4897]: I1121 15:54:28.072360 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ad1e6a8-2797-4538-974d-f32d2f359e09" containerName="extract-utilities" Nov 21 15:54:28 crc kubenswrapper[4897]: E1121 15:54:28.072393 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc2009f9-78da-4ba6-8af1-ceee9d597068" containerName="extract-utilities" Nov 21 15:54:28 crc kubenswrapper[4897]: I1121 15:54:28.072400 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc2009f9-78da-4ba6-8af1-ceee9d597068" containerName="extract-utilities" Nov 21 15:54:28 crc kubenswrapper[4897]: E1121 15:54:28.072411 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ad1e6a8-2797-4538-974d-f32d2f359e09" containerName="registry-server" Nov 21 15:54:28 crc kubenswrapper[4897]: I1121 15:54:28.072417 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ad1e6a8-2797-4538-974d-f32d2f359e09" containerName="registry-server" Nov 21 15:54:28 crc kubenswrapper[4897]: E1121 15:54:28.072450 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ad1e6a8-2797-4538-974d-f32d2f359e09" containerName="extract-content" Nov 21 15:54:28 crc kubenswrapper[4897]: I1121 15:54:28.072458 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ad1e6a8-2797-4538-974d-f32d2f359e09" containerName="extract-content" Nov 21 15:54:28 crc kubenswrapper[4897]: E1121 15:54:28.072471 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc2009f9-78da-4ba6-8af1-ceee9d597068" containerName="extract-content" Nov 21 15:54:28 crc kubenswrapper[4897]: I1121 15:54:28.072477 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc2009f9-78da-4ba6-8af1-ceee9d597068" containerName="extract-content" Nov 21 15:54:28 crc kubenswrapper[4897]: E1121 15:54:28.072487 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc2009f9-78da-4ba6-8af1-ceee9d597068" containerName="registry-server" Nov 21 15:54:28 crc kubenswrapper[4897]: I1121 15:54:28.072493 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc2009f9-78da-4ba6-8af1-ceee9d597068" containerName="registry-server" Nov 21 15:54:28 crc kubenswrapper[4897]: I1121 15:54:28.072720 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ad1e6a8-2797-4538-974d-f32d2f359e09" containerName="registry-server" Nov 21 15:54:28 crc kubenswrapper[4897]: I1121 15:54:28.072737 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc2009f9-78da-4ba6-8af1-ceee9d597068" containerName="registry-server" Nov 21 15:54:28 crc kubenswrapper[4897]: I1121 15:54:28.074478 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-w5bqr" Nov 21 15:54:28 crc kubenswrapper[4897]: I1121 15:54:28.113140 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-w5bqr"] Nov 21 15:54:28 crc kubenswrapper[4897]: I1121 15:54:28.179765 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba24b5c9-9a7c-44ba-a3ee-ffa1caf6a2e2-utilities\") pod \"redhat-marketplace-w5bqr\" (UID: \"ba24b5c9-9a7c-44ba-a3ee-ffa1caf6a2e2\") " pod="openshift-marketplace/redhat-marketplace-w5bqr" Nov 21 15:54:28 crc kubenswrapper[4897]: I1121 15:54:28.181440 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba24b5c9-9a7c-44ba-a3ee-ffa1caf6a2e2-catalog-content\") pod \"redhat-marketplace-w5bqr\" (UID: \"ba24b5c9-9a7c-44ba-a3ee-ffa1caf6a2e2\") " pod="openshift-marketplace/redhat-marketplace-w5bqr" Nov 21 15:54:28 crc kubenswrapper[4897]: I1121 15:54:28.183354 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qzxvf\" (UniqueName: \"kubernetes.io/projected/ba24b5c9-9a7c-44ba-a3ee-ffa1caf6a2e2-kube-api-access-qzxvf\") pod \"redhat-marketplace-w5bqr\" (UID: \"ba24b5c9-9a7c-44ba-a3ee-ffa1caf6a2e2\") " pod="openshift-marketplace/redhat-marketplace-w5bqr" Nov 21 15:54:28 crc kubenswrapper[4897]: I1121 15:54:28.285651 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qzxvf\" (UniqueName: \"kubernetes.io/projected/ba24b5c9-9a7c-44ba-a3ee-ffa1caf6a2e2-kube-api-access-qzxvf\") pod \"redhat-marketplace-w5bqr\" (UID: \"ba24b5c9-9a7c-44ba-a3ee-ffa1caf6a2e2\") " pod="openshift-marketplace/redhat-marketplace-w5bqr" Nov 21 15:54:28 crc kubenswrapper[4897]: I1121 15:54:28.286226 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba24b5c9-9a7c-44ba-a3ee-ffa1caf6a2e2-utilities\") pod \"redhat-marketplace-w5bqr\" (UID: \"ba24b5c9-9a7c-44ba-a3ee-ffa1caf6a2e2\") " pod="openshift-marketplace/redhat-marketplace-w5bqr" Nov 21 15:54:28 crc kubenswrapper[4897]: I1121 15:54:28.286374 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba24b5c9-9a7c-44ba-a3ee-ffa1caf6a2e2-catalog-content\") pod \"redhat-marketplace-w5bqr\" (UID: \"ba24b5c9-9a7c-44ba-a3ee-ffa1caf6a2e2\") " pod="openshift-marketplace/redhat-marketplace-w5bqr" Nov 21 15:54:28 crc kubenswrapper[4897]: I1121 15:54:28.286553 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba24b5c9-9a7c-44ba-a3ee-ffa1caf6a2e2-utilities\") pod \"redhat-marketplace-w5bqr\" (UID: \"ba24b5c9-9a7c-44ba-a3ee-ffa1caf6a2e2\") " pod="openshift-marketplace/redhat-marketplace-w5bqr" Nov 21 15:54:28 crc kubenswrapper[4897]: I1121 15:54:28.286753 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba24b5c9-9a7c-44ba-a3ee-ffa1caf6a2e2-catalog-content\") pod \"redhat-marketplace-w5bqr\" (UID: \"ba24b5c9-9a7c-44ba-a3ee-ffa1caf6a2e2\") " pod="openshift-marketplace/redhat-marketplace-w5bqr" Nov 21 15:54:28 crc kubenswrapper[4897]: I1121 15:54:28.306384 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qzxvf\" (UniqueName: \"kubernetes.io/projected/ba24b5c9-9a7c-44ba-a3ee-ffa1caf6a2e2-kube-api-access-qzxvf\") pod \"redhat-marketplace-w5bqr\" (UID: \"ba24b5c9-9a7c-44ba-a3ee-ffa1caf6a2e2\") " pod="openshift-marketplace/redhat-marketplace-w5bqr" Nov 21 15:54:28 crc kubenswrapper[4897]: I1121 15:54:28.405475 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-w5bqr" Nov 21 15:54:29 crc kubenswrapper[4897]: I1121 15:54:29.003093 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-w5bqr"] Nov 21 15:54:29 crc kubenswrapper[4897]: I1121 15:54:29.600908 4897 generic.go:334] "Generic (PLEG): container finished" podID="ba24b5c9-9a7c-44ba-a3ee-ffa1caf6a2e2" containerID="bcca416221a22373454a3b373a032f1f0c4ea527c6836a95552013e5af81b9d1" exitCode=0 Nov 21 15:54:29 crc kubenswrapper[4897]: I1121 15:54:29.601011 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w5bqr" event={"ID":"ba24b5c9-9a7c-44ba-a3ee-ffa1caf6a2e2","Type":"ContainerDied","Data":"bcca416221a22373454a3b373a032f1f0c4ea527c6836a95552013e5af81b9d1"} Nov 21 15:54:29 crc kubenswrapper[4897]: I1121 15:54:29.601291 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w5bqr" event={"ID":"ba24b5c9-9a7c-44ba-a3ee-ffa1caf6a2e2","Type":"ContainerStarted","Data":"fdb83c710118c5c32aeeec646bca9702c2aeac8d2934f6b35ab8383297ff21a3"} Nov 21 15:54:30 crc kubenswrapper[4897]: I1121 15:54:30.618119 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w5bqr" event={"ID":"ba24b5c9-9a7c-44ba-a3ee-ffa1caf6a2e2","Type":"ContainerStarted","Data":"10fdb96179b1330ef1674a7d15c188de6192cc705249de922fd0350e14fb90c6"} Nov 21 15:54:31 crc kubenswrapper[4897]: I1121 15:54:31.635063 4897 generic.go:334] "Generic (PLEG): container finished" podID="ba24b5c9-9a7c-44ba-a3ee-ffa1caf6a2e2" containerID="10fdb96179b1330ef1674a7d15c188de6192cc705249de922fd0350e14fb90c6" exitCode=0 Nov 21 15:54:31 crc kubenswrapper[4897]: I1121 15:54:31.635137 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w5bqr" event={"ID":"ba24b5c9-9a7c-44ba-a3ee-ffa1caf6a2e2","Type":"ContainerDied","Data":"10fdb96179b1330ef1674a7d15c188de6192cc705249de922fd0350e14fb90c6"} Nov 21 15:54:32 crc kubenswrapper[4897]: I1121 15:54:32.651597 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w5bqr" event={"ID":"ba24b5c9-9a7c-44ba-a3ee-ffa1caf6a2e2","Type":"ContainerStarted","Data":"f588fcb14661b7ab0a9fe320b560c92c954f2aec099737b2c1627dca2cbe4a9e"} Nov 21 15:54:32 crc kubenswrapper[4897]: I1121 15:54:32.671334 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-w5bqr" podStartSLOduration=2.046849362 podStartE2EDuration="4.671315048s" podCreationTimestamp="2025-11-21 15:54:28 +0000 UTC" firstStartedPulling="2025-11-21 15:54:29.607591165 +0000 UTC m=+6346.892184640" lastFinishedPulling="2025-11-21 15:54:32.232056841 +0000 UTC m=+6349.516650326" observedRunningTime="2025-11-21 15:54:32.666112429 +0000 UTC m=+6349.950705904" watchObservedRunningTime="2025-11-21 15:54:32.671315048 +0000 UTC m=+6349.955908523" Nov 21 15:54:38 crc kubenswrapper[4897]: I1121 15:54:38.405684 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-w5bqr" Nov 21 15:54:38 crc kubenswrapper[4897]: I1121 15:54:38.406263 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-w5bqr" Nov 21 15:54:38 crc kubenswrapper[4897]: I1121 15:54:38.464031 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-w5bqr" Nov 21 15:54:38 crc kubenswrapper[4897]: I1121 15:54:38.771085 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-w5bqr" Nov 21 15:54:42 crc kubenswrapper[4897]: I1121 15:54:42.051198 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-w5bqr"] Nov 21 15:54:42 crc kubenswrapper[4897]: I1121 15:54:42.052042 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-w5bqr" podUID="ba24b5c9-9a7c-44ba-a3ee-ffa1caf6a2e2" containerName="registry-server" containerID="cri-o://f588fcb14661b7ab0a9fe320b560c92c954f2aec099737b2c1627dca2cbe4a9e" gracePeriod=2 Nov 21 15:54:42 crc kubenswrapper[4897]: I1121 15:54:42.637536 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-w5bqr" Nov 21 15:54:42 crc kubenswrapper[4897]: I1121 15:54:42.766940 4897 generic.go:334] "Generic (PLEG): container finished" podID="ba24b5c9-9a7c-44ba-a3ee-ffa1caf6a2e2" containerID="f588fcb14661b7ab0a9fe320b560c92c954f2aec099737b2c1627dca2cbe4a9e" exitCode=0 Nov 21 15:54:42 crc kubenswrapper[4897]: I1121 15:54:42.767013 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w5bqr" event={"ID":"ba24b5c9-9a7c-44ba-a3ee-ffa1caf6a2e2","Type":"ContainerDied","Data":"f588fcb14661b7ab0a9fe320b560c92c954f2aec099737b2c1627dca2cbe4a9e"} Nov 21 15:54:42 crc kubenswrapper[4897]: I1121 15:54:42.767345 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w5bqr" event={"ID":"ba24b5c9-9a7c-44ba-a3ee-ffa1caf6a2e2","Type":"ContainerDied","Data":"fdb83c710118c5c32aeeec646bca9702c2aeac8d2934f6b35ab8383297ff21a3"} Nov 21 15:54:42 crc kubenswrapper[4897]: I1121 15:54:42.767367 4897 scope.go:117] "RemoveContainer" containerID="f588fcb14661b7ab0a9fe320b560c92c954f2aec099737b2c1627dca2cbe4a9e" Nov 21 15:54:42 crc kubenswrapper[4897]: I1121 15:54:42.767083 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-w5bqr" Nov 21 15:54:42 crc kubenswrapper[4897]: I1121 15:54:42.769607 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba24b5c9-9a7c-44ba-a3ee-ffa1caf6a2e2-utilities\") pod \"ba24b5c9-9a7c-44ba-a3ee-ffa1caf6a2e2\" (UID: \"ba24b5c9-9a7c-44ba-a3ee-ffa1caf6a2e2\") " Nov 21 15:54:42 crc kubenswrapper[4897]: I1121 15:54:42.769643 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba24b5c9-9a7c-44ba-a3ee-ffa1caf6a2e2-catalog-content\") pod \"ba24b5c9-9a7c-44ba-a3ee-ffa1caf6a2e2\" (UID: \"ba24b5c9-9a7c-44ba-a3ee-ffa1caf6a2e2\") " Nov 21 15:54:42 crc kubenswrapper[4897]: I1121 15:54:42.769697 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qzxvf\" (UniqueName: \"kubernetes.io/projected/ba24b5c9-9a7c-44ba-a3ee-ffa1caf6a2e2-kube-api-access-qzxvf\") pod \"ba24b5c9-9a7c-44ba-a3ee-ffa1caf6a2e2\" (UID: \"ba24b5c9-9a7c-44ba-a3ee-ffa1caf6a2e2\") " Nov 21 15:54:42 crc kubenswrapper[4897]: I1121 15:54:42.771684 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ba24b5c9-9a7c-44ba-a3ee-ffa1caf6a2e2-utilities" (OuterVolumeSpecName: "utilities") pod "ba24b5c9-9a7c-44ba-a3ee-ffa1caf6a2e2" (UID: "ba24b5c9-9a7c-44ba-a3ee-ffa1caf6a2e2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 15:54:42 crc kubenswrapper[4897]: I1121 15:54:42.785993 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ba24b5c9-9a7c-44ba-a3ee-ffa1caf6a2e2-kube-api-access-qzxvf" (OuterVolumeSpecName: "kube-api-access-qzxvf") pod "ba24b5c9-9a7c-44ba-a3ee-ffa1caf6a2e2" (UID: "ba24b5c9-9a7c-44ba-a3ee-ffa1caf6a2e2"). InnerVolumeSpecName "kube-api-access-qzxvf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 15:54:42 crc kubenswrapper[4897]: I1121 15:54:42.790441 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ba24b5c9-9a7c-44ba-a3ee-ffa1caf6a2e2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ba24b5c9-9a7c-44ba-a3ee-ffa1caf6a2e2" (UID: "ba24b5c9-9a7c-44ba-a3ee-ffa1caf6a2e2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 15:54:42 crc kubenswrapper[4897]: I1121 15:54:42.858552 4897 scope.go:117] "RemoveContainer" containerID="10fdb96179b1330ef1674a7d15c188de6192cc705249de922fd0350e14fb90c6" Nov 21 15:54:42 crc kubenswrapper[4897]: I1121 15:54:42.872393 4897 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba24b5c9-9a7c-44ba-a3ee-ffa1caf6a2e2-utilities\") on node \"crc\" DevicePath \"\"" Nov 21 15:54:42 crc kubenswrapper[4897]: I1121 15:54:42.872430 4897 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba24b5c9-9a7c-44ba-a3ee-ffa1caf6a2e2-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 21 15:54:42 crc kubenswrapper[4897]: I1121 15:54:42.872443 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qzxvf\" (UniqueName: \"kubernetes.io/projected/ba24b5c9-9a7c-44ba-a3ee-ffa1caf6a2e2-kube-api-access-qzxvf\") on node \"crc\" DevicePath \"\"" Nov 21 15:54:42 crc kubenswrapper[4897]: I1121 15:54:42.882063 4897 scope.go:117] "RemoveContainer" containerID="bcca416221a22373454a3b373a032f1f0c4ea527c6836a95552013e5af81b9d1" Nov 21 15:54:42 crc kubenswrapper[4897]: I1121 15:54:42.944948 4897 scope.go:117] "RemoveContainer" containerID="f588fcb14661b7ab0a9fe320b560c92c954f2aec099737b2c1627dca2cbe4a9e" Nov 21 15:54:42 crc kubenswrapper[4897]: E1121 15:54:42.946723 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f588fcb14661b7ab0a9fe320b560c92c954f2aec099737b2c1627dca2cbe4a9e\": container with ID starting with f588fcb14661b7ab0a9fe320b560c92c954f2aec099737b2c1627dca2cbe4a9e not found: ID does not exist" containerID="f588fcb14661b7ab0a9fe320b560c92c954f2aec099737b2c1627dca2cbe4a9e" Nov 21 15:54:42 crc kubenswrapper[4897]: I1121 15:54:42.946847 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f588fcb14661b7ab0a9fe320b560c92c954f2aec099737b2c1627dca2cbe4a9e"} err="failed to get container status \"f588fcb14661b7ab0a9fe320b560c92c954f2aec099737b2c1627dca2cbe4a9e\": rpc error: code = NotFound desc = could not find container \"f588fcb14661b7ab0a9fe320b560c92c954f2aec099737b2c1627dca2cbe4a9e\": container with ID starting with f588fcb14661b7ab0a9fe320b560c92c954f2aec099737b2c1627dca2cbe4a9e not found: ID does not exist" Nov 21 15:54:42 crc kubenswrapper[4897]: I1121 15:54:42.946933 4897 scope.go:117] "RemoveContainer" containerID="10fdb96179b1330ef1674a7d15c188de6192cc705249de922fd0350e14fb90c6" Nov 21 15:54:42 crc kubenswrapper[4897]: E1121 15:54:42.948098 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"10fdb96179b1330ef1674a7d15c188de6192cc705249de922fd0350e14fb90c6\": container with ID starting with 10fdb96179b1330ef1674a7d15c188de6192cc705249de922fd0350e14fb90c6 not found: ID does not exist" containerID="10fdb96179b1330ef1674a7d15c188de6192cc705249de922fd0350e14fb90c6" Nov 21 15:54:42 crc kubenswrapper[4897]: I1121 15:54:42.948140 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"10fdb96179b1330ef1674a7d15c188de6192cc705249de922fd0350e14fb90c6"} err="failed to get container status \"10fdb96179b1330ef1674a7d15c188de6192cc705249de922fd0350e14fb90c6\": rpc error: code = NotFound desc = could not find container \"10fdb96179b1330ef1674a7d15c188de6192cc705249de922fd0350e14fb90c6\": container with ID starting with 10fdb96179b1330ef1674a7d15c188de6192cc705249de922fd0350e14fb90c6 not found: ID does not exist" Nov 21 15:54:42 crc kubenswrapper[4897]: I1121 15:54:42.948164 4897 scope.go:117] "RemoveContainer" containerID="bcca416221a22373454a3b373a032f1f0c4ea527c6836a95552013e5af81b9d1" Nov 21 15:54:42 crc kubenswrapper[4897]: E1121 15:54:42.948619 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bcca416221a22373454a3b373a032f1f0c4ea527c6836a95552013e5af81b9d1\": container with ID starting with bcca416221a22373454a3b373a032f1f0c4ea527c6836a95552013e5af81b9d1 not found: ID does not exist" containerID="bcca416221a22373454a3b373a032f1f0c4ea527c6836a95552013e5af81b9d1" Nov 21 15:54:42 crc kubenswrapper[4897]: I1121 15:54:42.948730 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bcca416221a22373454a3b373a032f1f0c4ea527c6836a95552013e5af81b9d1"} err="failed to get container status \"bcca416221a22373454a3b373a032f1f0c4ea527c6836a95552013e5af81b9d1\": rpc error: code = NotFound desc = could not find container \"bcca416221a22373454a3b373a032f1f0c4ea527c6836a95552013e5af81b9d1\": container with ID starting with bcca416221a22373454a3b373a032f1f0c4ea527c6836a95552013e5af81b9d1 not found: ID does not exist" Nov 21 15:54:43 crc kubenswrapper[4897]: I1121 15:54:43.116026 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-w5bqr"] Nov 21 15:54:43 crc kubenswrapper[4897]: I1121 15:54:43.129447 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-w5bqr"] Nov 21 15:54:44 crc kubenswrapper[4897]: I1121 15:54:44.117016 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ba24b5c9-9a7c-44ba-a3ee-ffa1caf6a2e2" path="/var/lib/kubelet/pods/ba24b5c9-9a7c-44ba-a3ee-ffa1caf6a2e2/volumes" Nov 21 15:55:04 crc kubenswrapper[4897]: I1121 15:55:04.371118 4897 patch_prober.go:28] interesting pod/machine-config-daemon-krv5b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 21 15:55:04 crc kubenswrapper[4897]: I1121 15:55:04.372119 4897 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 21 15:55:18 crc kubenswrapper[4897]: I1121 15:55:18.143176 4897 generic.go:334] "Generic (PLEG): container finished" podID="b3a6d915-d988-45e8-a007-d135548ef81e" containerID="b8a02b6161e216b0fbfd9fc93366f53197a472ac54762a52ff31a8d73a41a20e" exitCode=0 Nov 21 15:55:18 crc kubenswrapper[4897]: I1121 15:55:18.143262 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"b3a6d915-d988-45e8-a007-d135548ef81e","Type":"ContainerDied","Data":"b8a02b6161e216b0fbfd9fc93366f53197a472ac54762a52ff31a8d73a41a20e"} Nov 21 15:55:19 crc kubenswrapper[4897]: I1121 15:55:19.701377 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Nov 21 15:55:19 crc kubenswrapper[4897]: I1121 15:55:19.762469 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/b3a6d915-d988-45e8-a007-d135548ef81e-ca-certs\") pod \"b3a6d915-d988-45e8-a007-d135548ef81e\" (UID: \"b3a6d915-d988-45e8-a007-d135548ef81e\") " Nov 21 15:55:19 crc kubenswrapper[4897]: I1121 15:55:19.762585 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/b3a6d915-d988-45e8-a007-d135548ef81e-test-operator-ephemeral-workdir\") pod \"b3a6d915-d988-45e8-a007-d135548ef81e\" (UID: \"b3a6d915-d988-45e8-a007-d135548ef81e\") " Nov 21 15:55:19 crc kubenswrapper[4897]: I1121 15:55:19.762716 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"b3a6d915-d988-45e8-a007-d135548ef81e\" (UID: \"b3a6d915-d988-45e8-a007-d135548ef81e\") " Nov 21 15:55:19 crc kubenswrapper[4897]: I1121 15:55:19.762766 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/b3a6d915-d988-45e8-a007-d135548ef81e-openstack-config\") pod \"b3a6d915-d988-45e8-a007-d135548ef81e\" (UID: \"b3a6d915-d988-45e8-a007-d135548ef81e\") " Nov 21 15:55:19 crc kubenswrapper[4897]: I1121 15:55:19.762884 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/b3a6d915-d988-45e8-a007-d135548ef81e-openstack-config-secret\") pod \"b3a6d915-d988-45e8-a007-d135548ef81e\" (UID: \"b3a6d915-d988-45e8-a007-d135548ef81e\") " Nov 21 15:55:19 crc kubenswrapper[4897]: I1121 15:55:19.762916 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4b626\" (UniqueName: \"kubernetes.io/projected/b3a6d915-d988-45e8-a007-d135548ef81e-kube-api-access-4b626\") pod \"b3a6d915-d988-45e8-a007-d135548ef81e\" (UID: \"b3a6d915-d988-45e8-a007-d135548ef81e\") " Nov 21 15:55:19 crc kubenswrapper[4897]: I1121 15:55:19.762932 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b3a6d915-d988-45e8-a007-d135548ef81e-ssh-key\") pod \"b3a6d915-d988-45e8-a007-d135548ef81e\" (UID: \"b3a6d915-d988-45e8-a007-d135548ef81e\") " Nov 21 15:55:19 crc kubenswrapper[4897]: I1121 15:55:19.762952 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b3a6d915-d988-45e8-a007-d135548ef81e-config-data\") pod \"b3a6d915-d988-45e8-a007-d135548ef81e\" (UID: \"b3a6d915-d988-45e8-a007-d135548ef81e\") " Nov 21 15:55:19 crc kubenswrapper[4897]: I1121 15:55:19.763027 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/b3a6d915-d988-45e8-a007-d135548ef81e-test-operator-ephemeral-temporary\") pod \"b3a6d915-d988-45e8-a007-d135548ef81e\" (UID: \"b3a6d915-d988-45e8-a007-d135548ef81e\") " Nov 21 15:55:19 crc kubenswrapper[4897]: I1121 15:55:19.764185 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b3a6d915-d988-45e8-a007-d135548ef81e-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "b3a6d915-d988-45e8-a007-d135548ef81e" (UID: "b3a6d915-d988-45e8-a007-d135548ef81e"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 15:55:19 crc kubenswrapper[4897]: I1121 15:55:19.765080 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b3a6d915-d988-45e8-a007-d135548ef81e-config-data" (OuterVolumeSpecName: "config-data") pod "b3a6d915-d988-45e8-a007-d135548ef81e" (UID: "b3a6d915-d988-45e8-a007-d135548ef81e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 15:55:19 crc kubenswrapper[4897]: I1121 15:55:19.769419 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3a6d915-d988-45e8-a007-d135548ef81e-kube-api-access-4b626" (OuterVolumeSpecName: "kube-api-access-4b626") pod "b3a6d915-d988-45e8-a007-d135548ef81e" (UID: "b3a6d915-d988-45e8-a007-d135548ef81e"). InnerVolumeSpecName "kube-api-access-4b626". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 15:55:19 crc kubenswrapper[4897]: I1121 15:55:19.770082 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b3a6d915-d988-45e8-a007-d135548ef81e-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "b3a6d915-d988-45e8-a007-d135548ef81e" (UID: "b3a6d915-d988-45e8-a007-d135548ef81e"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 15:55:19 crc kubenswrapper[4897]: I1121 15:55:19.773359 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "test-operator-logs") pod "b3a6d915-d988-45e8-a007-d135548ef81e" (UID: "b3a6d915-d988-45e8-a007-d135548ef81e"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 21 15:55:19 crc kubenswrapper[4897]: I1121 15:55:19.809975 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3a6d915-d988-45e8-a007-d135548ef81e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b3a6d915-d988-45e8-a007-d135548ef81e" (UID: "b3a6d915-d988-45e8-a007-d135548ef81e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 15:55:19 crc kubenswrapper[4897]: I1121 15:55:19.815102 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3a6d915-d988-45e8-a007-d135548ef81e-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "b3a6d915-d988-45e8-a007-d135548ef81e" (UID: "b3a6d915-d988-45e8-a007-d135548ef81e"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 15:55:19 crc kubenswrapper[4897]: I1121 15:55:19.835669 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3a6d915-d988-45e8-a007-d135548ef81e-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "b3a6d915-d988-45e8-a007-d135548ef81e" (UID: "b3a6d915-d988-45e8-a007-d135548ef81e"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 15:55:19 crc kubenswrapper[4897]: I1121 15:55:19.852105 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b3a6d915-d988-45e8-a007-d135548ef81e-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "b3a6d915-d988-45e8-a007-d135548ef81e" (UID: "b3a6d915-d988-45e8-a007-d135548ef81e"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 15:55:19 crc kubenswrapper[4897]: I1121 15:55:19.875983 4897 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Nov 21 15:55:19 crc kubenswrapper[4897]: I1121 15:55:19.876053 4897 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/b3a6d915-d988-45e8-a007-d135548ef81e-openstack-config\") on node \"crc\" DevicePath \"\"" Nov 21 15:55:19 crc kubenswrapper[4897]: I1121 15:55:19.876075 4897 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/b3a6d915-d988-45e8-a007-d135548ef81e-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Nov 21 15:55:19 crc kubenswrapper[4897]: I1121 15:55:19.876094 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4b626\" (UniqueName: \"kubernetes.io/projected/b3a6d915-d988-45e8-a007-d135548ef81e-kube-api-access-4b626\") on node \"crc\" DevicePath \"\"" Nov 21 15:55:19 crc kubenswrapper[4897]: I1121 15:55:19.876112 4897 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b3a6d915-d988-45e8-a007-d135548ef81e-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 21 15:55:19 crc kubenswrapper[4897]: I1121 15:55:19.876127 4897 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b3a6d915-d988-45e8-a007-d135548ef81e-config-data\") on node \"crc\" DevicePath \"\"" Nov 21 15:55:19 crc kubenswrapper[4897]: I1121 15:55:19.876143 4897 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/b3a6d915-d988-45e8-a007-d135548ef81e-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Nov 21 15:55:19 crc kubenswrapper[4897]: I1121 15:55:19.876158 4897 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/b3a6d915-d988-45e8-a007-d135548ef81e-ca-certs\") on node \"crc\" DevicePath \"\"" Nov 21 15:55:19 crc kubenswrapper[4897]: I1121 15:55:19.876173 4897 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/b3a6d915-d988-45e8-a007-d135548ef81e-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Nov 21 15:55:19 crc kubenswrapper[4897]: I1121 15:55:19.907060 4897 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Nov 21 15:55:19 crc kubenswrapper[4897]: I1121 15:55:19.979164 4897 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Nov 21 15:55:20 crc kubenswrapper[4897]: I1121 15:55:20.173686 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"b3a6d915-d988-45e8-a007-d135548ef81e","Type":"ContainerDied","Data":"8c6ceb51ed1cd6a4980a2c051cfafeab46e377ccd32e396d02240218c1d4c841"} Nov 21 15:55:20 crc kubenswrapper[4897]: I1121 15:55:20.173738 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Nov 21 15:55:20 crc kubenswrapper[4897]: I1121 15:55:20.173739 4897 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8c6ceb51ed1cd6a4980a2c051cfafeab46e377ccd32e396d02240218c1d4c841" Nov 21 15:55:29 crc kubenswrapper[4897]: I1121 15:55:29.299923 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Nov 21 15:55:29 crc kubenswrapper[4897]: E1121 15:55:29.300910 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba24b5c9-9a7c-44ba-a3ee-ffa1caf6a2e2" containerName="registry-server" Nov 21 15:55:29 crc kubenswrapper[4897]: I1121 15:55:29.300922 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba24b5c9-9a7c-44ba-a3ee-ffa1caf6a2e2" containerName="registry-server" Nov 21 15:55:29 crc kubenswrapper[4897]: E1121 15:55:29.300944 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba24b5c9-9a7c-44ba-a3ee-ffa1caf6a2e2" containerName="extract-utilities" Nov 21 15:55:29 crc kubenswrapper[4897]: I1121 15:55:29.300950 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba24b5c9-9a7c-44ba-a3ee-ffa1caf6a2e2" containerName="extract-utilities" Nov 21 15:55:29 crc kubenswrapper[4897]: E1121 15:55:29.300970 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3a6d915-d988-45e8-a007-d135548ef81e" containerName="tempest-tests-tempest-tests-runner" Nov 21 15:55:29 crc kubenswrapper[4897]: I1121 15:55:29.300976 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3a6d915-d988-45e8-a007-d135548ef81e" containerName="tempest-tests-tempest-tests-runner" Nov 21 15:55:29 crc kubenswrapper[4897]: E1121 15:55:29.301006 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba24b5c9-9a7c-44ba-a3ee-ffa1caf6a2e2" containerName="extract-content" Nov 21 15:55:29 crc kubenswrapper[4897]: I1121 15:55:29.301012 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba24b5c9-9a7c-44ba-a3ee-ffa1caf6a2e2" containerName="extract-content" Nov 21 15:55:29 crc kubenswrapper[4897]: I1121 15:55:29.301217 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba24b5c9-9a7c-44ba-a3ee-ffa1caf6a2e2" containerName="registry-server" Nov 21 15:55:29 crc kubenswrapper[4897]: I1121 15:55:29.301246 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3a6d915-d988-45e8-a007-d135548ef81e" containerName="tempest-tests-tempest-tests-runner" Nov 21 15:55:29 crc kubenswrapper[4897]: I1121 15:55:29.302187 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 21 15:55:29 crc kubenswrapper[4897]: I1121 15:55:29.304226 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-8gssp" Nov 21 15:55:29 crc kubenswrapper[4897]: I1121 15:55:29.319942 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Nov 21 15:55:29 crc kubenswrapper[4897]: I1121 15:55:29.415290 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pmwqm\" (UniqueName: \"kubernetes.io/projected/e42c0131-5677-4f0a-b71d-df730d0ab967-kube-api-access-pmwqm\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"e42c0131-5677-4f0a-b71d-df730d0ab967\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 21 15:55:29 crc kubenswrapper[4897]: I1121 15:55:29.415352 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"e42c0131-5677-4f0a-b71d-df730d0ab967\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 21 15:55:29 crc kubenswrapper[4897]: I1121 15:55:29.517944 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pmwqm\" (UniqueName: \"kubernetes.io/projected/e42c0131-5677-4f0a-b71d-df730d0ab967-kube-api-access-pmwqm\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"e42c0131-5677-4f0a-b71d-df730d0ab967\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 21 15:55:29 crc kubenswrapper[4897]: I1121 15:55:29.518008 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"e42c0131-5677-4f0a-b71d-df730d0ab967\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 21 15:55:29 crc kubenswrapper[4897]: I1121 15:55:29.519280 4897 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"e42c0131-5677-4f0a-b71d-df730d0ab967\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 21 15:55:29 crc kubenswrapper[4897]: I1121 15:55:29.537907 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pmwqm\" (UniqueName: \"kubernetes.io/projected/e42c0131-5677-4f0a-b71d-df730d0ab967-kube-api-access-pmwqm\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"e42c0131-5677-4f0a-b71d-df730d0ab967\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 21 15:55:29 crc kubenswrapper[4897]: I1121 15:55:29.550658 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"e42c0131-5677-4f0a-b71d-df730d0ab967\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 21 15:55:29 crc kubenswrapper[4897]: I1121 15:55:29.622340 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 21 15:55:30 crc kubenswrapper[4897]: I1121 15:55:30.148774 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Nov 21 15:55:30 crc kubenswrapper[4897]: I1121 15:55:30.294659 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"e42c0131-5677-4f0a-b71d-df730d0ab967","Type":"ContainerStarted","Data":"78e86aed064346d744271198436f421edc92d258dd297683687100fdbf7e34b1"} Nov 21 15:55:32 crc kubenswrapper[4897]: I1121 15:55:32.319283 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"e42c0131-5677-4f0a-b71d-df730d0ab967","Type":"ContainerStarted","Data":"0139b9b8756675fb31b669b71d30083dfa4d505f5c2593e0065b112a4c0d4917"} Nov 21 15:55:32 crc kubenswrapper[4897]: I1121 15:55:32.337897 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=2.27619833 podStartE2EDuration="3.337879995s" podCreationTimestamp="2025-11-21 15:55:29 +0000 UTC" firstStartedPulling="2025-11-21 15:55:30.157140647 +0000 UTC m=+6407.441734122" lastFinishedPulling="2025-11-21 15:55:31.218822312 +0000 UTC m=+6408.503415787" observedRunningTime="2025-11-21 15:55:32.332329135 +0000 UTC m=+6409.616922640" watchObservedRunningTime="2025-11-21 15:55:32.337879995 +0000 UTC m=+6409.622473460" Nov 21 15:55:34 crc kubenswrapper[4897]: I1121 15:55:34.370679 4897 patch_prober.go:28] interesting pod/machine-config-daemon-krv5b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 21 15:55:34 crc kubenswrapper[4897]: I1121 15:55:34.371349 4897 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 21 15:56:04 crc kubenswrapper[4897]: I1121 15:56:04.371444 4897 patch_prober.go:28] interesting pod/machine-config-daemon-krv5b container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 21 15:56:04 crc kubenswrapper[4897]: I1121 15:56:04.371990 4897 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 21 15:56:04 crc kubenswrapper[4897]: I1121 15:56:04.372052 4897 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" Nov 21 15:56:04 crc kubenswrapper[4897]: I1121 15:56:04.373067 4897 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0a45fdbba95fc7088c173ed5486d2f902ed547b99722f52d109c8eadeadc8938"} pod="openshift-machine-config-operator/machine-config-daemon-krv5b" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 21 15:56:04 crc kubenswrapper[4897]: I1121 15:56:04.373134 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" containerName="machine-config-daemon" containerID="cri-o://0a45fdbba95fc7088c173ed5486d2f902ed547b99722f52d109c8eadeadc8938" gracePeriod=600 Nov 21 15:56:05 crc kubenswrapper[4897]: E1121 15:56:05.131422 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:56:05 crc kubenswrapper[4897]: I1121 15:56:05.697692 4897 generic.go:334] "Generic (PLEG): container finished" podID="e7670227-d280-4847-b882-754429f56b0f" containerID="0a45fdbba95fc7088c173ed5486d2f902ed547b99722f52d109c8eadeadc8938" exitCode=0 Nov 21 15:56:05 crc kubenswrapper[4897]: I1121 15:56:05.697730 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" event={"ID":"e7670227-d280-4847-b882-754429f56b0f","Type":"ContainerDied","Data":"0a45fdbba95fc7088c173ed5486d2f902ed547b99722f52d109c8eadeadc8938"} Nov 21 15:56:05 crc kubenswrapper[4897]: I1121 15:56:05.698754 4897 scope.go:117] "RemoveContainer" containerID="050e0baf907b19cdd602d0b93f62921dca8e2aa352d875c030b097331cbb3b00" Nov 21 15:56:05 crc kubenswrapper[4897]: I1121 15:56:05.699652 4897 scope.go:117] "RemoveContainer" containerID="0a45fdbba95fc7088c173ed5486d2f902ed547b99722f52d109c8eadeadc8938" Nov 21 15:56:05 crc kubenswrapper[4897]: E1121 15:56:05.700018 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:56:19 crc kubenswrapper[4897]: I1121 15:56:19.089419 4897 scope.go:117] "RemoveContainer" containerID="0a45fdbba95fc7088c173ed5486d2f902ed547b99722f52d109c8eadeadc8938" Nov 21 15:56:19 crc kubenswrapper[4897]: E1121 15:56:19.091036 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:56:20 crc kubenswrapper[4897]: I1121 15:56:20.125226 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-w56h8/must-gather-cg6dj"] Nov 21 15:56:20 crc kubenswrapper[4897]: I1121 15:56:20.129721 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-w56h8/must-gather-cg6dj" Nov 21 15:56:20 crc kubenswrapper[4897]: I1121 15:56:20.138337 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-w56h8"/"kube-root-ca.crt" Nov 21 15:56:20 crc kubenswrapper[4897]: I1121 15:56:20.138592 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-w56h8"/"openshift-service-ca.crt" Nov 21 15:56:20 crc kubenswrapper[4897]: I1121 15:56:20.251144 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/c80d7a41-98ec-44af-a19b-d9d7d2018ad9-must-gather-output\") pod \"must-gather-cg6dj\" (UID: \"c80d7a41-98ec-44af-a19b-d9d7d2018ad9\") " pod="openshift-must-gather-w56h8/must-gather-cg6dj" Nov 21 15:56:20 crc kubenswrapper[4897]: I1121 15:56:20.251326 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mqmp4\" (UniqueName: \"kubernetes.io/projected/c80d7a41-98ec-44af-a19b-d9d7d2018ad9-kube-api-access-mqmp4\") pod \"must-gather-cg6dj\" (UID: \"c80d7a41-98ec-44af-a19b-d9d7d2018ad9\") " pod="openshift-must-gather-w56h8/must-gather-cg6dj" Nov 21 15:56:20 crc kubenswrapper[4897]: I1121 15:56:20.253976 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-w56h8/must-gather-cg6dj"] Nov 21 15:56:20 crc kubenswrapper[4897]: I1121 15:56:20.356474 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/c80d7a41-98ec-44af-a19b-d9d7d2018ad9-must-gather-output\") pod \"must-gather-cg6dj\" (UID: \"c80d7a41-98ec-44af-a19b-d9d7d2018ad9\") " pod="openshift-must-gather-w56h8/must-gather-cg6dj" Nov 21 15:56:20 crc kubenswrapper[4897]: I1121 15:56:20.356609 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mqmp4\" (UniqueName: \"kubernetes.io/projected/c80d7a41-98ec-44af-a19b-d9d7d2018ad9-kube-api-access-mqmp4\") pod \"must-gather-cg6dj\" (UID: \"c80d7a41-98ec-44af-a19b-d9d7d2018ad9\") " pod="openshift-must-gather-w56h8/must-gather-cg6dj" Nov 21 15:56:20 crc kubenswrapper[4897]: I1121 15:56:20.357096 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/c80d7a41-98ec-44af-a19b-d9d7d2018ad9-must-gather-output\") pod \"must-gather-cg6dj\" (UID: \"c80d7a41-98ec-44af-a19b-d9d7d2018ad9\") " pod="openshift-must-gather-w56h8/must-gather-cg6dj" Nov 21 15:56:20 crc kubenswrapper[4897]: I1121 15:56:20.380715 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mqmp4\" (UniqueName: \"kubernetes.io/projected/c80d7a41-98ec-44af-a19b-d9d7d2018ad9-kube-api-access-mqmp4\") pod \"must-gather-cg6dj\" (UID: \"c80d7a41-98ec-44af-a19b-d9d7d2018ad9\") " pod="openshift-must-gather-w56h8/must-gather-cg6dj" Nov 21 15:56:20 crc kubenswrapper[4897]: I1121 15:56:20.463543 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-w56h8/must-gather-cg6dj" Nov 21 15:56:20 crc kubenswrapper[4897]: I1121 15:56:20.992203 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-w56h8/must-gather-cg6dj"] Nov 21 15:56:21 crc kubenswrapper[4897]: I1121 15:56:21.916374 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-w56h8/must-gather-cg6dj" event={"ID":"c80d7a41-98ec-44af-a19b-d9d7d2018ad9","Type":"ContainerStarted","Data":"5479e104954ed613f0ee1c2f0e98bfa6a8da0618df058e7aa7ab95cd437b08dd"} Nov 21 15:56:30 crc kubenswrapper[4897]: I1121 15:56:30.090383 4897 scope.go:117] "RemoveContainer" containerID="0a45fdbba95fc7088c173ed5486d2f902ed547b99722f52d109c8eadeadc8938" Nov 21 15:56:30 crc kubenswrapper[4897]: E1121 15:56:30.092096 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:56:41 crc kubenswrapper[4897]: E1121 15:56:41.068550 4897 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/openstack-must-gather:latest" Nov 21 15:56:41 crc kubenswrapper[4897]: E1121 15:56:41.072397 4897 kuberuntime_manager.go:1274] "Unhandled Error" err=< Nov 21 15:56:41 crc kubenswrapper[4897]: container &Container{Name:gather,Image:quay.io/openstack-k8s-operators/openstack-must-gather:latest,Command:[/bin/bash -c Nov 21 15:56:41 crc kubenswrapper[4897]: echo "[disk usage checker] Started" Nov 21 15:56:41 crc kubenswrapper[4897]: target_dir="/must-gather" Nov 21 15:56:41 crc kubenswrapper[4897]: usage_percentage_limit="30" Nov 21 15:56:41 crc kubenswrapper[4897]: while true; do Nov 21 15:56:41 crc kubenswrapper[4897]: disk_usage=$(du -s "$target_dir" | awk '{print $1}') Nov 21 15:56:41 crc kubenswrapper[4897]: disk_space=$(df -P "$target_dir" | awk 'NR==2 {print $2}') Nov 21 15:56:41 crc kubenswrapper[4897]: usage_percentage=$(( (disk_usage * 100) / disk_space )) Nov 21 15:56:41 crc kubenswrapper[4897]: echo "[disk usage checker] Volume usage percentage: current = ${usage_percentage} ; allowed = ${usage_percentage_limit}" Nov 21 15:56:41 crc kubenswrapper[4897]: if [ "$usage_percentage" -gt "$usage_percentage_limit" ]; then Nov 21 15:56:41 crc kubenswrapper[4897]: echo "[disk usage checker] Disk usage exceeds the volume percentage of ${usage_percentage_limit} for mounted directory, terminating..." Nov 21 15:56:41 crc kubenswrapper[4897]: pkill --signal SIGKILL gather Nov 21 15:56:41 crc kubenswrapper[4897]: exit 1 Nov 21 15:56:41 crc kubenswrapper[4897]: fi Nov 21 15:56:41 crc kubenswrapper[4897]: sleep 5 Nov 21 15:56:41 crc kubenswrapper[4897]: done & ADDITIONAL_NAMESPACES=kuttl,openshift-storage,openshift-marketplace,openshift-operators,sushy-emulator,tobiko OPENSTACK_DATABASES=ALL SOS_EDPM=all SOS_DECOMPRESS=0 gather; sync && echo 'Caches written to disk'],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:must-gather-output,ReadOnly:false,MountPath:/must-gather,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mqmp4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod must-gather-cg6dj_openshift-must-gather-w56h8(c80d7a41-98ec-44af-a19b-d9d7d2018ad9): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled Nov 21 15:56:41 crc kubenswrapper[4897]: > logger="UnhandledError" Nov 21 15:56:41 crc kubenswrapper[4897]: E1121 15:56:41.075347 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"gather\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"copy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-must-gather:latest\\\"\"]" pod="openshift-must-gather-w56h8/must-gather-cg6dj" podUID="c80d7a41-98ec-44af-a19b-d9d7d2018ad9" Nov 21 15:56:41 crc kubenswrapper[4897]: I1121 15:56:41.089910 4897 scope.go:117] "RemoveContainer" containerID="0a45fdbba95fc7088c173ed5486d2f902ed547b99722f52d109c8eadeadc8938" Nov 21 15:56:41 crc kubenswrapper[4897]: E1121 15:56:41.090603 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:56:41 crc kubenswrapper[4897]: E1121 15:56:41.182866 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"gather\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-must-gather:latest\\\"\", failed to \"StartContainer\" for \"copy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-must-gather:latest\\\"\"]" pod="openshift-must-gather-w56h8/must-gather-cg6dj" podUID="c80d7a41-98ec-44af-a19b-d9d7d2018ad9" Nov 21 15:56:52 crc kubenswrapper[4897]: I1121 15:56:52.091707 4897 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 21 15:56:53 crc kubenswrapper[4897]: I1121 15:56:53.320751 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-w56h8/must-gather-cg6dj" event={"ID":"c80d7a41-98ec-44af-a19b-d9d7d2018ad9","Type":"ContainerStarted","Data":"2b3d3cebaacc450dc4c8f80187df1f4af5cdd41dd8009fcb2f876f58c19069ce"} Nov 21 15:56:53 crc kubenswrapper[4897]: I1121 15:56:53.321441 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-w56h8/must-gather-cg6dj" event={"ID":"c80d7a41-98ec-44af-a19b-d9d7d2018ad9","Type":"ContainerStarted","Data":"39c7035d19c640376e049b31611f2a48c4cf3f8ac190ef475e93bb4e6661c500"} Nov 21 15:56:53 crc kubenswrapper[4897]: I1121 15:56:53.353197 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-w56h8/must-gather-cg6dj" podStartSLOduration=1.832013745 podStartE2EDuration="33.353157954s" podCreationTimestamp="2025-11-21 15:56:20 +0000 UTC" firstStartedPulling="2025-11-21 15:56:20.98501641 +0000 UTC m=+6458.269609885" lastFinishedPulling="2025-11-21 15:56:52.506160619 +0000 UTC m=+6489.790754094" observedRunningTime="2025-11-21 15:56:53.337700937 +0000 UTC m=+6490.622294412" watchObservedRunningTime="2025-11-21 15:56:53.353157954 +0000 UTC m=+6490.637751469" Nov 21 15:56:55 crc kubenswrapper[4897]: I1121 15:56:55.595273 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-w56h8/must-gather-cg6dj"] Nov 21 15:56:55 crc kubenswrapper[4897]: I1121 15:56:55.596103 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-w56h8/must-gather-cg6dj" podUID="c80d7a41-98ec-44af-a19b-d9d7d2018ad9" containerName="gather" containerID="cri-o://39c7035d19c640376e049b31611f2a48c4cf3f8ac190ef475e93bb4e6661c500" gracePeriod=2 Nov 21 15:56:55 crc kubenswrapper[4897]: I1121 15:56:55.596263 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-w56h8/must-gather-cg6dj" podUID="c80d7a41-98ec-44af-a19b-d9d7d2018ad9" containerName="copy" containerID="cri-o://2b3d3cebaacc450dc4c8f80187df1f4af5cdd41dd8009fcb2f876f58c19069ce" gracePeriod=2 Nov 21 15:56:55 crc kubenswrapper[4897]: I1121 15:56:55.612963 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-w56h8/must-gather-cg6dj"] Nov 21 15:56:56 crc kubenswrapper[4897]: I1121 15:56:56.089753 4897 scope.go:117] "RemoveContainer" containerID="0a45fdbba95fc7088c173ed5486d2f902ed547b99722f52d109c8eadeadc8938" Nov 21 15:56:56 crc kubenswrapper[4897]: E1121 15:56:56.090337 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:56:56 crc kubenswrapper[4897]: I1121 15:56:56.368775 4897 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-w56h8_must-gather-cg6dj_c80d7a41-98ec-44af-a19b-d9d7d2018ad9/copy/0.log" Nov 21 15:56:56 crc kubenswrapper[4897]: I1121 15:56:56.369178 4897 generic.go:334] "Generic (PLEG): container finished" podID="c80d7a41-98ec-44af-a19b-d9d7d2018ad9" containerID="2b3d3cebaacc450dc4c8f80187df1f4af5cdd41dd8009fcb2f876f58c19069ce" exitCode=143 Nov 21 15:56:58 crc kubenswrapper[4897]: I1121 15:56:58.134474 4897 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-w56h8_must-gather-cg6dj_c80d7a41-98ec-44af-a19b-d9d7d2018ad9/copy/0.log" Nov 21 15:56:58 crc kubenswrapper[4897]: I1121 15:56:58.138764 4897 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-w56h8_must-gather-cg6dj_c80d7a41-98ec-44af-a19b-d9d7d2018ad9/gather/0.log" Nov 21 15:56:58 crc kubenswrapper[4897]: I1121 15:56:58.139071 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-w56h8/must-gather-cg6dj" Nov 21 15:56:58 crc kubenswrapper[4897]: I1121 15:56:58.259358 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mqmp4\" (UniqueName: \"kubernetes.io/projected/c80d7a41-98ec-44af-a19b-d9d7d2018ad9-kube-api-access-mqmp4\") pod \"c80d7a41-98ec-44af-a19b-d9d7d2018ad9\" (UID: \"c80d7a41-98ec-44af-a19b-d9d7d2018ad9\") " Nov 21 15:56:58 crc kubenswrapper[4897]: I1121 15:56:58.259562 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/c80d7a41-98ec-44af-a19b-d9d7d2018ad9-must-gather-output\") pod \"c80d7a41-98ec-44af-a19b-d9d7d2018ad9\" (UID: \"c80d7a41-98ec-44af-a19b-d9d7d2018ad9\") " Nov 21 15:56:58 crc kubenswrapper[4897]: I1121 15:56:58.260301 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c80d7a41-98ec-44af-a19b-d9d7d2018ad9-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "c80d7a41-98ec-44af-a19b-d9d7d2018ad9" (UID: "c80d7a41-98ec-44af-a19b-d9d7d2018ad9"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 15:56:58 crc kubenswrapper[4897]: I1121 15:56:58.278578 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c80d7a41-98ec-44af-a19b-d9d7d2018ad9-kube-api-access-mqmp4" (OuterVolumeSpecName: "kube-api-access-mqmp4") pod "c80d7a41-98ec-44af-a19b-d9d7d2018ad9" (UID: "c80d7a41-98ec-44af-a19b-d9d7d2018ad9"). InnerVolumeSpecName "kube-api-access-mqmp4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 15:56:58 crc kubenswrapper[4897]: I1121 15:56:58.282054 4897 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/c80d7a41-98ec-44af-a19b-d9d7d2018ad9-must-gather-output\") on node \"crc\" DevicePath \"\"" Nov 21 15:56:58 crc kubenswrapper[4897]: I1121 15:56:58.282326 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mqmp4\" (UniqueName: \"kubernetes.io/projected/c80d7a41-98ec-44af-a19b-d9d7d2018ad9-kube-api-access-mqmp4\") on node \"crc\" DevicePath \"\"" Nov 21 15:56:58 crc kubenswrapper[4897]: I1121 15:56:58.401147 4897 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-w56h8_must-gather-cg6dj_c80d7a41-98ec-44af-a19b-d9d7d2018ad9/copy/0.log" Nov 21 15:56:58 crc kubenswrapper[4897]: I1121 15:56:58.402395 4897 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-w56h8_must-gather-cg6dj_c80d7a41-98ec-44af-a19b-d9d7d2018ad9/gather/0.log" Nov 21 15:56:58 crc kubenswrapper[4897]: I1121 15:56:58.402462 4897 generic.go:334] "Generic (PLEG): container finished" podID="c80d7a41-98ec-44af-a19b-d9d7d2018ad9" containerID="39c7035d19c640376e049b31611f2a48c4cf3f8ac190ef475e93bb4e6661c500" exitCode=137 Nov 21 15:56:58 crc kubenswrapper[4897]: I1121 15:56:58.402579 4897 scope.go:117] "RemoveContainer" containerID="2b3d3cebaacc450dc4c8f80187df1f4af5cdd41dd8009fcb2f876f58c19069ce" Nov 21 15:56:58 crc kubenswrapper[4897]: I1121 15:56:58.402851 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-w56h8/must-gather-cg6dj" Nov 21 15:56:58 crc kubenswrapper[4897]: I1121 15:56:58.448006 4897 scope.go:117] "RemoveContainer" containerID="39c7035d19c640376e049b31611f2a48c4cf3f8ac190ef475e93bb4e6661c500" Nov 21 15:56:58 crc kubenswrapper[4897]: I1121 15:56:58.499414 4897 scope.go:117] "RemoveContainer" containerID="2b3d3cebaacc450dc4c8f80187df1f4af5cdd41dd8009fcb2f876f58c19069ce" Nov 21 15:56:58 crc kubenswrapper[4897]: E1121 15:56:58.500081 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b3d3cebaacc450dc4c8f80187df1f4af5cdd41dd8009fcb2f876f58c19069ce\": container with ID starting with 2b3d3cebaacc450dc4c8f80187df1f4af5cdd41dd8009fcb2f876f58c19069ce not found: ID does not exist" containerID="2b3d3cebaacc450dc4c8f80187df1f4af5cdd41dd8009fcb2f876f58c19069ce" Nov 21 15:56:58 crc kubenswrapper[4897]: I1121 15:56:58.500206 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b3d3cebaacc450dc4c8f80187df1f4af5cdd41dd8009fcb2f876f58c19069ce"} err="failed to get container status \"2b3d3cebaacc450dc4c8f80187df1f4af5cdd41dd8009fcb2f876f58c19069ce\": rpc error: code = NotFound desc = could not find container \"2b3d3cebaacc450dc4c8f80187df1f4af5cdd41dd8009fcb2f876f58c19069ce\": container with ID starting with 2b3d3cebaacc450dc4c8f80187df1f4af5cdd41dd8009fcb2f876f58c19069ce not found: ID does not exist" Nov 21 15:56:58 crc kubenswrapper[4897]: I1121 15:56:58.500310 4897 scope.go:117] "RemoveContainer" containerID="39c7035d19c640376e049b31611f2a48c4cf3f8ac190ef475e93bb4e6661c500" Nov 21 15:56:58 crc kubenswrapper[4897]: E1121 15:56:58.500938 4897 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"39c7035d19c640376e049b31611f2a48c4cf3f8ac190ef475e93bb4e6661c500\": container with ID starting with 39c7035d19c640376e049b31611f2a48c4cf3f8ac190ef475e93bb4e6661c500 not found: ID does not exist" containerID="39c7035d19c640376e049b31611f2a48c4cf3f8ac190ef475e93bb4e6661c500" Nov 21 15:56:58 crc kubenswrapper[4897]: I1121 15:56:58.501004 4897 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39c7035d19c640376e049b31611f2a48c4cf3f8ac190ef475e93bb4e6661c500"} err="failed to get container status \"39c7035d19c640376e049b31611f2a48c4cf3f8ac190ef475e93bb4e6661c500\": rpc error: code = NotFound desc = could not find container \"39c7035d19c640376e049b31611f2a48c4cf3f8ac190ef475e93bb4e6661c500\": container with ID starting with 39c7035d19c640376e049b31611f2a48c4cf3f8ac190ef475e93bb4e6661c500 not found: ID does not exist" Nov 21 15:57:00 crc kubenswrapper[4897]: I1121 15:57:00.107899 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c80d7a41-98ec-44af-a19b-d9d7d2018ad9" path="/var/lib/kubelet/pods/c80d7a41-98ec-44af-a19b-d9d7d2018ad9/volumes" Nov 21 15:57:11 crc kubenswrapper[4897]: I1121 15:57:11.089784 4897 scope.go:117] "RemoveContainer" containerID="0a45fdbba95fc7088c173ed5486d2f902ed547b99722f52d109c8eadeadc8938" Nov 21 15:57:11 crc kubenswrapper[4897]: E1121 15:57:11.090889 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:57:24 crc kubenswrapper[4897]: I1121 15:57:24.101194 4897 scope.go:117] "RemoveContainer" containerID="0a45fdbba95fc7088c173ed5486d2f902ed547b99722f52d109c8eadeadc8938" Nov 21 15:57:24 crc kubenswrapper[4897]: E1121 15:57:24.103963 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:57:33 crc kubenswrapper[4897]: E1121 15:57:33.036133 4897 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/system.slice/systemd-hostnamed.service\": RecentStats: unable to find data in memory cache]" Nov 21 15:57:38 crc kubenswrapper[4897]: I1121 15:57:38.091611 4897 scope.go:117] "RemoveContainer" containerID="0a45fdbba95fc7088c173ed5486d2f902ed547b99722f52d109c8eadeadc8938" Nov 21 15:57:38 crc kubenswrapper[4897]: E1121 15:57:38.092840 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:57:47 crc kubenswrapper[4897]: I1121 15:57:42.332008 4897 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-6sfv4 container/packageserver namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"https://10.217.0.29:5443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 21 15:57:47 crc kubenswrapper[4897]: I1121 15:57:42.332648 4897 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6sfv4" podUID="11301ff8-fa01-48da-a95e-fc244fd0a575" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.29:5443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 21 15:57:47 crc kubenswrapper[4897]: I1121 15:57:44.081765 4897 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-gzzl8" podUID="d4c39700-beb8-4272-9f54-1c501332af0d" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.111:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 21 15:57:47 crc kubenswrapper[4897]: I1121 15:57:44.721739 4897 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="c628be92-b88f-4e02-a981-4b8fe6fd7c13" containerName="ceilometer-central-agent" probeResult="failure" output="command timed out" Nov 21 15:57:47 crc kubenswrapper[4897]: I1121 15:57:47.565585 4897 trace.go:236] Trace[682996555]: "iptables ChainExists" (21-Nov-2025 15:57:44.091) (total time: 3474ms): Nov 21 15:57:47 crc kubenswrapper[4897]: Trace[682996555]: [3.474280435s] [3.474280435s] END Nov 21 15:57:47 crc kubenswrapper[4897]: I1121 15:57:47.565597 4897 trace.go:236] Trace[1906503703]: "iptables ChainExists" (21-Nov-2025 15:57:44.093) (total time: 3472ms): Nov 21 15:57:47 crc kubenswrapper[4897]: Trace[1906503703]: [3.472492758s] [3.472492758s] END Nov 21 15:57:47 crc kubenswrapper[4897]: I1121 15:57:47.593272 4897 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 2.038191193s: [/var/lib/containers/storage/overlay/fcc5a0f06853e2f0bd9510b3c554f02106bf4f6b6f44cb9705670e26e27fa7f9/diff /var/log/pods/openshift-logging_logging-loki-gateway-78b4f7fc55-dbkmd_6dda2ab1-6304-4116-90f3-b9219aff2664/opa/0.log]; will not log again for this container unless duration exceeds 2s Nov 21 15:57:47 crc kubenswrapper[4897]: I1121 15:57:47.595305 4897 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.551587118s: [/var/lib/containers/storage/overlay/d383db55289b24b901b4adfa973934ead6177029b0f2bb4d87617bfb3583be4f/diff /var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log]; will not log again for this container unless duration exceeds 2s Nov 21 15:57:47 crc kubenswrapper[4897]: I1121 15:57:47.597548 4897 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.105751923s: [/var/lib/containers/storage/overlay/eb3a408c91ce162c9e561056b82b0b959d68a8c078104fe3768353ed6adf05be/diff /var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log]; will not log again for this container unless duration exceeds 2s Nov 21 15:57:47 crc kubenswrapper[4897]: I1121 15:57:47.588276 4897 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 2.007698681s: [/var/lib/containers/storage/overlay/171de22bf8bee0732057ec8cd4d5b6a77a7c1a24d7fa7043910d8077b1e94e45/diff /var/log/pods/openshift-logging_logging-loki-gateway-78b4f7fc55-txlqp_268847f7-dd43-4094-a01c-e5d795722e82/opa/0.log]; will not log again for this container unless duration exceeds 2s Nov 21 15:57:47 crc kubenswrapper[4897]: E1121 15:57:47.652069 4897 kubelet.go:2526] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="1.564s" Nov 21 15:57:49 crc kubenswrapper[4897]: I1121 15:57:49.090175 4897 scope.go:117] "RemoveContainer" containerID="0a45fdbba95fc7088c173ed5486d2f902ed547b99722f52d109c8eadeadc8938" Nov 21 15:57:49 crc kubenswrapper[4897]: E1121 15:57:49.091307 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:58:01 crc kubenswrapper[4897]: I1121 15:58:01.697281 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-kscx7"] Nov 21 15:58:01 crc kubenswrapper[4897]: E1121 15:58:01.700960 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c80d7a41-98ec-44af-a19b-d9d7d2018ad9" containerName="gather" Nov 21 15:58:01 crc kubenswrapper[4897]: I1121 15:58:01.700976 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="c80d7a41-98ec-44af-a19b-d9d7d2018ad9" containerName="gather" Nov 21 15:58:01 crc kubenswrapper[4897]: E1121 15:58:01.701007 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c80d7a41-98ec-44af-a19b-d9d7d2018ad9" containerName="copy" Nov 21 15:58:01 crc kubenswrapper[4897]: I1121 15:58:01.701013 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="c80d7a41-98ec-44af-a19b-d9d7d2018ad9" containerName="copy" Nov 21 15:58:01 crc kubenswrapper[4897]: I1121 15:58:01.701223 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="c80d7a41-98ec-44af-a19b-d9d7d2018ad9" containerName="gather" Nov 21 15:58:01 crc kubenswrapper[4897]: I1121 15:58:01.701254 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="c80d7a41-98ec-44af-a19b-d9d7d2018ad9" containerName="copy" Nov 21 15:58:01 crc kubenswrapper[4897]: I1121 15:58:01.706910 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kscx7" Nov 21 15:58:01 crc kubenswrapper[4897]: I1121 15:58:01.750279 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kscx7"] Nov 21 15:58:01 crc kubenswrapper[4897]: I1121 15:58:01.876292 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3efbbb26-0d50-4b7e-8aef-403a35cb92a3-utilities\") pod \"community-operators-kscx7\" (UID: \"3efbbb26-0d50-4b7e-8aef-403a35cb92a3\") " pod="openshift-marketplace/community-operators-kscx7" Nov 21 15:58:01 crc kubenswrapper[4897]: I1121 15:58:01.876663 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nn944\" (UniqueName: \"kubernetes.io/projected/3efbbb26-0d50-4b7e-8aef-403a35cb92a3-kube-api-access-nn944\") pod \"community-operators-kscx7\" (UID: \"3efbbb26-0d50-4b7e-8aef-403a35cb92a3\") " pod="openshift-marketplace/community-operators-kscx7" Nov 21 15:58:01 crc kubenswrapper[4897]: I1121 15:58:01.876870 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3efbbb26-0d50-4b7e-8aef-403a35cb92a3-catalog-content\") pod \"community-operators-kscx7\" (UID: \"3efbbb26-0d50-4b7e-8aef-403a35cb92a3\") " pod="openshift-marketplace/community-operators-kscx7" Nov 21 15:58:01 crc kubenswrapper[4897]: I1121 15:58:01.980062 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3efbbb26-0d50-4b7e-8aef-403a35cb92a3-catalog-content\") pod \"community-operators-kscx7\" (UID: \"3efbbb26-0d50-4b7e-8aef-403a35cb92a3\") " pod="openshift-marketplace/community-operators-kscx7" Nov 21 15:58:01 crc kubenswrapper[4897]: I1121 15:58:01.980249 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3efbbb26-0d50-4b7e-8aef-403a35cb92a3-utilities\") pod \"community-operators-kscx7\" (UID: \"3efbbb26-0d50-4b7e-8aef-403a35cb92a3\") " pod="openshift-marketplace/community-operators-kscx7" Nov 21 15:58:01 crc kubenswrapper[4897]: I1121 15:58:01.980277 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nn944\" (UniqueName: \"kubernetes.io/projected/3efbbb26-0d50-4b7e-8aef-403a35cb92a3-kube-api-access-nn944\") pod \"community-operators-kscx7\" (UID: \"3efbbb26-0d50-4b7e-8aef-403a35cb92a3\") " pod="openshift-marketplace/community-operators-kscx7" Nov 21 15:58:01 crc kubenswrapper[4897]: I1121 15:58:01.980753 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3efbbb26-0d50-4b7e-8aef-403a35cb92a3-catalog-content\") pod \"community-operators-kscx7\" (UID: \"3efbbb26-0d50-4b7e-8aef-403a35cb92a3\") " pod="openshift-marketplace/community-operators-kscx7" Nov 21 15:58:01 crc kubenswrapper[4897]: I1121 15:58:01.980753 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3efbbb26-0d50-4b7e-8aef-403a35cb92a3-utilities\") pod \"community-operators-kscx7\" (UID: \"3efbbb26-0d50-4b7e-8aef-403a35cb92a3\") " pod="openshift-marketplace/community-operators-kscx7" Nov 21 15:58:02 crc kubenswrapper[4897]: I1121 15:58:02.007197 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nn944\" (UniqueName: \"kubernetes.io/projected/3efbbb26-0d50-4b7e-8aef-403a35cb92a3-kube-api-access-nn944\") pod \"community-operators-kscx7\" (UID: \"3efbbb26-0d50-4b7e-8aef-403a35cb92a3\") " pod="openshift-marketplace/community-operators-kscx7" Nov 21 15:58:02 crc kubenswrapper[4897]: I1121 15:58:02.038906 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kscx7" Nov 21 15:58:02 crc kubenswrapper[4897]: I1121 15:58:02.090087 4897 scope.go:117] "RemoveContainer" containerID="0a45fdbba95fc7088c173ed5486d2f902ed547b99722f52d109c8eadeadc8938" Nov 21 15:58:02 crc kubenswrapper[4897]: E1121 15:58:02.090385 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:58:03 crc kubenswrapper[4897]: I1121 15:58:03.071081 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kscx7"] Nov 21 15:58:03 crc kubenswrapper[4897]: I1121 15:58:03.839862 4897 generic.go:334] "Generic (PLEG): container finished" podID="3efbbb26-0d50-4b7e-8aef-403a35cb92a3" containerID="1a66bbffd4d3f368c05da8f810645d486fd7c17a881b717c3946672aeda6fd96" exitCode=0 Nov 21 15:58:03 crc kubenswrapper[4897]: I1121 15:58:03.839948 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kscx7" event={"ID":"3efbbb26-0d50-4b7e-8aef-403a35cb92a3","Type":"ContainerDied","Data":"1a66bbffd4d3f368c05da8f810645d486fd7c17a881b717c3946672aeda6fd96"} Nov 21 15:58:03 crc kubenswrapper[4897]: I1121 15:58:03.840303 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kscx7" event={"ID":"3efbbb26-0d50-4b7e-8aef-403a35cb92a3","Type":"ContainerStarted","Data":"3c319cb328f266066f0a80b10d782936934c8d9c1400565ea897f9846a816b4e"} Nov 21 15:58:04 crc kubenswrapper[4897]: I1121 15:58:04.853358 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kscx7" event={"ID":"3efbbb26-0d50-4b7e-8aef-403a35cb92a3","Type":"ContainerStarted","Data":"f81d0fd73e325684537bfe5b5d5e2926ef522ccc03ce86027537cf1dc3e46641"} Nov 21 15:58:07 crc kubenswrapper[4897]: I1121 15:58:07.891806 4897 generic.go:334] "Generic (PLEG): container finished" podID="3efbbb26-0d50-4b7e-8aef-403a35cb92a3" containerID="f81d0fd73e325684537bfe5b5d5e2926ef522ccc03ce86027537cf1dc3e46641" exitCode=0 Nov 21 15:58:07 crc kubenswrapper[4897]: I1121 15:58:07.891874 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kscx7" event={"ID":"3efbbb26-0d50-4b7e-8aef-403a35cb92a3","Type":"ContainerDied","Data":"f81d0fd73e325684537bfe5b5d5e2926ef522ccc03ce86027537cf1dc3e46641"} Nov 21 15:58:09 crc kubenswrapper[4897]: I1121 15:58:09.915833 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kscx7" event={"ID":"3efbbb26-0d50-4b7e-8aef-403a35cb92a3","Type":"ContainerStarted","Data":"8b7dbe210706f78ee8e3c9deae025f0d08ac1df8e988e74da417a84cd0f30f8b"} Nov 21 15:58:09 crc kubenswrapper[4897]: I1121 15:58:09.938483 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-kscx7" podStartSLOduration=3.7793305630000003 podStartE2EDuration="8.938454292s" podCreationTimestamp="2025-11-21 15:58:01 +0000 UTC" firstStartedPulling="2025-11-21 15:58:03.842731982 +0000 UTC m=+6561.127325457" lastFinishedPulling="2025-11-21 15:58:09.001855721 +0000 UTC m=+6566.286449186" observedRunningTime="2025-11-21 15:58:09.934571878 +0000 UTC m=+6567.219165373" watchObservedRunningTime="2025-11-21 15:58:09.938454292 +0000 UTC m=+6567.223047767" Nov 21 15:58:12 crc kubenswrapper[4897]: I1121 15:58:12.039543 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-kscx7" Nov 21 15:58:12 crc kubenswrapper[4897]: I1121 15:58:12.040337 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-kscx7" Nov 21 15:58:13 crc kubenswrapper[4897]: I1121 15:58:13.092471 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-kscx7" podUID="3efbbb26-0d50-4b7e-8aef-403a35cb92a3" containerName="registry-server" probeResult="failure" output=< Nov 21 15:58:13 crc kubenswrapper[4897]: timeout: failed to connect service ":50051" within 1s Nov 21 15:58:13 crc kubenswrapper[4897]: > Nov 21 15:58:14 crc kubenswrapper[4897]: I1121 15:58:14.097605 4897 scope.go:117] "RemoveContainer" containerID="0a45fdbba95fc7088c173ed5486d2f902ed547b99722f52d109c8eadeadc8938" Nov 21 15:58:14 crc kubenswrapper[4897]: E1121 15:58:14.100482 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:58:23 crc kubenswrapper[4897]: I1121 15:58:23.092832 4897 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-kscx7" podUID="3efbbb26-0d50-4b7e-8aef-403a35cb92a3" containerName="registry-server" probeResult="failure" output=< Nov 21 15:58:23 crc kubenswrapper[4897]: timeout: failed to connect service ":50051" within 1s Nov 21 15:58:23 crc kubenswrapper[4897]: > Nov 21 15:58:29 crc kubenswrapper[4897]: I1121 15:58:29.089398 4897 scope.go:117] "RemoveContainer" containerID="0a45fdbba95fc7088c173ed5486d2f902ed547b99722f52d109c8eadeadc8938" Nov 21 15:58:29 crc kubenswrapper[4897]: E1121 15:58:29.090179 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:58:32 crc kubenswrapper[4897]: I1121 15:58:32.106717 4897 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-kscx7" Nov 21 15:58:32 crc kubenswrapper[4897]: I1121 15:58:32.161899 4897 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-kscx7" Nov 21 15:58:32 crc kubenswrapper[4897]: I1121 15:58:32.905039 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-kscx7"] Nov 21 15:58:33 crc kubenswrapper[4897]: I1121 15:58:33.187152 4897 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-kscx7" podUID="3efbbb26-0d50-4b7e-8aef-403a35cb92a3" containerName="registry-server" containerID="cri-o://8b7dbe210706f78ee8e3c9deae025f0d08ac1df8e988e74da417a84cd0f30f8b" gracePeriod=2 Nov 21 15:58:34 crc kubenswrapper[4897]: I1121 15:58:34.204077 4897 generic.go:334] "Generic (PLEG): container finished" podID="3efbbb26-0d50-4b7e-8aef-403a35cb92a3" containerID="8b7dbe210706f78ee8e3c9deae025f0d08ac1df8e988e74da417a84cd0f30f8b" exitCode=0 Nov 21 15:58:34 crc kubenswrapper[4897]: I1121 15:58:34.204535 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kscx7" event={"ID":"3efbbb26-0d50-4b7e-8aef-403a35cb92a3","Type":"ContainerDied","Data":"8b7dbe210706f78ee8e3c9deae025f0d08ac1df8e988e74da417a84cd0f30f8b"} Nov 21 15:58:34 crc kubenswrapper[4897]: I1121 15:58:34.363565 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kscx7" Nov 21 15:58:34 crc kubenswrapper[4897]: I1121 15:58:34.421707 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3efbbb26-0d50-4b7e-8aef-403a35cb92a3-utilities\") pod \"3efbbb26-0d50-4b7e-8aef-403a35cb92a3\" (UID: \"3efbbb26-0d50-4b7e-8aef-403a35cb92a3\") " Nov 21 15:58:34 crc kubenswrapper[4897]: I1121 15:58:34.422664 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nn944\" (UniqueName: \"kubernetes.io/projected/3efbbb26-0d50-4b7e-8aef-403a35cb92a3-kube-api-access-nn944\") pod \"3efbbb26-0d50-4b7e-8aef-403a35cb92a3\" (UID: \"3efbbb26-0d50-4b7e-8aef-403a35cb92a3\") " Nov 21 15:58:34 crc kubenswrapper[4897]: I1121 15:58:34.422880 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3efbbb26-0d50-4b7e-8aef-403a35cb92a3-catalog-content\") pod \"3efbbb26-0d50-4b7e-8aef-403a35cb92a3\" (UID: \"3efbbb26-0d50-4b7e-8aef-403a35cb92a3\") " Nov 21 15:58:34 crc kubenswrapper[4897]: I1121 15:58:34.426724 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3efbbb26-0d50-4b7e-8aef-403a35cb92a3-utilities" (OuterVolumeSpecName: "utilities") pod "3efbbb26-0d50-4b7e-8aef-403a35cb92a3" (UID: "3efbbb26-0d50-4b7e-8aef-403a35cb92a3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 15:58:34 crc kubenswrapper[4897]: I1121 15:58:34.446813 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3efbbb26-0d50-4b7e-8aef-403a35cb92a3-kube-api-access-nn944" (OuterVolumeSpecName: "kube-api-access-nn944") pod "3efbbb26-0d50-4b7e-8aef-403a35cb92a3" (UID: "3efbbb26-0d50-4b7e-8aef-403a35cb92a3"). InnerVolumeSpecName "kube-api-access-nn944". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 15:58:34 crc kubenswrapper[4897]: I1121 15:58:34.483180 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3efbbb26-0d50-4b7e-8aef-403a35cb92a3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3efbbb26-0d50-4b7e-8aef-403a35cb92a3" (UID: "3efbbb26-0d50-4b7e-8aef-403a35cb92a3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 21 15:58:34 crc kubenswrapper[4897]: I1121 15:58:34.526886 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nn944\" (UniqueName: \"kubernetes.io/projected/3efbbb26-0d50-4b7e-8aef-403a35cb92a3-kube-api-access-nn944\") on node \"crc\" DevicePath \"\"" Nov 21 15:58:34 crc kubenswrapper[4897]: I1121 15:58:34.527929 4897 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3efbbb26-0d50-4b7e-8aef-403a35cb92a3-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 21 15:58:34 crc kubenswrapper[4897]: I1121 15:58:34.528032 4897 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3efbbb26-0d50-4b7e-8aef-403a35cb92a3-utilities\") on node \"crc\" DevicePath \"\"" Nov 21 15:58:35 crc kubenswrapper[4897]: I1121 15:58:35.219158 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kscx7" event={"ID":"3efbbb26-0d50-4b7e-8aef-403a35cb92a3","Type":"ContainerDied","Data":"3c319cb328f266066f0a80b10d782936934c8d9c1400565ea897f9846a816b4e"} Nov 21 15:58:35 crc kubenswrapper[4897]: I1121 15:58:35.219822 4897 scope.go:117] "RemoveContainer" containerID="8b7dbe210706f78ee8e3c9deae025f0d08ac1df8e988e74da417a84cd0f30f8b" Nov 21 15:58:35 crc kubenswrapper[4897]: I1121 15:58:35.219226 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kscx7" Nov 21 15:58:35 crc kubenswrapper[4897]: I1121 15:58:35.254898 4897 scope.go:117] "RemoveContainer" containerID="f81d0fd73e325684537bfe5b5d5e2926ef522ccc03ce86027537cf1dc3e46641" Nov 21 15:58:35 crc kubenswrapper[4897]: I1121 15:58:35.268103 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-kscx7"] Nov 21 15:58:35 crc kubenswrapper[4897]: I1121 15:58:35.281950 4897 scope.go:117] "RemoveContainer" containerID="1a66bbffd4d3f368c05da8f810645d486fd7c17a881b717c3946672aeda6fd96" Nov 21 15:58:35 crc kubenswrapper[4897]: I1121 15:58:35.286046 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-kscx7"] Nov 21 15:58:36 crc kubenswrapper[4897]: I1121 15:58:36.107386 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3efbbb26-0d50-4b7e-8aef-403a35cb92a3" path="/var/lib/kubelet/pods/3efbbb26-0d50-4b7e-8aef-403a35cb92a3/volumes" Nov 21 15:58:42 crc kubenswrapper[4897]: I1121 15:58:42.089767 4897 scope.go:117] "RemoveContainer" containerID="0a45fdbba95fc7088c173ed5486d2f902ed547b99722f52d109c8eadeadc8938" Nov 21 15:58:42 crc kubenswrapper[4897]: E1121 15:58:42.090946 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:58:53 crc kubenswrapper[4897]: I1121 15:58:53.090041 4897 scope.go:117] "RemoveContainer" containerID="0a45fdbba95fc7088c173ed5486d2f902ed547b99722f52d109c8eadeadc8938" Nov 21 15:58:53 crc kubenswrapper[4897]: E1121 15:58:53.091139 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:59:08 crc kubenswrapper[4897]: I1121 15:59:08.089289 4897 scope.go:117] "RemoveContainer" containerID="0a45fdbba95fc7088c173ed5486d2f902ed547b99722f52d109c8eadeadc8938" Nov 21 15:59:08 crc kubenswrapper[4897]: E1121 15:59:08.090990 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:59:23 crc kubenswrapper[4897]: I1121 15:59:23.090310 4897 scope.go:117] "RemoveContainer" containerID="0a45fdbba95fc7088c173ed5486d2f902ed547b99722f52d109c8eadeadc8938" Nov 21 15:59:23 crc kubenswrapper[4897]: E1121 15:59:23.091160 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:59:35 crc kubenswrapper[4897]: I1121 15:59:35.089130 4897 scope.go:117] "RemoveContainer" containerID="0a45fdbba95fc7088c173ed5486d2f902ed547b99722f52d109c8eadeadc8938" Nov 21 15:59:35 crc kubenswrapper[4897]: E1121 15:59:35.090137 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 15:59:46 crc kubenswrapper[4897]: I1121 15:59:46.090127 4897 scope.go:117] "RemoveContainer" containerID="0a45fdbba95fc7088c173ed5486d2f902ed547b99722f52d109c8eadeadc8938" Nov 21 15:59:46 crc kubenswrapper[4897]: E1121 15:59:46.091222 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 16:00:00 crc kubenswrapper[4897]: I1121 16:00:00.089630 4897 scope.go:117] "RemoveContainer" containerID="0a45fdbba95fc7088c173ed5486d2f902ed547b99722f52d109c8eadeadc8938" Nov 21 16:00:00 crc kubenswrapper[4897]: E1121 16:00:00.090560 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 16:00:00 crc kubenswrapper[4897]: I1121 16:00:00.185495 4897 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29395680-p4f8z"] Nov 21 16:00:00 crc kubenswrapper[4897]: E1121 16:00:00.186127 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3efbbb26-0d50-4b7e-8aef-403a35cb92a3" containerName="registry-server" Nov 21 16:00:00 crc kubenswrapper[4897]: I1121 16:00:00.186164 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="3efbbb26-0d50-4b7e-8aef-403a35cb92a3" containerName="registry-server" Nov 21 16:00:00 crc kubenswrapper[4897]: E1121 16:00:00.186209 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3efbbb26-0d50-4b7e-8aef-403a35cb92a3" containerName="extract-utilities" Nov 21 16:00:00 crc kubenswrapper[4897]: I1121 16:00:00.186219 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="3efbbb26-0d50-4b7e-8aef-403a35cb92a3" containerName="extract-utilities" Nov 21 16:00:00 crc kubenswrapper[4897]: E1121 16:00:00.186244 4897 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3efbbb26-0d50-4b7e-8aef-403a35cb92a3" containerName="extract-content" Nov 21 16:00:00 crc kubenswrapper[4897]: I1121 16:00:00.186253 4897 state_mem.go:107] "Deleted CPUSet assignment" podUID="3efbbb26-0d50-4b7e-8aef-403a35cb92a3" containerName="extract-content" Nov 21 16:00:00 crc kubenswrapper[4897]: I1121 16:00:00.186585 4897 memory_manager.go:354] "RemoveStaleState removing state" podUID="3efbbb26-0d50-4b7e-8aef-403a35cb92a3" containerName="registry-server" Nov 21 16:00:00 crc kubenswrapper[4897]: I1121 16:00:00.187651 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29395680-p4f8z" Nov 21 16:00:00 crc kubenswrapper[4897]: I1121 16:00:00.191102 4897 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 21 16:00:00 crc kubenswrapper[4897]: I1121 16:00:00.191341 4897 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 21 16:00:00 crc kubenswrapper[4897]: I1121 16:00:00.202935 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29395680-p4f8z"] Nov 21 16:00:00 crc kubenswrapper[4897]: I1121 16:00:00.327114 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5b20dc2b-f17f-4aeb-b43b-027b668a8e42-config-volume\") pod \"collect-profiles-29395680-p4f8z\" (UID: \"5b20dc2b-f17f-4aeb-b43b-027b668a8e42\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29395680-p4f8z" Nov 21 16:00:00 crc kubenswrapper[4897]: I1121 16:00:00.327715 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5b20dc2b-f17f-4aeb-b43b-027b668a8e42-secret-volume\") pod \"collect-profiles-29395680-p4f8z\" (UID: \"5b20dc2b-f17f-4aeb-b43b-027b668a8e42\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29395680-p4f8z" Nov 21 16:00:00 crc kubenswrapper[4897]: I1121 16:00:00.327778 4897 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j4zcb\" (UniqueName: \"kubernetes.io/projected/5b20dc2b-f17f-4aeb-b43b-027b668a8e42-kube-api-access-j4zcb\") pod \"collect-profiles-29395680-p4f8z\" (UID: \"5b20dc2b-f17f-4aeb-b43b-027b668a8e42\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29395680-p4f8z" Nov 21 16:00:00 crc kubenswrapper[4897]: I1121 16:00:00.432349 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5b20dc2b-f17f-4aeb-b43b-027b668a8e42-config-volume\") pod \"collect-profiles-29395680-p4f8z\" (UID: \"5b20dc2b-f17f-4aeb-b43b-027b668a8e42\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29395680-p4f8z" Nov 21 16:00:00 crc kubenswrapper[4897]: I1121 16:00:00.432468 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5b20dc2b-f17f-4aeb-b43b-027b668a8e42-secret-volume\") pod \"collect-profiles-29395680-p4f8z\" (UID: \"5b20dc2b-f17f-4aeb-b43b-027b668a8e42\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29395680-p4f8z" Nov 21 16:00:00 crc kubenswrapper[4897]: I1121 16:00:00.432539 4897 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j4zcb\" (UniqueName: \"kubernetes.io/projected/5b20dc2b-f17f-4aeb-b43b-027b668a8e42-kube-api-access-j4zcb\") pod \"collect-profiles-29395680-p4f8z\" (UID: \"5b20dc2b-f17f-4aeb-b43b-027b668a8e42\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29395680-p4f8z" Nov 21 16:00:00 crc kubenswrapper[4897]: I1121 16:00:00.433336 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5b20dc2b-f17f-4aeb-b43b-027b668a8e42-config-volume\") pod \"collect-profiles-29395680-p4f8z\" (UID: \"5b20dc2b-f17f-4aeb-b43b-027b668a8e42\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29395680-p4f8z" Nov 21 16:00:00 crc kubenswrapper[4897]: I1121 16:00:00.440612 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5b20dc2b-f17f-4aeb-b43b-027b668a8e42-secret-volume\") pod \"collect-profiles-29395680-p4f8z\" (UID: \"5b20dc2b-f17f-4aeb-b43b-027b668a8e42\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29395680-p4f8z" Nov 21 16:00:00 crc kubenswrapper[4897]: I1121 16:00:00.450656 4897 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j4zcb\" (UniqueName: \"kubernetes.io/projected/5b20dc2b-f17f-4aeb-b43b-027b668a8e42-kube-api-access-j4zcb\") pod \"collect-profiles-29395680-p4f8z\" (UID: \"5b20dc2b-f17f-4aeb-b43b-027b668a8e42\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29395680-p4f8z" Nov 21 16:00:00 crc kubenswrapper[4897]: I1121 16:00:00.520662 4897 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29395680-p4f8z" Nov 21 16:00:01 crc kubenswrapper[4897]: I1121 16:00:01.135975 4897 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29395680-p4f8z"] Nov 21 16:00:01 crc kubenswrapper[4897]: I1121 16:00:01.367941 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29395680-p4f8z" event={"ID":"5b20dc2b-f17f-4aeb-b43b-027b668a8e42","Type":"ContainerStarted","Data":"153a026eeb81f9aa9e852095c68a22a6a71909547cfacbe69f0882fae3db2b52"} Nov 21 16:00:02 crc kubenswrapper[4897]: I1121 16:00:02.381865 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29395680-p4f8z" event={"ID":"5b20dc2b-f17f-4aeb-b43b-027b668a8e42","Type":"ContainerStarted","Data":"10c113ad4338863bd09466b9e9e6ae4702f3f53a687a8618fdcc2725d16082f8"} Nov 21 16:00:02 crc kubenswrapper[4897]: I1121 16:00:02.408005 4897 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29395680-p4f8z" podStartSLOduration=2.407982257 podStartE2EDuration="2.407982257s" podCreationTimestamp="2025-11-21 16:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-21 16:00:02.398602534 +0000 UTC m=+6679.683196019" watchObservedRunningTime="2025-11-21 16:00:02.407982257 +0000 UTC m=+6679.692575732" Nov 21 16:00:03 crc kubenswrapper[4897]: I1121 16:00:03.395963 4897 generic.go:334] "Generic (PLEG): container finished" podID="5b20dc2b-f17f-4aeb-b43b-027b668a8e42" containerID="10c113ad4338863bd09466b9e9e6ae4702f3f53a687a8618fdcc2725d16082f8" exitCode=0 Nov 21 16:00:03 crc kubenswrapper[4897]: I1121 16:00:03.396011 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29395680-p4f8z" event={"ID":"5b20dc2b-f17f-4aeb-b43b-027b668a8e42","Type":"ContainerDied","Data":"10c113ad4338863bd09466b9e9e6ae4702f3f53a687a8618fdcc2725d16082f8"} Nov 21 16:00:04 crc kubenswrapper[4897]: I1121 16:00:04.850251 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29395680-p4f8z" Nov 21 16:00:05 crc kubenswrapper[4897]: I1121 16:00:05.051673 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j4zcb\" (UniqueName: \"kubernetes.io/projected/5b20dc2b-f17f-4aeb-b43b-027b668a8e42-kube-api-access-j4zcb\") pod \"5b20dc2b-f17f-4aeb-b43b-027b668a8e42\" (UID: \"5b20dc2b-f17f-4aeb-b43b-027b668a8e42\") " Nov 21 16:00:05 crc kubenswrapper[4897]: I1121 16:00:05.051980 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5b20dc2b-f17f-4aeb-b43b-027b668a8e42-config-volume\") pod \"5b20dc2b-f17f-4aeb-b43b-027b668a8e42\" (UID: \"5b20dc2b-f17f-4aeb-b43b-027b668a8e42\") " Nov 21 16:00:05 crc kubenswrapper[4897]: I1121 16:00:05.052010 4897 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5b20dc2b-f17f-4aeb-b43b-027b668a8e42-secret-volume\") pod \"5b20dc2b-f17f-4aeb-b43b-027b668a8e42\" (UID: \"5b20dc2b-f17f-4aeb-b43b-027b668a8e42\") " Nov 21 16:00:05 crc kubenswrapper[4897]: I1121 16:00:05.052493 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5b20dc2b-f17f-4aeb-b43b-027b668a8e42-config-volume" (OuterVolumeSpecName: "config-volume") pod "5b20dc2b-f17f-4aeb-b43b-027b668a8e42" (UID: "5b20dc2b-f17f-4aeb-b43b-027b668a8e42"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 21 16:00:05 crc kubenswrapper[4897]: I1121 16:00:05.053370 4897 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5b20dc2b-f17f-4aeb-b43b-027b668a8e42-config-volume\") on node \"crc\" DevicePath \"\"" Nov 21 16:00:05 crc kubenswrapper[4897]: I1121 16:00:05.059794 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b20dc2b-f17f-4aeb-b43b-027b668a8e42-kube-api-access-j4zcb" (OuterVolumeSpecName: "kube-api-access-j4zcb") pod "5b20dc2b-f17f-4aeb-b43b-027b668a8e42" (UID: "5b20dc2b-f17f-4aeb-b43b-027b668a8e42"). InnerVolumeSpecName "kube-api-access-j4zcb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 21 16:00:05 crc kubenswrapper[4897]: I1121 16:00:05.064940 4897 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b20dc2b-f17f-4aeb-b43b-027b668a8e42-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "5b20dc2b-f17f-4aeb-b43b-027b668a8e42" (UID: "5b20dc2b-f17f-4aeb-b43b-027b668a8e42"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 21 16:00:05 crc kubenswrapper[4897]: I1121 16:00:05.158487 4897 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j4zcb\" (UniqueName: \"kubernetes.io/projected/5b20dc2b-f17f-4aeb-b43b-027b668a8e42-kube-api-access-j4zcb\") on node \"crc\" DevicePath \"\"" Nov 21 16:00:05 crc kubenswrapper[4897]: I1121 16:00:05.158539 4897 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5b20dc2b-f17f-4aeb-b43b-027b668a8e42-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 21 16:00:05 crc kubenswrapper[4897]: I1121 16:00:05.422276 4897 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29395680-p4f8z" event={"ID":"5b20dc2b-f17f-4aeb-b43b-027b668a8e42","Type":"ContainerDied","Data":"153a026eeb81f9aa9e852095c68a22a6a71909547cfacbe69f0882fae3db2b52"} Nov 21 16:00:05 crc kubenswrapper[4897]: I1121 16:00:05.422587 4897 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="153a026eeb81f9aa9e852095c68a22a6a71909547cfacbe69f0882fae3db2b52" Nov 21 16:00:05 crc kubenswrapper[4897]: I1121 16:00:05.422323 4897 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29395680-p4f8z" Nov 21 16:00:05 crc kubenswrapper[4897]: I1121 16:00:05.489102 4897 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29395635-svgmt"] Nov 21 16:00:05 crc kubenswrapper[4897]: I1121 16:00:05.501195 4897 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29395635-svgmt"] Nov 21 16:00:06 crc kubenswrapper[4897]: I1121 16:00:06.108403 4897 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a75d805d-9fcb-4bfd-8309-cf2f821e43d1" path="/var/lib/kubelet/pods/a75d805d-9fcb-4bfd-8309-cf2f821e43d1/volumes" Nov 21 16:00:13 crc kubenswrapper[4897]: I1121 16:00:13.089207 4897 scope.go:117] "RemoveContainer" containerID="0a45fdbba95fc7088c173ed5486d2f902ed547b99722f52d109c8eadeadc8938" Nov 21 16:00:13 crc kubenswrapper[4897]: E1121 16:00:13.090051 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" Nov 21 16:00:28 crc kubenswrapper[4897]: I1121 16:00:28.089400 4897 scope.go:117] "RemoveContainer" containerID="0a45fdbba95fc7088c173ed5486d2f902ed547b99722f52d109c8eadeadc8938" Nov 21 16:00:28 crc kubenswrapper[4897]: E1121 16:00:28.090408 4897 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-krv5b_openshift-machine-config-operator(e7670227-d280-4847-b882-754429f56b0f)\"" pod="openshift-machine-config-operator/machine-config-daemon-krv5b" podUID="e7670227-d280-4847-b882-754429f56b0f" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515110106246024441 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015110106247017357 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015110070575016506 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015110070575015456 5ustar corecore